A U.K. woman was photographed standing in a mirror where her reflections didn’t match, but not because of a glitch in the Matrix. Instead, it’s a simple iPhone computational photography mistake.
A U.K. woman was photographed standing in a mirror where her reflections didn’t match, but not because of a glitch in the Matrix. Instead, it’s a simple iPhone computational photography mistake.
But… no? When I change the key frame (the image) of the video, then send that to someone as a picture, the picture is of that key frame. If I send over iMessage as Live Photo, it’s both a pic and a video.
So it gives the the ability to slightly adjust when the picture was taken to get that perfect shot with no one blinking and things look just right. Basically, every picture is like 50 pictures and I can pick the best. By default, it picks the middle.
Call it what you want, but it’s integrated pic + vid on every picture. Yes. Coupled with simple tools to leverage that for some nice functionality.
If your argument is just “but it’s simple, it’s just a video” then you’re ignoring the UX execution entirely.
90% of people share their photos from the messages app, they aren’t going into their photos and explicitly sharing what they already think is a photo, as a photo.
I get where you’re coming from, but “this buried feature most don’t know about” isn’t what I’m talking about, I’m talking about what the majority of people experience when using them.
For instance, if you take a photo in the messages app, and don’t turn off the Live Photo option, you are presented with what looks like a photo to send. You aren’t given options to change the key frame, you aren’t given much warning that it’s actually a video, for all intents and purposes it’s a photo to most users. Which is why I’ve been sent bits and pieces of conversations people didn’t know they were sharing.