Product Reviews

The Pixel 10 Pro puts generative AI right inside the camera

The Pixel 10 Pro puts generative AI right inside the camera


At The Verge, we like to ask “What is a photo?” when we’re trying to sort out real and unreal images — especially those taken with phone cameras. But I think there’s another question that we’ll want to add to the mix starting right now: what is a camera? With the introduction of the Pixel 10 Pro and Pro XL, that answer is more wild and complicated than ever, because generative AI isn’t just something you can use to edit a photo you’ve already taken; it’s baked right into the camera itself.

I’m talking about Pro Res Zoom, which is not to be confused with Apple’s ProRes video format or Google’s Super Res Zoom, so help us all. Pro Res Zoom kicks in when past 30x, all the way up to 100x digital zoom. Typically, the camera uses an algorithm to help fill in the gaps left by upscaling a small portion of your photo to the original resolution. Typically, the results look like hot garbage, especially when you get all the way to 75x or 100x, despite every camera maker’s best efforts over the past two decades. Pro Res Zoom aims to give you a usable image where you wouldn’t have gotten one before — and that’s where the diffusion model comes in.

It’s a latent diffusion model, Google’s Pixel camera product manager Isaac Reynolds tells me. He doesn’t see it as an entirely new process — more like a variation on what phone cameras have done for years. Algorithms have long helped identify subjects and improve detail, producing unwanted artifacts as a byproduct that engineers squash in subsequent updates. “Generative AI is just a different algorithm with different artifacts,” he says. But as opposed to a more conventional neural network, a diffusion model is “pretty good at killing the artifacts.”

That might be an understatement. In the handful of demos I saw, Pro Res Zoom cleaned up some pretty gnarly 100x zoom photos remarkably well. The processing all happens on device after you take the photo. Reynolds tells me that when Google started developing the feature, it took around a minute to run the diffusion model on the phone; his team got the runtime down to four or five seconds. Once the processing is done, the new version is saved alongside the original. I only saw it work a handful of times, but the results I saw looked pretty darn good.

1/3

The original photo before Pro Res Zoom.

Pro Res Zoom has one important guardrail: it doesn’t work on people. If it detects a person in the image, it’ll work around them and enhance everything else, leaving the human be. This is a good idea, not only because I do not want a phone camera hallucinating different features onto my face, but also because it could be problematic from a creepiness standpoint.

Google has also taken a responsible step to tag photos taken with the phone using C2PA content credentials, labeling Pro Res Zoom photos as “edited with AI tools.” But it doesn’t stop there — all photos taken with the Pixel 10 get tagged to indicate that they were taken with a camera and whether AI played a role. If a photo is the result of merging multiple frames, like a panorama, that’ll be noted in the content credentials, too.

The Pixel 10 labels all photos taken with its camera using C2PA content credentials.

It’s all in an effort to reduce the “implied truth effect,” Reynolds explains. If you only apply labels to AI-generated images, then anything without an AI label seems to be authentic. But that only really means that the origin of an image is unknown, especially in an age of easy access to AI editing and image generation tools. It could have been edited with AI and not tagged as such, or the tag could have been removed by taking a screenshot and sharing that image instead.

The thing is, C2PA credentials can’t be modified once they’re created. Looking for a tag to positively identify an image as being camera-created becomes one of the only surefire ways of knowing that what you’re looking at isn’t AI. If that’s the future we’re moving toward, then there’s a massive gap between that reality and the one we live in now.

“I do think there’s going to be a period of education,” Reynolds acknowledges. He thinks that phase is already well underway, and I agree. But there is still potential for real harm — to people and our institutions — between now and that future, and that’s what makes me most uncomfortable about this whole moment.

Is a camera that uses AI to clean up your crappy zoom photos still just a camera? Probably, for now

Misgivings aside, I still had one question I needed an answer to: what exactly is an image taken with Pro Res Zoom? A memory? A robot’s best guess at what a tree looks like? A moment lost in time, like tears in the rain? If I take a Pro Res Zoom picture of the Statue of Liberty, is it really a photo that I took? Reynolds thinks so.

“Pro Res Zoom is tuned very carefully to just be a picture,” he says. “There’s nothing about Pro Res Zoom that changes what you’re expecting from a camera. Because that’s how we built it, that’s what we wanted it to be.”

Is a camera that uses AI to clean up your crappy zoom photos still just a camera? Probably, for now. But there’s a door open for someone who wants to build something else — and a lot of questions to ask in the meantime.

Photos by Allison Johnson / The Verge

Follow topics and authors from this story to see more like this in your personalized homepage feed and to receive email updates.


Leave a Reply

Your email address will not be published. Required fields are marked *