Skip to main content

Google is revolutionizing smartphone photos with computing, not lenses

Google Pixel 3
Julian Chokkattu/Digital Trends

The new Google Pixel 3 and Pixel 3 XL smartphones take low-light images, high-resolution photos, and well-timed shots — but these major photo features aren’t realized solely by the cameras packed inside. Instead, Google is tackling tasks typically left to larger cameras with computing power — specifically machines learning — not lenses and high-resolution sensors.

(Read our hands-on preview of the new Google Pixel 3 and Pixel 3 XL.)

Like the Pixel 2, Google integrated a special chip designed just for photos, the Pixel Visual Core, and a dual-pixel sensor that enables dual-lens effects with a single lens. And like the original Pixel phone, the Pixel 3 shoots and merges multiple images without a delay by using HDR+. And like the first two generations of Google smartphones, Google isn’t done leveraging artificial intelligence and computational photography to take better photos.

Good-bye, crappy smartphone zoom?

Smartphone cameras have either a slight zoom using two lenses or digital zoom — and all digital zooms produce poor results by cropping the photos. You just can’t fit a big zoom lens inside a small smartphone. Google is promising better zoom with a fixed, single lens smartphone (on the rear, anyways) using Super Res Zoom.

The feature doesn’t appear to require a tripod, since it actually needs those small movements in your hands.

Super Res Zoom revamps an existing idea and reworks the concept to solve a new problem — that crappy smartphone zoom. Digital zoom doesn’t work well because the resolution is drastically reduced — but what if the image you started with had a higher resolution?

Super Res Zoom takes a burst of photos. Small movements in your hands will make those photos taken from a slightly different position. By stitching those slightly different photos together, the Pixel 3 creates a higher resolution image. And with a higher resolution image, you can use digital zoom with results that aren’t so cringe-worthy.

Perhaps what’s even more intriguing is that the feature doesn’t appear to require a tripod, since it actually needs those small movements in your hands. Panasonic, Olympus, and Pentax cameras have similar modes using pixel shift, but they are designed to create a higher resolution final file, not as an artificial zoom, and tripods are recommended.

A good low-light smartphone?

Speaking of cringe-worthy, Google’s Liza Ma says that the Pixel 3’s new low-light mode called Night Sight is so good, you’ll never use the flash. Like the Super Res Zoom, the feature is powered by machine learning. The Night Sight doesn’t use any of the usual hardware solutions for a better low light shot like a larger sensor and brighter aperture — instead, machine learning re-colors the photo to create brighter, more vivid colors even without using the flash.

Google didn’t dive much into detail how machine learning is used to brighten the photos, but says the A.I. recolors the image for a brighter shot without the flash. We’ll have to wait to see just how well that recoloring works — the feature isn’t launching until next month via software.

Top Shot mixes HDR+ with A.I. that chooses your best photos for you

The Top Shot feature inside the Pixel 3 is essentially burst mode — a fast series of photos — and a feature that DSLRs and even smartphones have long had. But what Google is doing different with Top Shot is automatically choosing which moment out of that burst is the best one.

Top Shot takes a fast burst of photos. The Pixel 3 highlights the one with your actual timing, and also highlights a recommended photo. Machine learning, Google says, determines which image in that burst is the best option. By feeding a computer a bunch of good photos and bad photos, essentially, the software learned that, yes, photos are better with everyone’s eyes open and a smile in the frame. And if you don’t agree with the A.I.’s pick, you can dig through the burst and choose the image yourself.

Pixel 3: Top Shot

Google says the alternate shots are still also captured in HDR+ — so essentially, that burst mode is also taking smaller bursts to layer together for a more detailed image. HDR+ already impressed in earlier Pixel models, but managing both burst shots and multiple images at once suggests impressive computing power. (And yes, those photos will probably take up a lot of space, but Google is including unlimited Google Photos storage with the Pixel 3).

The idea of using A.I. to choose your best shots is nothing new — Adobe announced a beta tool for Lightroom to do just that a year ago. But what the Pixel 3 does is mix that new concept of automatically flagging your best shots without sifting through the bad ones with the old school burst mode. And it’s all done on one device.

So where does hardware fit in?

While the biggest new features are powered by A.I., the Pixel 3 doesn’t leave camera hardware unchanged. The front of the phone now houses two cameras — one an expected 8-megapixel camera, the other a wide angle lens with a 97-degree field of view so you can actually fit everyone into a groupie. A portrait booth mode will also trigger the shot hands-free by looking for a smile or a funny face, Google says.

Google Pixel 3 and Pixel 3 XL
Julian Chokkattu/Digital Trends

The camera keeps a single lens at the back, yet manages to continue the impressive portrait mode from earlier models using dual pixel technology instead of dual lenses. That portrait mode is getting a boost, Google says — the Pixel 3 can edit the result, including changing the subject for a sharp background instead.

The camera’s dual pixel autofocus can also now track subjects — a feature that’s been around for some time on advanced cameras but is a nice addition to see integrated into a smartphone. The rear camera also includes optical and electronic image stabilization, a flicker sensor and a bright f/1.8 lens.

Video is shot at up to 30 fps 4K or 120 fps in 1080p.

Filmed on Pixel 3 | In collaboration with Terrence Malick

Google may have made some claims that are no big deal for DSLR fans like tracking autofocus, but pit the Pixel 3’s camera against other smartphones and those A.I. features could give the Pixel 3 an edge. Annie Leibovitz, at least, agrees — she’s entered into a partnership with Google, the first time the photographer has signed an agreement with a brand. She’s not saying anything about leaving her dedicated camera behind, but Leibovitz did use the Pixel 3 for portraits and places while traveling, Google says.

Of course, we will be putting these features to the test when they become available, so stay tuned for our full reviews of both products.

Editors' Recommendations

Hillary K. Grigonis
Hillary never planned on becoming a photographer—and then she was handed a camera at her first writing job and she's been…
Don’t update your Google Pixel phone — you might break it
A person holding the Google Pixel 8, showing the screen.

One of the reasons to buy a Google Pixel phone is to be first in line to receive software updates — from new Android versions to important security patches. Unfortunately, one of the latest updates from Google is breaking some Pixel phones.

Over the weekend, a Reddit user on the r/GooglePixel subreddit compiled a list of threads from nearly a dozen Pixel owners reporting issues with their phones after downloading the most recent January 2024 Google Play system update.

Read more
The Google Pixel 8 is getting a surprise update next week
Someone holding the Bay blue Google Pixel 8 Pro.

You're in luck if you've been holding off on ordering a new Google Pixel 8 Pro. According to a post on the official Made by Google account on X (formerly Twitter), a new color option is coming.

The Pixel 8 Pro is currently available in three colors: the blueish Bay, the black-like Obsidian, and Porcelain. The Bay Pixel 8 Pro is drawn over in a light green shade in the post. A link in the post takes you to a "Minty Fresh" page that notes Thursday, January 25, as the big reveal date. Furthermore, the binary code in the post translates to "Fresh year, fresh drop."

Read more
There’s something Samsung didn’t tell you about the Galaxy S24
The Samsung Galaxy S24 Ultra in its launch colors.

“Look, dude, I don’t know how they are going to manage the costs of licensing from AI companies and cloud partnerships, among other associated factors. Nothing comes free.” That’s what a machine learning engineer told me a few days ago when I explained to him how Qualcomm and MediaTek are bringing some neat generative AI tricks to phones.

Well, Samsung has confirmed those fears and quietly dropped the bombshell that at least some of its snazzy AI tricks for the Galaxy S24 series phones will eventually ask you to cough up some cash. That’s going to happen next year, but we don’t know how much you'll have to pay and in what way — at least not right now.

Read more