Skip to main content

This tiny sensor is about to change your phone camera forever

“We believe there’s a real opportunity to develop and use a spectral imager in a smartphone. Despite all the progress which has been made with different cameras and the computing power of a smartphone, none can really identify the true color of a picture.”

This is how Spectricity CEO Vincent Mouret described the company’s mission to Digital Trends in a recent interview, as well as the reason why it’s making a miniaturized spectral image sensor that’s ready for use in a smartphone. But what exactly is a spectral sensor, and how does it work? It turns out, it can do a whole lot more than just capture pretty colors.

Have you ever seen this color?

A headshot showing Spectricity CEO Vincent Mouret and application engineer Michael Jacobs.
Spectricity CEO Vincent Mouret (from left) and application engineer Michael Jacobs Spectricity

Before we get to all that, let’s talk about what Spectricity’s S1 multispectral image sensor does for cameras. Most phone cameras use three-color RGB (red, green, and blue) sensors, but the Spectricity S1 sensor looks further into the visible light and near-infrared range to reproduce more natural, more consistent colors, along with considerably improved white balance.

“A standard camera integrated into a smartphone has an RGB sensor that sees red, green, and blue,” Mouret said. “We add filters to create up to 16 different images with different colors, different wavelengths of light, light coming from different sources, and the reflected light coming from the object of the scene. You can identify many different properties thanks to these different images compared to a standard RGB.”

What this means is that no matter the lighting conditions, the images taken by a phone with the S1 sensor will have more consistent colors throughout, as you can see in the example image below. During our conversation, I saw a live demonstration that replicated the color reproduction and consistency seen in the example, but it’s worth noting the sensor was operated by a PC and not a smartphone.

An example of how the Spectricity S1 sensor differs from smartphone cameras.
A chart showing how the Spectricity S1 sensor differs from smartphone cameras. Spectricity

“You will see with Xiaomi, Samsung, Apple, and Huawei cameras the colors are very different in different lighting conditions. In the same scene taken with our spectral imager, there’s slight differences, but minor ones. The colors out of images are the colors seen by our naked eyes. You can see all the smartphones produce very different colors.”

This color accuracy will also improve the ability of smartphone cameras to better reproduce different skin tones. In another example image, the S1 sensor appeared to show considerable improvements here too.

“You can see the skin tones are totally different depending on the lighting condition,” Mouret pointed out. “The solution is to use a spectral imager to analyze the lighting conditions, to really give the right tone. This is the only way. You can put a lot of AI behind it, but it’s not enough. You need to have some additional hardware.”

Unfortunately, Mouret didn’t have a Google Pixel phone on hand to compare the S1 sensor with Google’s Real Tone computational technology, which promises to do something similar, only by using software enhancements.

Is it a sensor, or a camera?

Spectricity's S1 Spectral Image Sensor.

The benefits of a spectral image sensor like the S1 seem to be clear, and who doesn’t want more effective white balance and natural, consistent colors in their photographs? But the way Mouret described the S1’s ability made it sound more like a camera than a sensor. Which is it? Can it really fit inside a smartphone? If so, what modifications will be needed?

Spectricity application engineer Michael Jacobs, who ran the demo I watched, explained what the S1 actually is.

“It’s really a companion sensor, very similar to a depth sensor or a 3D sensor used with an RGB camera,” he confirmed.

However, it does technically take photos, just not ones you’d want to use individually, as Mouret explained:

“Our image sensor has a VGA resolution, 800 x 600 pixels. You need to combine this smaller spectral image with an RGB image. The module itself is really very small. It has been designed to be integrated in a smartphone. So it’s 5mm by 5mm by 6mm,” Mouret told me, before saying more about how the sensor is built. “It’s not easy, but it’s nothing very sophisticated, I would say,” he explained, saying the module is built using the same methods as regular CMOS sensors. “In terms of overall cost,  the complete module will be the same cost as a high-volume standard camera.”

Example of the differences between an RGB sensor and Spectricity's S1 sensor.
A graphic showing the difference between an RGB sensor and the Spectricity S1 spectral imager. Spectricity

Right now, the sensor will integrate with a standard Image Signal Processor (ISP) used by smartphone processors from Qualcomm and MediaTek, but it will require additional software to operate.

The S1 multispectral image sensor isn’t a one-trick pony, and actually has an intriguing second function, in that it can recognize and analyze skin biomarkers. What this means is it’s able to tell if it’s “seeing” a real person, which could be used for security applications — for example, it knows if someone is wearing a mask to conceal their identity — skin care, and even in health applications when combined with AI and other software. It could also be used to enhance Portrait Mode photos.

When will it be used in a phone?

Oppo Find N2's camera module.
Andy Boxall/Digital Trends

Manufacturers have been working to improve color reproduction on phones for a while, going back to the LG G5 in 2016, which used a dedicated color spectrum sensor alongside its laser autofocus sensor to improve color performance. Huawei took a different approach and abandoned an RGB sensor on the P30 Pro in favor of an RYYB sensor, in its effort to better reproduce colors and improve lowlight performance. Olympus and Imec have also experimented with near-infrared RGB sensors for different applications.

The Spectricity S1 sensor is a world’s first and goes in a different direction agin from these examples. Based on the demo and example images, it’s showing considerable promise — but when should we expect to see it on a phone?

“In the beginning, it will be introduced on high-end phones,” Mouret said. “It will be small volume in 2024, a higher volume in 2025, and then starting in 2026, it will be more widespread.”

The Huawei Mate 50 Pro and Apple iPhone 14 Pro camera modules.
Andy Boxall/Digital Trends

Mouret also expects the first examples to come from Chinese smartphone manufacturers, and not from companies like Samsung and Apple. However, Mouret expects this to change in the future, and in the CES 2023 press release for the S1 sensor, he didn’t hold back on his expectations, saying:

“We expect the first smartphone models with the S1 to be released in 2024, and we expect all smartphones to include our technology within the coming years.”

Yes, all smartphones. It’s a big target, but if the performance we saw in the demo is what we will eventually see from a phone camera with the S1 alongside it, few manufacturers will want to be left behind.

Editors' Recommendations

Andy Boxall
Senior Mobile Writer
Andy is a Senior Writer at Digital Trends, where he concentrates on mobile technology, a subject he has written about for…
Samsung’s new 200-megapixel sensor may appear on S23 Ultra
Galaxy S22 Ultra's screen.

Last year in September, Samsung showcased its first 200-megapixel sensor and a fresh report suggests that the company has another one in the works. It is supposed that the S23 Ultra will come equipped with this new sensor. The Korean publication ET News broke out the news stating that the 200-megapixel sensor is in the final stage of development. According to the report, Samsung Electronics mobile division and Samsung Electro-Mechanics are the two divisions behind the making of the sensor. Thirty percent of the camera sensors will be manufactured by the former while the latter will be producing the other seventy percent.

Samsung flagships have been shipping with massive 108-megapixel sensors for a while now. But it's likely not enough as the company plans to incorporate a 200-megapixel camera on its upcoming flagship phone. At the moment there isn’t a lot to talk about this new sensor but we expect Samsung to reveal more details soon.

Read more
Galaxy S22 Ultra’s camera takes on the iPhone 13 Pro’s
Galaxy S22 Ultra and iPhone 13 Pro cameras seen from the back.

Samsung and Apple will always have a heated rivalry, and the Galaxy S22 Ultra’s natural opposite is the Apple iPhone 13 Pro or iPhone 13 Pro Max. The camera is a major selling point for both these phones, and even if someone considering the S22 Ultra would maybe never buy an iPhone (and vice versa), it’s always good to know if one is better than the other.

I took a selection of photos over a few days with both smartphones to see which one wins the bragging rights of having the best camera.
Camera specs
The Galaxy S22 Ultra’s headline 108-megapixel camera is joined by a 12MP 120-degree wide-angle camera, a 3x optical telephoto zoom camera, and a 10x optical zoom telephoto. The iPhone 13 Pro is quite different on paper from the Galaxy S22 Ultra. It has three 12-megapixel cameras: A main camera, one providing a 120-degree field-of-view wide-angle shot, and another for 3x telephoto optical zoom.

Read more
Oppo details supersmooth optical zoom and new camera sensor for future phones
Oppo Find X3 Pro camera module

Oppo has detailed exciting new camera technology it’s working on at the moment, and some of it may hit phones we can buy before the end of the year. There are three main new developments, with the first focused on a new Red, Green, Blue, White (RGBW) camera sensor. The others concern an optical zoom telephoto camera and image stabilization technology.

The RGBW sensor promises to deliver photos with better color performance and accuracy, as well as improved lowlight ability. The sensor uses a red, green, blue, and white filter, unlike the more common red, green, green, and blue (RGGB) sensors on most other smartphones, along with 4-in-1 pixel technology to let in 60% more light. It reduces image noise by 35%, and Oppo also says the sensor will show more realistic skin tones.

Read more