Skip to main content

A.I. helps scientists inch closer to the ‘holy grail’ of computer graphics

Fur Real - Scientists Improve Computer Rendering of Animal Fur

Computer scientists at the University of California, San Diego, and UC Berkeley devised a way to make animals in movies and video games more realistic by improving the look of computer-generated fur. It might not sound like much but the researchers call photorealistic fur a “holy grail” of computer graphics.

“Creating photorealistic … characters has long been one of the holy grails of computer graphics in film production, virtual reality, and for predictive design,” Ravi Ramamoorthi, a professor of computer science at UC San Diego, who worked on the project, told Digital Trends. “Realistic rendering of animal fur is a key aspect to creating believable animal characters in special effects, movies, or augmented reality.”

To do so, they leveraged artificial intelligence to better reflect the way light bounces between the fur of an animal pelt, which has a surprisingly significant effect on realism.

Existing models were designed to depict human hair and were less focused on animal fur. However, while human hair and fur both contain an interior cylinder called a medulla, the medulla in fur is much bigger than in hair, and creates an unusual scattering of light. Most existing models haven’t taken the medulla into account in this complex scattering of light.

But the team from UC San Diego and UC Berkeley turned to a concept called subsurface scattering and employed an A.I. algorithm to lend a hand.

“A key innovation is to relate fur rendering to subsurface scattering, which has earlier been used for things like clouds or human skin,” Ramamoorthi said. “There are many techniques to render subsurface scattering efficiently, but the parameters are completely different physically from those used to describe fur reflectance. We have introduced a simple neural network that relates them, enabling one to translate a fur reflectance model to comparable subsurface parameters for fast rendering.”

In terms of speed, Ramamoorthi said his team’s model can generate more accurate simulations ten times faster than the current state-of-the-art models. They shared their findings last week at the SIGGRAPH Asia conference in Thailand.

The new model has future potential in fields from virtual reality to video games, but Ramamoorthi seemed most enthusiastic about its current use for special effects in films.

“Our fur reflectance model is already used, for example in the Rise of the Planet of the Apes, nominated for a visual effects Oscar this year,” he said.

Dyllan Furness
Dyllan Furness is a freelance writer from Florida. He covers strange science and emerging tech for Digital Trends, focusing…
This AI cloned my voice using just three minutes of audio
acapela group voice cloning ad

There's a scene in Mission Impossible 3 that you might recall. In it, our hero Ethan Hunt (Tom Cruise) tackles the movie's villain, holds him at gunpoint, and forces him to read a bizarre series of sentences aloud.

"The pleasure of Busby's company is what I most enjoy," he reluctantly reads. "He put a tack on Miss Yancy's chair, and she called him a horrible boy. At the end of the month, he was flinging two kittens across the width of the room ..."

Read more
Digital Trends’ Top Tech of CES 2023 Awards
Best of CES 2023 Awards Our Top Tech from the Show Feature

Let there be no doubt: CES isn’t just alive in 2023; it’s thriving. Take one glance at the taxi gridlock outside the Las Vegas Convention Center and it’s evident that two quiet COVID years didn’t kill the world’s desire for an overcrowded in-person tech extravaganza -- they just built up a ravenous demand.

From VR to AI, eVTOLs and QD-OLED, the acronyms were flying and fresh technologies populated every corner of the show floor, and even the parking lot. So naturally, we poked, prodded, and tried on everything we could. They weren’t all revolutionary. But they didn’t have to be. We’ve watched enough waves of “game-changing” technologies that never quite arrive to know that sometimes it’s the little tweaks that really count.

Read more
Digital Trends’ Tech For Change CES 2023 Awards
Digital Trends CES 2023 Tech For Change Award Winners Feature

CES is more than just a neon-drenched show-and-tell session for the world’s biggest tech manufacturers. More and more, it’s also a place where companies showcase innovations that could truly make the world a better place — and at CES 2023, this type of tech was on full display. We saw everything from accessibility-minded PS5 controllers to pedal-powered smart desks. But of all the amazing innovations on display this year, these three impressed us the most:

Samsung's Relumino Mode
Across the globe, roughly 300 million people suffer from moderate to severe vision loss, and generally speaking, most TVs don’t take that into account. So in an effort to make television more accessible and enjoyable for those millions of people suffering from impaired vision, Samsung is adding a new picture mode to many of its new TVs.
[CES 2023] Relumino Mode: Innovation for every need | Samsung
Relumino Mode, as it’s called, works by adding a bunch of different visual filters to the picture simultaneously. Outlines of people and objects on screen are highlighted, the contrast and brightness of the overall picture are cranked up, and extra sharpness is applied to everything. The resulting video would likely look strange to people with normal vision, but for folks with low vision, it should look clearer and closer to "normal" than it otherwise would.
Excitingly, since Relumino Mode is ultimately just a clever software trick, this technology could theoretically be pushed out via a software update and installed on millions of existing Samsung TVs -- not just new and recently purchased ones.

Read more