Third-dimensional movies could soon be venturing outside the theaters — and ditching those red and blue paper glasses too. New research out of the Massachusetts Institute of Technology and the Computer Science and Artificial Intelligence Laboratory (CSAIL), published on July 12, devises a way to watch 3D movies at home, without the need for glasses. Called Home3D, the new platform converts existing stereoscopic 3D movies by using artificial intelligence, rather than using a new camera system for natively recording the content.
The 3D glasses correct a pair of offset, polarized images to create a sense of depth. Home 3D, however, uses what’s called an automultiscopic display. The display is actually three (or more) images, but those images are presented in a slightly offset way that looks different when the screen is viewed from different angles. That allows the brain to see a coherent image with depth information, creating a 3D effect without the glasses.
Automultiscopic displays have already allowed for glasses-free TVs. While the display type recently introduced is promising, it requires creating a 3D video using 30 different cameras. The unusual format creates a sort of chicken-and-egg problem: Producers aren’t going to create this 30-camera content when no consumers actually own a automultiscopic TV, but no consumers are actually going to buy a automultiscopic TV when there’s no content being produced for it. Companies like Toshiba already have glasses-free displays on the market, but they’ve failed to grow in popularity because of limited content.
What the latest MIT/CSAIL research does is to create a way to convert existing, glasses-required 3D content into the proper format for automultiscopic displays (the study is based off a similar work that focused on glasses-free viewing at movie theaters last year). By converting content that already exists into the new glasses-free format, automultiscopic TVs could actually wind up in homes since it only requires some software to convert content to the new format. That’s why when YouTube announced a new 180-degree format, they launched a way to view the content and, with manufacturing partners, a way to create the content at the same time.
“Automultiscopic displays aren’t as popular as they could be because they can’t actually play the stereo formats that traditional 3D movies use in theaters,” said Petr Kellnhofer, lead author on the Home3D research. “By converting existing 3D movies to this format, our system helps open the door to bringing 3D TVs into people’s homes.”
The MIT/CSAIL study isn’t the first to try computerized conversions to the new format, but reverses some of the current limitations. Phase-based rendering is a quick and accurate method for converting to the video type, but it can’t handle every type of 3D image. A technique called depth image-based rendering can handle those types of images, but is low-resolution and loses finer details like in transparent objects or motion blur.
The new software uses AI to mix the two conversion methods, creating a high-resolution 3D result without the glasses and without the limitations of earlier methods. The software can convert videos in real-time, running on a graphics processing unit (GPU), which means that the program could run off existing gaming devices like an Xbox or PlayStation, effectively mixing existing videos with existing hardware through software conversion. Future media players or smart automultiscopic TVs could integrate a GPU to enable the conversion as well, the research group says.
When Home3D-converted movies including The Avengers and Big Buck Bunny were presented to a study group, 60 percent of participants said the quality was better than existing 3D videos.
As we mentioned in our DT10 series on the future of the television, glasses-free 3D TVs will be a future technology manufacturers tap into, and this research backs up this trend. While the new method brings glasses-free home TVs closer to reality, the researchers say the current program does create some ghosting, or offset images, an issue the team plans to continue to refine through software.
- What is MPEG-H? The burgeoning 3D audio standard explained
- What is an audio codec? A jargon-free explainer
- How do you film what isn’t real? Joe Hunting on his documentary We Met in Virtual Reality
- LED vs. LCD TVs explained: What’s the difference?
- ATSC 3.0: Everything you need to know about broadcast TV’s next big thing