Skip to main content

OpenAI’s Advanced Voice Mode can now see your screen and analyze videos

Advanced Santa voice mode
OpenAI

OpenAI’s “12 Days of OpenAI” continued apace on Wednesday with the development team announcing a new seasonal voice for ChatGPT’s Advanced Voice Mode (AVM), as well as new video and screen-sharing capabilities for the conversational AI feature.

Santa Mode, as OpenAI is calling it, is a seasonal feature for AVM, and offers St. Nick’s dulcet tones as a preset voice option. It is being released to Plus and Pro subscribers through the website and mobile and desktop apps starting today and will remain so until early January. To access the limited-time feature, first sign in to your Plus or Pro account, then click on the snowflake icon next to the text prompt window.

Recommended Videos

Select Santa’s voice from the popup menu, confirm your choice, and start chatting. I, for one, am not entirely clear on why you’d want to talk to a large language model masquerading as a fictional religious figure, much less shell out $20 for the privilege, but OpenAI seems to believe it holds value. Note that the system will not log your chats with Santa, they won’t be saved to your chat history, nor will they impact the memory of ChatGPT.

Just in time for the holidays, video and screensharing are now starting to roll out in Advanced Voice in the ChatGPT mobile app. pic.twitter.com/HFHX2E33S8

— OpenAI (@OpenAI) December 12, 2024

The company is also rolling out a long-awaited feature for Advanced Voice Mode: the ability to analyze video and screen shares through the mobile AVM interface. With it, you’ll be able to share your screen or video feed with ChatGPT, hold real-time conversations, and get it to answer questions about what you see, without needing to describe your surroundings or upload photos.

The new feature is rolling out to Plus and Pro subscribers “in most countries” according to the company, as well as to all Teams users. Stringent privacy laws are delaying the feature’s release in the EU, Switzerland, Iceland, Norway, and Liechtenstein, though the company hopes to get it to Plus and Pro subscribers in those regions “soon.” Enterprise and Edu users will have to wait until January to try it for themselves. If you have access, you can launch the new feature by opening voice mode, then tapping the video camera icon in the lower left. To launch screen share, just tap the three-dot menu and select “Share Screen.”

Wednesday’s announcement marks the fourth day of OpenAI’s live stream event. The company has already unveiled its fully-functional 01 reasoning model, its Sora video generation model, a $200/month Pro subscription tier, and updates to ChatGPT’s Canvas.

Andrew Tarantola
Andrew Tarantola is a journalist with more than a decade reporting on emerging technologies ranging from robotics and machine…
OpenAI teases its ‘breakthrough’ next-generation o3 reasoning model
Sam Altman describing the o3 model's capabilities

For the finale of its 12 Days of OpenAI livestream event, CEO Sam Altman revealed its next foundation model, and successor to the recently announced o1 family of reasoning AIs, dubbed o3 and 03-mini.

And no, you aren't going crazy -- OpenAI skipped right over o2, apparently to avoid infringing on the copyright of British telecom provider O2.

Read more
Yes, it’s real: ChatGPT has its own 800 number
1-800-chatgpt

On the 10th of its "12 Days of OpenAI" media event, the company announced that it has set up an 800 number (1-800-ChatGPT, of course) where anyone in the U.S. with a phone line can dial in and speak with the AI via Advanced Voice Mode. Because why not.

“[The goal of] OpenAI is to make artificial general intelligence beneficial to all of humanity, and part of that is making it as accessible as possible to as many people as we can,” the company's chief product officer, Kevin Weil, said during the Wednesday live stream. “Today, we’re taking the next step and bringing ChatGPT to your telephone.”

Read more
OpenAI opens up developer access to the full o1 reasoning model
The openAI o1 logo

On the ninth day of OpenAI's holiday press blitz, the company announced that it is releasing the full version of its o1 reasoning model to select developers through the company's API. Until Tuesday's news, devs could only access the less-capable o1-preview model.

According to the company, the full o1 model will begin rolling out to folks in OpenAI's "Tier 5" developer category. Those are users that have had an account for more than a month and who spend at least $1,000 with the company. The new service is especially pricey for users (on account of the added compute resources o1 requires), costing $15 for every (roughly) 750,000 words analyzed and $60 for every (roughly) 750,000 words generated by the model. That's three to four times the cost of performing the same tasks with GPT-4o.

Read more