Skip to main content

OpenAI’s new AI-made videos are blowing people’s minds

An AI image portraying two mammoths that walk through snow, with mountains and a forest in the background.
OpenAI

OpenAI’s latest venture into AI might be its most impressive one to date. Dubbed “Sora,” this new text-to-video AI model has just opened its doors to a limited number of users who will get to test it. The company launched it by showing several videos made entirely by AI, and the end results are shockingly realistic.

OpenAI introduces Sora by saying that it can create realistic scenes based on text prompts, and the videos shared on its website serve to prove it. The prompts are descriptive, but short; I’ve personally used longer prompts just interacting with ChatGPT. For instance, to generate the video of wooly mammoths pictured above, Sora required a 67-word prompt that described the animals, the surroundings, and the camera placement.

Introducing Sora, our text-to-video model.

Sora can create videos of up to 60 seconds featuring highly detailed scenes, complex camera motion, and multiple characters with vibrant emotions. https://t.co/7j2JN27M3W

Prompt: “Beautiful, snowy… pic.twitter.com/ruTEWn87vf

— OpenAI (@OpenAI) February 15, 2024

“Sora can generate videos up to a minute long while maintaining visual quality and adherence to the user’s prompt,” said OpenAI in its announcement. The AI can generate complex scenes filled with many characters, scenery, and accurate motion. To that end, OpenAI says that Sora predicts and reads between the lines as needed.

“The model understands not only what the user has asked for in the prompt, but also how those things exist in the physical world,” OpenAI said. The model doesn’t just tackle characters, clothing, or backgrounds, but also creates “compelling characters that express vibrant emotions.”

Sora can also fill in the gaps in an existing video or make it longer, as well as generate a video based on an image, so it’s not all just text prompts.

While the videos look good as screenshotted stills, they’re borderline mind-blowing in motion. OpenAI served up a wide range of videos to show off the new tech, including Cyberpunk-esque Tokyo streets and “historical footage” of California during the Gold Rush. There’s more, too, including an extreme close-up of a human eye. The prompts cover anything from cartoons to wildlife photography.

Sora still made some mistakes. Looking closer reveals that, for instance, some figures out in the crowd don’t have heads or move strangely. The awkward motion stood out at first glance in some samples, but the general weirdness took multiple viewings to spot.

It might be a while before OpenAI opens Sora to the general public. Right now, the model will be tested by red teamers who will assess potential risks. Some creators will also get to start testing it now, while it’s still in the early stages of development.

AI is still imperfect, so I went in expecting something quite messy. Whether it’s the low expectations or Sora’s capabilities, I’m walking away impressed, but also mildly worried. We’re already living in a world where it’s hard to tell a fake from something real, and now, it’s not just images that are in jeopardy — videos are, too. However, Sora is hardly the first text-to-video model we’ve seen, such as Pika.

Others are raising the flag as well, such as the popular tech YouTuber, Marques Brownlee, who tweeted that “if this doesn’t concern you at least a little bit, nothing will” in response to the Sora videos.

Every single one of these videos is AI-generated, and if this doesn't concern you at least a little bit, nothing will

The newest model: https://t.co/zkDWU8Be9S

(Remember Will Smith eating spaghetti? I have so many questions) pic.twitter.com/TQ44wvNlQw

— Marques Brownlee (@MKBHD) February 15, 2024

If OpenAI’s Sora is this good now, it’s hard to imagine what it’ll be capable of after a few years of further development and testing. This is the kind of tech that has the potential to displace many jobs — but, hopefully, like ChatGPT, it will instead coexist alongside human professionals.

Editors' Recommendations

Monica J. White
Monica is a UK-based freelance writer and self-proclaimed geek. A firm believer in the "PC building is just like expensive…
Google Gemini vs. GPT-4: Which is the best AI?
A person typing on a laptop that is showing the ChatGPT generative AI website.

Google's Gemini artificial intelligence and OpenAI's ChatGPT that uses the GPT-4 model are two of the most advanced artificial intelligence (AI) solutions available today. They can comprehend and interact with text, images, video, audio, and code, as well as output various alterations of each. they also provide expertise that would cost a lot to replicate with an expert human.

But if you're weighing which tool to put your time and energies into learning how to use, you want to pick the best one. Which is the more capable AI tool? Gemini or GPT-4?
Availability and pricing
Gemini is available in Pro and Nano form, though Ultra has yet to be released. Image used with permission by copyright holder

Read more
OpenAI and Microsoft sued by NY Times for copyright infringement
A phone with the OpenAI logo in front of a large Microsoft logo.

The New York Times has become the first major media organization to take on AI firms in the courts, accusing OpenAI and its backer, Microsoft, of infringing its copyright by using its content to train AI-powered products such as OpenAI's ChatGPT.

In a lawsuit filed in Federal District Court in Manhattan, the media giant claims that “millions” of its copyrighted articles were used to train its AI technologies, enabling it to compete with the New York Times as a content provider.

Read more
Here’s why people are claiming GPT-4 just got way better
A person sits in front of a laptop. On the laptop screen is the home page for OpenAI's ChatGPT artificial intelligence chatbot.

It appears that OpenAI is busy playing cleanup with its GPT language models after accusations that GPT-4 has been getting "lazy," "dumb," and has been experiencing errors outside of the norm for the ChatGPT chatbot circulated social media in late November.

Some are even speculating that GPT-4.5 has secretly been rolled out to some users, based on some responses from ChatGPT itself. Regardless of whether or not that's true, there's definitely been some positive internal changes over the past behind GPT-4.
More GPUs, better performance?
Posts started rolling in as early as last Thursday that noticed the improvement in GPT-4's performance. Wharton Professor Ethan Mollick, who previously commented on the sharp downturn in GPT-4 performance in November, has also noted a revitalization in the model, without seeing any proof of a switch to GPT-4.5 for himself. Consistently using a code interpreter to fix his code, he described the change as "night and day, for both speed and answer quality" after experiencing ChatGPT-4 being "unreliable and a little dull for weeks."

Read more