OpenAI unveils powerful, creepy new text-to-video generator that it calls 'a foundation for models that can understand and simulate the real world'
Sora is another step on the road to "artificial general intelligence," the ChatGPT company says.
Keep up to date with the most important stories and the best deals, as picked by the PC Gamer team.
You are now subscribed
Your newsletter sign-up was successful
Want to add more newsletters?
Every Friday
GamesRadar+
Your weekly update on everything you could ever want to know about the games you already love, games we know you're going to love in the near future, and tales from the communities that surround them.
Every Thursday
GTA 6 O'clock
Our special GTA 6 newsletter, with breaking news, insider info, and rumor analysis from the award-winning GTA 6 O'clock experts.
Every Friday
Knowledge
From the creators of Edge: A weekly videogame industry newsletter with analysis from expert writers, guidance from professionals, and insight into what's on the horizon.
Every Thursday
The Setup
Hardware nerds unite, sign up to our free tech newsletter for a weekly digest of the hottest new tech, the latest gadgets on the test bench, and much more.
Every Wednesday
Switch 2 Spotlight
Sign up to our new Switch 2 newsletter, where we bring you the latest talking points on Nintendo's new console each week, bring you up to date on the news, and recommend what games to play.
Every Saturday
The Watchlist
Subscribe for a weekly digest of the movie and TV news that matters, direct to your inbox. From first-look trailers, interviews, reviews and explainers, we've got you covered.
Once a month
SFX
Get sneak previews, exclusive competitions and details of special events each month!
The generative AI company behind ChatGPT and DALL-E has a new toy: Sora, a text-to-video model that can (sometimes) generate pretty convincing 60-second clips from prompts like "a stylish woman walks down a Tokyo street..." and "a movie trailer featuring the adventures of the 30 year old space man wearing a red wool knitted motorcycle helmet..."
A lot of the AI video generation we've seen so far fails to sustain a consistent reality, redesigning faces and clothing and objects from one frame to the next. Sora, however, "understands not only what the user has asked for in the prompt, but also how those things exist in the physical world," says OpenAI in its announcement post (using the word "understands" loosely).
View post on imgur.com"
The Sora clips are impressive. If I weren't looking closely—say, I was just scrolling past them on social media—I'd probably think many of them were real. The prompt "a Chinese Lunar New Year celebration video with Chinese Dragon" looks at first like typical documentary footage of a parade. But then you realize that the people are oddly proportioned, and seem to be stumbling—it's like the moment in a dream when you suddenly notice that everything is a little bit wrong. Creepy.
"The current model has weaknesses," writes OpenAI. "It may struggle with accurately simulating the physics of a complex scene, and may not understand specific instances of cause and effect. For example, a person might take a bite out of a cookie, but afterward, the cookie may not have a bite mark. The model may also confuse spatial details of a prompt, for example, mixing up left and right, and may struggle with precise descriptions of events that take place over time, like following a specific camera trajectory."
My favorite demonstration of Sora's weaknesses is a video in which a plastic chair begins morphing into a Cronenberg lifeform. Behold:
View post on imgur.com"
Sora is not widely available yet, and OpenAI says it's assessing social risks of the model and working on mitigating them, for instance with "a detection classifier that can tell when a video was generated by Sora."
It's fascinating as a research project, but OpenAI isn't just interested in doing cool computer science. If it can outmaneuver copyright critics and legislators, it's here to make bank. The company says it's currently "granting [Sora] access to a number of visual artists, designers, and filmmakers to gain feedback on how to advance the model to be most helpful for creative professionals."
Keep up to date with the most important stories and the best deals, as picked by the PC Gamer team.
One commenter on X optimistically wondered if models like Sora will one day allow the public to wrest control of filmmaking away from Hollywood by making movies purely with prompts—but I wonder where the source material for all this generated video will come from if not, you know, filmmakers? Big Hollywood movies may already look pretty homogenous, but auto-reproducing Marvel Cinematic Universe-style CGI and car commercial drone shots isn't exactly bringing creative expression to the masses. (The blog post notably doesn't mention Sora's training material.)
View post on imgur.com"
Despite the often clumsy results of current generative AI models and the legal, ethical quagmire it presents, we're already seeing it used in professional creative media. That includes videogames, both in ways that are directly visible to us, like to generate art and voices and on-the-fly dialogue, and in ways that are less obvious, like generating code snippets or early concept art. A recent survey found that 31% of game development professionals use generative AI in some capacity. Combined with other software, I wonder what this kind of machine learning-driven video simulation could do besides generate slightly-off CG-like clips?
I don't think anyone really knows how generative AI will be used in five or ten years or what the consequences of continued development will be, but it isn't slowing down, so it appears we'll find out. OpenAI and other companies are explicitly working not just toward better image and video and text generators, but toward "artificial general intelligence" or AGI—as in, the science fiction idea of what AI is.
"Sora serves as a foundation for models that can understand and simulate the real world, a capability we believe will be an important milestone for achieving AGI," says OpenAI.

Tyler grew up in Silicon Valley during the '80s and '90s, playing games like Zork and Arkanoid on early PCs. He was later captivated by Myst, SimCity, Civilization, Command & Conquer, all the shooters they call "boomer shooters" now, and PS1 classic Bushido Blade (that's right: he had Bleem!). Tyler joined PC Gamer in 2011, and today he's focused on the site's news coverage. His hobbies include amateur boxing and adding to his 1,200-plus hours in Rocket League.

