Content
summary Summary

OpenAI's new text-to-video model, Sora, will likely remain in development for some time before a public release.

According to Bloomberg, OpenAI has not yet set an exact release schedule. There are two reasons for this: One is that OpenAI does not want to take any safety risks, given the number of elections this year. The second reason is that the model is not yet technically ready for release.

When OpenAI unveiled Sora, the company pointed out shortcomings in the model's physical understanding and consistency. Bloomberg's tests with two OpenAI-generated prompts confirmed these issues. For example, in the video below, the parrot turns into a monkey at the end.

Prompt: "An overhead view of a parrot flying through a verdant Costa Rica jungle, then landing on a tree branch to eat a piece of fruit with a group of monkeys. Golden hour, 35mm film." | Video: Rachel Metz via X

Ad
Ad

"Sora remains a research project, very few people have access to it, and there were clear limitations for the tool during our private demo with OpenAI," Bloomberg writes.

It's worth noting that Bloomberg gave OpenAI four prompts, but OpenAI only generated two videos. The company cited time constraints on the part of its researchers. This suggests that the generation process is lengthy and/or immature.

Generate and snack

OpenAI has not given any details about Sora's speed. But it's going to take longer than standard image generators. For now, you can "definitely" grab a snack while waiting for the finished video, says OpenAI researcher Bill Peebles. Of course, this could change before the release.

Peebles also acknowledges Sora's generation errors, but still calls the system a "significant leap" in AI video generation, especially in terms of scene complexity.

Sora is currently in the red teaming phase, and selected artists, filmmakers, and designers have been given access to the system. This was announced by OpenAI when the model was unveiled.

Recommendation

Looking at OpenAI's model rollout plan, this seems to be the "Evaluation & Iterative Development" phase. The next step would then be a staged rollout phase with private betas, use case pilots, and further safety testing.

Image: OpenAI

The project also caught the attention of US filmmaker Tyler Perry, who said he canceled an $800 million expansion of his studio after seeing Sora demos. Perry believes that in the future he will be able to generate movie scenes solely from text.

Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • OpenAI's Sora text-to-video model is still under development and has no set release schedule. The model is not yet technically mature, and OpenAI wants to minimize safety risks due to the large number of elections this year.
  • Sora has shortcomings in physical understanding and consistency, and the generation process is still slow. Still, OpenAI researcher Bill Peebles calls it a "significant leap" in video generation.
  • The model is currently in the "evaluation and iterative development" phase, in which selected artists, filmmakers, and designers have been given access to the system; the next step would be a staggered rollout phase with private betas, use-case pilots, and further safety testing.
Sources
Online journalist Matthias is the co-founder and publisher of THE DECODER. He believes that artificial intelligence will fundamentally change the relationship between humans and computers.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.