OpenAI's new Sora 2 model pushes AI video closer to the mainstream, adding more realistic physics, better control, and, for the first time, high-quality audio. The launch also includes a Sora iOS app built for sharing AI-generated videos with friends.
Inside OpenAI, Sora 2 is seen as a serious upgrade. The original model was their "GPT-1 moment" for video - an early proof of concept with clear limits. Sora 2, on the other hand, is what the team calls the "GPT-3.5 moment" for generative video, hitting the point where the tech feels usable, much like the jump language models made a few years ago.
There’s still no official word on technical specs like resolution or maximum video length, but sample clips look like they run at 720p and 30 FPS, lasting about five to ten seconds.
Sora 2’s big leap is in simulating complex physical processes. OpenAI says the model can handle things like paddleboard backflips with realistic buoyancy or gymnastic stunts that actually look right.
Earlier video models often glitched or warped moving objects. Sora 2, by comparison, can show a basketball bouncing off the backboard if it misses, which OpenAI points to as evidence of better physical reasoning. In the long run, OpenAI sees Sora 2 as a step toward general world simulators for physically interactive AI.
Control, Consistency, and Sound
Sora 2 can follow complex, multi-step instructions across several scenes without losing track of what's happening, according to OpenAI. It handles a range of visual styles - from photorealistic to cinematic to anime - and now, for the first time, generates believable background noise, speech, and sound effects. Like Google's Veo 3, Sora 2 aims to keep visuals and audio in sync.
A new feature lets users put themselves into the videos they generate. By recording their voice and appearance once, people can make "cameos" that show up in any scene, with their look and voice carried over. Animals and objects can be added too. The demo video includes a cameo from OpenAI CEO Sam Altman.
OpenAI says users always control their own cameos. Only people you authorize can use your cameo, and you can see every video - including drafts - where your cameo appears. You can revoke access or delete your cameo whenever you want.
Extra protections are in place for minors, including stricter controls, reduced visibility, and default safety limits. Deepfakes of public figures are technically possible, but OpenAI says these are blocked unless the person opts in.
Sora App with Social Features
Sora 2 is rolling out through a new iOS app where users can create videos, remix other people's content, and browse a personalized feed. The app is launching by invitation in the US and Canada, with plans to expand to more countries soon. At first, Sora 2 will be free to use, with what OpenAI calls "generous limits."
The feed highlights videos from people you interact with and clips that have strong remix potential. Recommendations are powered by OpenAI's language models and can be tweaked using text prompts.
OpenAI says they're "not optimizing for time spent in feed," and built the app to encourage creation over mindless scrolling. The company even published its own "Feed philosophy", promising to stick to these principles as the app evolves. Meta is also working on a similar AI-generated video feed.
Sora.com still requires an invite code to access the upgraded Sora 2 Pro model with "higher quality" videos, and OpenAI says an API is on the way. More demos can be seen in the Sora 2 announcement livestream.