Home
Sora OpenAI Finally Obeys the Laws of Physics
Sora OpenAI Finally Obeys the Laws of Physics
Sora 2 is no longer just a technical curiosity hidden behind research whitepapers. It has officially transitioned from the "GPT-1 moment" of video generation—where we were impressed that objects simply existed—to what is effectively the "GPT-3.5 moment." The leap from the original Sora announcement to the current Sora 2 model represents a fundamental shift in how artificial intelligence perceives and replicates the physical world. While the first iteration was a master of aesthetics, Sora 2 is becoming a master of kinetics.
Testing the new Sora 2 environment feels different because the hallucinations have changed. In the early days of generative video, we accepted that a person walking through a door might emerge as a different person, or that a cup of coffee might merge into the table it sat upon. Sora 2 has largely corrected these "logic errors." When you watch a video generated by this model today, the persistence of objects and the adherence to gravity feel intentional rather than accidental.
The End of Teleporting Objects
One of the most frustrating aspects of early video models was their "optimistic" nature. If a model was prompted to show a basketball player missing a shot, it would often force the ball to teleport into the hoop because it understood the concept of "scoring" better than the concept of "missing." In our extensive testing of Sora 2, we’ve seen a radical departure from this behavior.
If you prompt Sora 2 with a scene of a basketball hitting the backboard, the ball now rebounds with a trajectory that feels weighted and realistic. It doesn't just bounce; it obeys the laws of buoyancy and rigidity. We ran a series of prompts involving a guy doing a backflip on a paddleboard. In previous versions, the board would remain static, or the water would react like liquid mercury. In Sora 2, the displacement of the water and the slight dip of the board under the athlete’s weight align with real-world fluid dynamics. This is what OpenAI refers to as "world simulation," and it's the primary reason this model feels like a step toward AGI.
However, it isn't perfect. We noticed that in complex scenes involving multiple interacting agents—like a chaotic kitchen scene with five chefs—the model can still struggle with occlusion. A spatula might disappear behind a pot and emerge as a spoon. But these instances are becoming the exception rather than the rule. The model is implicitly modeling an internal agent that understands what should happen next based on physics, not just what pixel should come next based on probability.
Synchronized Audio: The Silent Era is Over
For the past year, the AI video space has been strangely silent. We were used to generating beautiful visuals and then spending hours in post-production trying to find or generate matching sound effects. Sora 2 eliminates this friction. It is a general-purpose video-audio generation system that creates sophisticated background soundscapes, speech, and sound effects with high fidelity.
In one test, we prompted: "Two mountain explorers in bright technical shells, ice-crusted faces, eyes narrowed with urgency shout in the snow, one at a time." The result was jarring in its realism. Not only were the facial expressions synchronized with the vocalizations, but the crunch of the snow under their boots and the howling wind in the background were spatially mapped to the visual scene. The audio isn't just an overlay; it feels like it was recorded on-site. This synchronization of dialogue and foley is perhaps the most underrated update in the Sora OpenAI ecosystem, as it suddenly makes the tool viable for rapid prototyping in filmmaking without needing a secondary audio pipeline.
The Sora App and the "Cameos" Revolution
The most significant shift in strategy is the release of the standalone Sora iOS app. OpenAI is clearly moving away from being a mere API provider and toward becoming a social platform. The app isn't just a place to type prompts; it’s a community-driven feed where you can remix others' generations and, most importantly, insert yourself into the AI world via "Cameos."
We spent several days with the Cameos feature, and it is arguably the most addictive part of the new ecosystem. After a one-time video and audio recording to verify identity and capture likeness, the model can drop you into any environment. Want to see yourself as a Viking launching a longship in the North Sea? Or as a futuristic explorer on a neon-lit cyberpunk street? The fidelity is remarkable. Unlike traditional deepfakes, which often look like a mask layered over a body, Sora 2 reconstructs your likeness within the lighting and texture of the generated scene. If the scene has a cool winter daylight, your face reflects that specific light.
The social aspect is built around this. You can send "Sora messages" where you are the protagonist of an impossible scenario. The app's feed philosophy, according to OpenAI, is not optimized for "time spent" or doomscrolling, but for inspiration. By default, the feed shows content from people you follow, prioritizing videos that you might want to use as a base for your own remixes. It feels less like TikTok and more like a collaborative studio.
Technical Underpinnings: Patches vs. Tokens
To understand why Sora 2 works, you have to look at its architecture. While Large Language Models (LLMs) use tokens, Sora uses "visual patches." Think of these as the visual equivalent of words. The model compresses videos into a lower-dimensional latent space and then decomposes them into spacetime patches.
This approach allows Sora 2 to be highly scalable. By training on diverse datasets—including publicly available internet data and proprietary partnerships with companies like Shutterstock—the model has learned the "grammar" of the physical world. It uses the recaptioning technique pioneered in DALL·E 3, which means it translates highly descriptive captions into complex visual sequences. This is why it follows intricate instructions spanning multiple shots while accurately persisting the world state. If a character is wearing a specific torn jacket in shot one, that tear remains in exactly the same place in shot five, even if the camera angle has shifted 180 degrees.
Safety and the Provenance Problem
With great power comes the inevitable risk of misuse. OpenAI has been transparent about the "Safety Stack" implemented in Sora 2. Every video generated contains C2PA metadata, a verifiable industry standard that proves the content's origin. Additionally, there is a visible moving watermark on all downloads from the app.
During our testing, we found the safety filters to be quite stringent. The model uses multi-modal moderation classifiers that scan not just the input prompt, but also the output video frames, audio transcripts, and even the scene descriptions. If you try to generate a video involving a photorealistic person without their consent—or if the prompt triggers any of the disallowed content categories (violence, hate symbols, explicit material)—the system blocks it before generation is even complete.
For minors, the protections are even tighter. Users under 18 have default limits on how many generations they can see and restricted permissions on the Cameos feature. There are also parental controls integrated via ChatGPT that allow parents to manage direct message settings and turn off algorithmic personalization. This focus on "launching responsibly" is a clear attempt to avoid the pitfalls that have plagued other social platforms.
The Experience of a "World Simulator"
One of our favorite prompts during the review period was: "Bigfoot is really kind to him, a little too kind, like oddly kind. Bigfoot wants to hang out but he wants to hang too much." This prompt tests a model’s ability to understand subtle human (or cryptid) emotions and social awkwardness.
Sora 2 didn't just generate a furry monster. It generated a creature with expressive eyes and body language that conveyed "clinging needy friend." The human character in the video showed visible discomfort that felt genuine. This is the difference between a video generator and a world simulator. A world simulator understands context, subtext, and the consequences of actions.
When we talk about Sora OpenAI moving toward AGI, we are talking about this ability to simulate reality so accurately that the AI can eventually be used to train other models—like those for robotics—to function in the physical world. If an AI can understand the complex dynamics of a backflip or the subtle social cues of an awkward conversation, it is one step closer to understanding the human experience.
The Reality Check: Limitations and Costs
Despite the glowing praise, we must address the friction points. The computational cost of Sora 2 is clearly immense. While the app is currently invite-only, the monetization plan involves paying for extra generations once a user exceeds their base limit. For high-resolution, 1080p videos that last up to 20 seconds, the wait time can still be several minutes depending on server load. This isn't "instant" creation yet.
There is also the issue of the "internal agent" making mistakes. Sometimes the physics are too perfect, leading to a look that feels sterile or "over-simulated." And while the likeness protection is robust within the app, the broader world is still grappling with how to handle AI-generated media that circulates without metadata. OpenAI’s internal detection tools are a good start, but they only work for content generated by their own products.
Final Thoughts on the New Era
Sora 2 is a validation that scaling neural networks on video data works. We are entering a completely new era for co-creative experiences. The ability to remix a video with the same ease that we currently quote-tweet a post is going to fundamentally change the creator economy.
For filmmakers, it’s a tool for previz and storytelling. For creators, it’s a way to bring themselves into impossible worlds. For the rest of us, it’s a new way to communicate that goes beyond text, emojis, and voice notes. The social iOS app is a bold bet that the future of social media isn't just watching videos, but living in them. Sora OpenAI has set a high bar, and while the competition is fierce, the "world simulator" approach seems to be the most promising path forward for the industry.