I was intrigued and impressed when OpenAI first demonstrated Sora earlier this year. There seemed no limit to the films the AI video model could produce out of a text prompt. Sora could effortlessly transform descriptions into immersive, realistic videos, and OpenAI coyly hinted at a general rollout in the near future. Months later, only professional filmmakers partnered with OpenAI have any real access (a recent, brief leak doesn't count). The same goes for other much-hyped AI video generators, including Meta’s Movie Gen and Google's Veo.
Many great AI video generators have emerged since Sora blew people away, but it's hard not to feel like a kid with their nose pressed up against the glass of the toy store, wondering why we can't play with the toys just a little bit. Here's why I think OpenAI and the rest of the reticent AI video creation models are still locked away.
Movie trailers always lead to disappointment
Maybe I'm just a skeptic, but I find it odd how OpenAI, Meta, and Google all seemingly couldn't wait to show off demos of their respective AI video generators without even a vague sense of a rollout date. It makes me think of movie teaser trailers that come out a year before a film and promise far more than the final cut can deliver. I wonder if Sora, Movie Gen, and Veo might have more than a little cooking left to do before we get our hands on them.
The meticulously curated demos might not only be the best examples of the AI models but also the only ones worth showing to the public. Sora's standard output might be more fever dream than celestial vision. Perhaps asking for a "serene sunset over a lake” only occasionally nets a tranquil evening on the water. If nine out of ten Sora clips depict a lake melting into a neon green abyss under a sun flickering like a haunted strobe light, I wouldn't blame OpenAI for holding Sora back for now.
Ethics (or legal exposure)
The companies behind AI tools for making images and videos usually make a point of highlighting their ethical training and output controls where they can. Sora is no exception, but the lines of ethical model limits get a lot blurrier for videos compared to images, especially since the video is essentially a huge number of images strung together.
Unapproved data scraping to make deepfakes of real people without their knowledge and producing films with trademarked characters and logos without permission open the gates of a vast legal and ethical minefield. Working with professional filmmakers and commercial video directors eliminates those problems because the tech company can closely watch the AI's output and prevent casual infringement.
Where's the profit?
As much as OpenAI, Adobe, Google, and Meta might like showing off their technology, the people controlling the purse strings want to know where the return on this investment comes from and when. The goal is a polished, marketable AI video generator, not a cool toy. A free-for-all AI video playground to experiment and make mistakes is a step in the path, not the destination.
While we don't know the exact cost, it's possible high-end AI videomakers are just as expensive to run compared to Runway or Dream Machine. The processing power required is certainly staggering compared to AI text composition. Scaling it up without restrictions might cause a server meltdown. Allowing bored students to make short clips of a dog playing the violin in a submarine may not seem worth the expense of running Sora around the clock for millions of users. Limiting access to approved professionals gives the companies more control.
OpenAI is almost certainly working on strategies for making money from hobbyists, smaller marketing firms, and film studios willing to pay for ongoing access to advanced AI video generators like Sora. But until they are as accessible as the subscriptions for premium versions of ChatGPT, Gemini, and other AI chatbots, only the deepest of filmmaking pockets will likely get access to Sora and its sister models. Until then, we're just spectators.