At its Dev Day on Monday, OpenAI announced new API updates, including GPT-5 Pro, its latest language model, Sora 2, its new video creation model, and a smaller, less expensive voice model.
The upgrades were part of a series of announcements aimed at attracting developers to OpenAI’s ecosystem, which also included the debut of an agent-building tool and the option to create apps in ChatGPT.
The addition of GPT-5 Pro may appeal to developers working on applications in finance, legal, and healthcare fields that require “high accuracy and depth of reasoning,” according to OpenAI CEO Sam Altman.
Altman also stated that speech skills will be critical in the future, as they quickly become one of the key ways people interact with AI. To that purpose, OpenAI is introducing “gpt-realtime mini,” a smaller, less expensive voice model in the API that supports low-latency streaming interactions for audio and speech.
The new model is 70% cheaper than OpenAI’s previous sophisticated voice model while promising the “same voice quality and expressiveness.
Finally, creators within OpenAI’s developer ecosystem can now access Sora 2 in preview via the API. Last week, OpenAI unveiled Sora 2, its latest music and video generator, coupled with the Sora app, a TikTok competitor that features brief AI-generated videos. The Sora app enables users to create films of themselves, friends, or anyone else based on a prompt and distribute them through a TikTok-style algorithmic feed.
“[Developers] now have access to the same model that powers Sora 2’s stunning video outputs right in your own app,” according to Altman.
Sora 2 improves on its predecessor with more realistic, physically consistent scenes, synchronised sound, and increased creative options, ranging from detailed camera direction to stylised images.
“For example, you can take the iPhone view and prompt Sora to expand it into a sweeping, cinematic wide shot,” Altman pointed out. “But one of the most exciting things that we’ve been working on is how well this new model pairs sound with visuals, not just speech, but rich soundscapes, ambient audio, synchronised effects that are grounded in what you’re seeing.”
Sora 2 is positioned as a tool for concept development, whether it’s a visual starting point for an ad based on a product’s overall vibe or a Mattel designer transforming a sketch into a toy concept — an example Altman provided at Dev Day that sheds light on OpenAI’s agreement with the Barbie-maker to bring generative AI into the toy-making pipeline.