OpenAI's New Real-Time Audio Models Are Changing How Voice AI Works

OpenAI launched three real-time audio models for developers: GPT-Realtime-2, GPT-Realtime-Translate, and GPT-Realtime-Whisper.

A colorful mosaic tunnel wall is shown.

OpenAI's New Real-Time Audio Models Are Changing How Voice AI Works

Voice AI has been stuck in an awkward middle ground for years. You could talk to a chatbot and get a response, but the experience felt transactional at best. You had to pause, wait for it to catch up, and repeat yourself whenever it lost the thread of what you were asking.

OpenAI is taking a direct swing at that problem. On May 7, 2026, the company released three new audio models through its developer API: GPT-Realtime-2, GPT-Realtime-Translate, and GPT-Realtime-Whisper. The launch marks a shift from voice as a feature layered on top of a text model to voice as a first-class, task-capable interface in its own right.

What OpenAI Actually Released

The three models are distinct tools built for different parts of the voice pipeline, and understanding each one separately is worth the time. They are not incremental upgrades to existing transcription tools. Each one targets a specific gap in how voice AI has historically fallen short.

GPT-Realtime-2 is the flagship of the three. OpenAI describes it as the first voice model with GPT-5-class reasoning, which is a meaningful distinction from what came before. Previous voice models could respond to commands and hold short conversations, but they struggled when things got complicated: multi-step requests, interruptions mid-sentence, or context that needed to carry across a longer session. GPT-Realtime-2 is designed to handle all of that, and it introduces parallel tool calling so the model can pull from multiple external sources simultaneously while keeping the user informed throughout. The context window expanded from 32,000 to 128,000 tokens, and developers can set the reasoning effort level anywhere from minimal to extra-high depending on what the use case demands.

The benchmark results reflect that improvement. GPT-Realtime-2 scored 96.6% on Big Bench Audio with high reasoning enabled, compared to 81.4% for GPT-Realtime-1.5. On the Audio MultiChallenge instruction following benchmark, the score jumped from 34.7% to 48.5% at extra-high reasoning. Those numbers come directly from OpenAI's published model documentation.

GPT-Realtime-Translate is built for live multilingual voice experiences. It handles translation from more than 70 input languages into 13 output languages in real time, while keeping pace with how people actually speak, including regional pronunciations, mid-sentence context shifts, and domain-specific vocabulary. Customer support, education, and cross-language professional settings are the obvious targets for this model. OpenAI noted that Deutsche Telekom and Vimeo are already testing it, and BolnaAI has been working with it specifically for India-focused voice applications.

GPT-Realtime-Whisper handles the transcription layer of the pipeline. Unlike traditional speech-to-text systems that process audio after a speaker finishes a sentence, this model transcribes continuously as someone speaks, which means live captions, real-time meeting notes, and in-app assistants can all update instantly rather than in delayed chunks. The use cases span customer support, healthcare documentation, recruiting, and any setting where lag between speech and text creates friction.

Why the Tool-Calling Piece Matters Most

The voice AI space has had capable technology for a while, but what it has consistently lacked is voice AI that can act on what it hears. The ability to call external tools mid-conversation is the part of this announcement with the most direct impact for anyone building production voice applications. It closes the gap between a voice assistant that can respond and one that can actually complete work.

Consider what that enables in practice. A voice agent handling a customer billing issue can pull account records, check payment history, apply a credit, and confirm the resolution, all while the conversation is still happening. That is a categorically different product than a voice chatbot that reads from a knowledge base and escalates when things get complicated. Developers building in healthcare, financial services, and customer operations have been waiting for this capability in a voice-native form.

The tone control and adjustable reasoning features in GPT-Realtime-2 are also worth paying attention to for anyone thinking through deployment contexts. A voice agent handling intake calls for a medical practice operates in a different register than one supporting a SaaS help desk, and the ability to configure that at the model level gives developers more control than trying to manage it entirely through prompting and scripting.

Pricing and Availability

All three models are available now through OpenAI's Realtime API, and developers can test them in the OpenAI Playground before building toward production. GPT-Realtime-2 is priced at $32 per million audio input tokens and $64 per million audio output tokens, with cached input tokens available at $0.40 per million. GPT-Realtime-Translate runs $0.034 per minute, and GPT-Realtime-Whisper is priced at $0.017 per minute.

For high-volume applications like large-scale customer support operations or telehealth platforms, those costs will factor meaningfully into the unit economics of a product. For lower-volume or higher-value interactions, the pricing is workable and in line with what the capabilities are worth. OpenAI also confirmed the API includes active content classifiers and supports EU data residency requirements, which matters for teams building in regulated industries or serving European markets.

Where This Goes From Here

OpenAI noted that the ChatGPT consumer voice experience is still running on older models and that upgrades to the public-facing product are coming separately. The Realtime API launch is aimed at developers for now, which means the consumer version of these capabilities is still a step behind what builders can access today.

The broader direction is hard to miss. The gap between what you can build in a voice interface and what you can build in a text interface is closing, and it is closing quickly. Parallel tool calls, extended context, live multilingual translation, and streaming transcription were all either technically impossible or deeply impractical in voice applications until recently. They are now available through a single API at pricing that makes production deployment viable. For teams building AI-powered products where voice is a real channel and not just a demo feature, the infrastructure has gotten significantly more serious.

Start building with agents in minutes

Start building with agents in minutes

Start building with agents in minutes

Start building with agents in minutes