OpenAI unveils audio-native models bringing GPT-5 level reasoning to real-time voice interactions
New audio-native models bring GPT-5 level reasoning to live conversation, featuring adjustable intensity for solving complex, real-time problems.
May 7, 2026

The barrier between high-level cognitive reasoning and instantaneous voice interaction has significantly narrowed as OpenAI introduces a new suite of audio-native models designed to handle complex, real-time tasks.[1] In a major expansion of its developer ecosystem, the company has released three specialized models—GPT-Realtime-2, GPT-Realtime-Translate, and GPT-Realtime-Whisper—that collectively represent a shift away from traditional voice-to-text-to-voice pipelines.[2] The centerpiece of this release is GPT-Realtime-2, a model OpenAI claims possesses reasoning capabilities on par with its most advanced text-based system, GPT-5.[2] This development marks the first time that sophisticated reasoning and large-scale context have been successfully integrated into a low-latency audio environment, potentially transforming how humans interact with digital agents across industries ranging from real estate to global logistics.
The technical foundation of GPT-Realtime-2 centers on a dramatic increase in both memory and processing flexibility. Unlike its predecessors, which often struggled to maintain coherence over long sessions, the new model features a context window of 128,000 tokens. This expansion allows the AI to recall and reference details from hours of continuous conversation, making it suitable for deep technical support or long-form collaborative planning. Furthermore, OpenAI has introduced a granular control mechanism called "reasoning intensity," which allows developers to toggle between five distinct levels: minimal, low, medium, high, and xhigh.[1] This architectural choice addresses the inherent trade-off between intelligence and speed.[3] While the "low" setting is optimized for rapid-fire banter and simple information retrieval, the "xhigh" setting enables the model to pause and engage in complex problem-solving, such as debugging code or analyzing financial data, before speaking its conclusion.
To manage the human-centric nature of live conversation, GPT-Realtime-2 utilizes a series of "stalling tricks" to bridge the gap between processing time and audible response. When the model is set to higher reasoning levels, it is programmed to use natural-sounding filler phrases like "let me check that" or "I am looking into that right now" to maintain the flow of dialogue while it calculates a response. This behavioral layer is combined with an improved ability to handle interruptions and corrections.[2][4] The model can now detect when a user has cut it off mid-sentence, process the new input instantly, and pivot its reasoning accordingly without the jarring restarts typical of earlier voice assistants. This fluidity is further enhanced by parallel tool calling, which allows the model to perform multiple backend actions—such as checking a database, sending an email, and updating a calendar—simultaneously while still engaged in the verbal exchange.
Alongside the primary reasoning model, OpenAI is shipping GPT-Realtime-Translate and GPT-Realtime-Whisper to provide specialized infrastructure for multilingual and administrative workflows. The translation model supports more than 70 input languages and can output speech in 13 languages with minimal delay, aiming to preserve the speaker's original tone and emphasis.[2][1] This is a significant departure from standard machine translation, as the model is designed to keep pace with natural speech rates, handling regional accents and specialized terminology in real time. Meanwhile, GPT-Realtime-Whisper provides a low-latency streaming transcription service focused on generating immediate text records for meetings and live events.[4] Together, these models are priced to encourage broad adoption: GPT-Realtime-2 costs $32 per million audio input tokens and $64 per million output tokens, while the Translate and Whisper models are billed at $0.034 and $0.017 per minute, respectively.[4]
The immediate application of these models is already visible through early partnerships with major consumer and enterprise brands. Zillow is utilizing the reasoning power of GPT-Realtime-2 to build a voice-based real estate assistant capable of identifying potential properties based on nuanced user preferences and autonomously scheduling home tours by interacting with scheduling APIs. Priceline is also integrating the technology to facilitate comprehensive trip management, where travelers can modify flights, hotels, and dinner reservations through a single, continuous voice conversation that requires the model to reason through complex logistics and availability. Other early adopters, including Deutsche Telekom and Glean, are leveraging the models to create more empathetic and capable customer service agents that can handle frustrated users with tailored emotional tones—switching from calm problem-solving to an upbeat demeanor once a resolution is reached.
This move toward audio-native reasoning has profound implications for the AI industry, as it challenges the dominance of traditional text-first interfaces. By eliminating the need to transcribe audio into text before processing it, OpenAI reduces the loss of information that occurs when nuances like sarcasm, urgency, or hesitation are stripped away. This "speech-to-speech" architecture ensures that the reasoning engine perceives the full spectrum of human communication. It also places immense pressure on competitors like Google and Anthropic to accelerate their own multimodal releases. The transition from "voice assistants" that follow simple commands to "voice agents" that can reason, plan, and act represents a fundamental evolution in human-computer interaction, moving the technology closer to the long-promised vision of a seamless digital collaborator.
In the broader context of artificial intelligence development, the integration of GPT-5-level reasoning into real-time voice signals a maturation of the technology from a novelty into a utility. The ability to dial reasoning effort up or down gives developers a sophisticated toolkit to manage the high compute costs associated with large-scale models, while the expanded context window ensures that these agents can become deeply integrated into a user’s long-term workflow. As organizations begin to deploy these models in production, the focus will likely shift from basic performance metrics to the ethical and operational challenges of managing agents that sound and think with human-like complexity. By providing a platform where reasoning and conversation are no longer separate functions, OpenAI is setting a new standard for the next generation of interactive intelligence.