OpenAI is pushing deeper into voice. The company just launched three new realtime audio models in its API. GPT-Realtime-2 for conversational reasoning, GPT-Realtime-Translate for live multilingual translation, and GPT-Realtime-Whisper for streaming speech transcription. GPT-Realtime-2 can now handle longer conversations, recover from interruptions more naturally, use tools while someone is still talking, and respond with different reasoning levels depending on the task. OpenAI says the model is designed for things like customer support, scheduling, travel assistance, and other workflows where the AI actually has to keep track of context instead of just replying quickly. OpenAI is no longer treating voice as a side feature attached to chatbots. It’s starting to position voice as the interface itself. That means live translation during conversations. Real time transcription while meetings are still happening. AI agents that can check your calendar, pull information from apps, or complete actions while the conversation keeps moving.