Share via

Azure Open AI Realtime GA Model not working

Christopher Rajan 0 Reputation points
2026-03-23T06:04:49.4633333+00:00

I have developed a conversational AI agent using Azure OpenAI Realtime Preview model which is getting depracted in April. All the audio functions were working properly. When I moved this to Realtime GA model when ever I start speaking its terminating. All the parameters are passed as per the API specification documentation. Do anyone know why is this happening and are there any specific things I need to correct. Or should I file a Azure ticket since the issue is with the Realtime GA Model.

Azure AI Speech
Azure AI Speech

An Azure service that integrates speech processing into apps and services.


1 answer

Sort by: Most helpful
  1. Q&A Assist
    2026-03-23T06:05:07.1333333+00:00
    AI generated content. This question contains an answer created with AI from Q&A Assist. Learn more

    The behavior is consistent with known migration and protocol differences between the Realtime Preview and Realtime GA APIs. Several GA-specific requirements can cause the session to terminate as soon as audio starts, even if parameters appear correct.

    Key areas to verify and correct before opening a support ticket:

    1. Use the GA endpoint format
    • GA no longer accepts date-based api-version or the api-version query parameter.
    • WebSocket endpoint must be:
        wss://<resource>.openai.azure.com/openai/v1/realtime
      
    • If the connection is still using something like:
        wss://<resource>.openai.azure.com/openai/realtime?api-version=2025-04-01-preview
      
      the service can fail or behave unexpectedly. Remove api-version and ensure /openai/v1 is present.
    1. Update protocol and event names for GA If a custom client was built against the Preview protocol, several event names and message formats changed and must be updated. Using old event names can cause the server to close the session when audio starts.

    Important GA changes:

    • session.update now requires a type field:
      • "realtime" for speech-to-speech
      • "transcription" for realtime audio transcription Example:
        {
          "type": "session.update",
          "session": {
            "type": "realtime",
            "model": "gpt-realtime"
            // other config
          }
        }
      
      If type is missing, the server can reject the configuration and terminate.
    • Event name changes:
      • response.text.deltaresponse.output_text.delta
      • response.audio.deltaresponse.output_audio.delta
      • response.audio_transcript.deltaresponse.output_audio_transcript.delta
    • Assistant message content types changed:
      • type: "text"type: "output_text"
      • type: "audio"type: "output_audio"
    • All conversation item events now include object: "realtime.item".

    Ensure all event handlers, message builders, and parsing logic are updated to these GA names and shapes. If the client still sends or expects Preview event types, the conversation can appear to “terminate” when audio is sent.

    1. Remove Preview-only headers
    • Do not send the OpenAI-Beta header in any GA requests. Keeping this header from Preview code can cause protocol issues.
    1. Audio format and transport checks Even if this worked in Preview, confirm that the audio being sent still matches GA expectations:
    • PCM 16-bit (pcm16)
    • Mono
    • 24 kHz sample rate
    • For JSON transport, audio chunks must be base64-encoded and reasonably small (around 100 ms per chunk). Oversized or malformed chunks can cause the server to close the connection.
    1. SDK and client implementation
    • GA protocol is only supported in the official OpenAI SDKs and in custom clients that fully implement the GA protocol.
    • For .NET, GA requires OpenAI .NET SDK version 2.9.0 or later. Earlier versions do not support the GA Realtime protocol and can fail when audio starts.
    1. When to open an Azure support ticket After verifying all of the following:
    • Endpoint uses /openai/v1 and no api-version query parameter.
    • No OpenAI-Beta header is sent.
    • session.update includes session.type set correctly ("realtime" or "transcription").
    • All event names and content types are updated to GA (response.output_* and output_* content types).
    • Audio format and chunking match the documented requirements.
    • A supported SDK version or a correctly updated custom client is used.

    If the session still terminates immediately when speaking, collect:

    • Exact endpoint URL (with resource and path, redacting keys)
    • Region and deployment name
    • Timestamps and correlation IDs if available
    • Example session.update payload
    • A small example of the audio payload

    Then open an Azure support ticket, as this may indicate a service-side issue with the Realtime GA model.


    References:

    0 comments No comments

Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.