Share via

Azure OpenAI Realtime client_secrets API returns OperationNotSupported for valid deployment

Pranit Awasthi 0 Reputation points
2026-03-19T03:45:20.2133333+00:00

I am trying to generate a WebRTC ephemeral token using the Azure OpenAI Realtime API:

POST /openai/v1/realtime/client_secrets

The request is made with a valid Azure AD bearer token using DefaultAzureCredential, and the endpoint, region, and resource configuration are correct.

Request payload:

{ "session": { "type": "realtime", "model": "gpt-realtime-mini", "instructions": "You are a helpful assistant." } }

However, the API returns the following error:

{ "error": { "code": "OpperationNotSupported", "message": "The realtime operation does not work with the specified model." } }

Additional details:

  • The deployment exists and the name is correct

The request is sent to the GA endpoint: /openai/v1/realtime/client_secrets

No preview headers or api-version parameters are used

Authentication is successful and the bearer token is valid

Questions:

Does the deployment need to be created specifically from a Realtime-capable model family (e.g., gpt-realtime or gpt-realtime-mini)?

Are there additional configuration steps required to enable Realtime support on a deployment?

Could this error indicate that the deployment is based on a model that does not support the Realtime API, even if the deployment name is valid?

Any clarification on the requirements for using the Realtime client_secrets endpoint would be helpful.I am trying to generate a WebRTC ephemeral token using the Azure OpenAI Realtime API:

POST /openai/v1/realtime/client_secrets

The request is made with a valid Azure AD bearer token using DefaultAzureCredential, and the endpoint, region, and resource configuration are correct.

Request payload:

{
"session": {
"type": "realtime",
"model": "gpt-realtime-mini",
"instructions": "You are a helpful assistant."
}
}

However, the API returns the following error:

{
"error": {
"code": "OpperationNotSupported",
"message": "The realtime operation does not work with the specified model."
}
}

Additional details:

The deployment exists and the name is correct

The request is sent to the GA endpoint: /openai/v1/realtime/client_secrets

No preview headers or api-version parameters are used

Authentication is successful and the bearer token is valid

Questions:

Does the deployment need to be created specifically from a Realtime-capable model family (e.g., gpt-realtime or gpt-realtime-mini)?

Are there additional configuration steps required to enable Realtime support on a deployment?

Could this error indicate that the deployment is based on a model that does not support the Realtime API, even if the deployment name is valid?

Any clarification on the requirements for using the Realtime client_secrets endpoint would be helpful.

Azure OpenAI Service
Azure OpenAI Service

An Azure service that provides access to OpenAI’s GPT-3 models with enterprise capabilities.

{count} votes

2 answers

Sort by: Most helpful
  1. Karnam Venkata Rajeswari 565 Reputation points Microsoft External Staff Moderator
    2026-03-25T18:01:56.5433333+00:00

    Hello Pranit Awasthi,

    Welcome to Microsoft Q&A and Thank you for reaching out.

    The POST /openai/v1/realtime/client_secrets call is reaching the service successfully (authentication is valid), but the service is rejecting the request because the deployment being targeted is not recognized as a Realtime-capable deployment for the Realtime operation. The error text - realtime operation does not work with the specified model This is most commonly encountered when

    1. The deployment is not actually a GPT Realtime model deployment (even if the deployment name exists and looks correct). Realtime endpoints only work with the GPT Realtime model families listed as supported for Realtime.
    2. A non-supported region/resource is being used for Realtime, because Realtime model availability is region-dependent. The Realtime WebRTC doc lists supported regions for global deployments as East US 2 and Sweden Central.
    3. The request is passing a model name when the endpoint expects a deployment name .If the deployment name is different from gpt-realtime-mini, the call can be routed incorrectly and result in operation/model mismatch errors.

    As asked if the deployment needs to be created specifically from a Realtime-capable model family (for example, gpt-realtime / gpt-realtime-mini) - yes. Only the Realtime API supports specific GPT Realtime models/versions.

    As asked if there any additional configuration steps required to enable Realtime support on a deployment - No special “enable” toggle is required beyond meeting the documented prerequisites

    No special “enable” toggle is required beyond meeting the documented prerequisites. The documented prerequisites are:

    • A resource created in a supported region, and
    • A deployment of a GPT Realtime model in that supported region, and
    • Using the GA endpoint format with /openai/v1 in the URL

    For WebRTC ephemeral tokens specifically, please use the GA client secrets endpoint (…/openai/v1/realtime/client_secrets) as the mechanism to get the ephemeral token

    As asked if this error indicates that the deployment is based on a model that does not support the Realtime API, even if the deployment name is valid - Yes. This error is consistent with the Realtime operation being invoked against a deployment that is not a supported Realtime model deployment or is deployed in a non-supported region for Realtime.

    Please note that for Azure OpenAI, model should refer to the deployment name chosen during deployment. If the deployment name is not exactly gpt-realtime-mini, set "model" to the deployment name instead.

    Please consider the following troubleshooting steps

    1. Confirm the deployed base model is actually a GPT Realtime model
    2. Confirm the resource region is supported for Realtime
    3. Use the deployment name in the request payload
    4. Keep GA endpoint format exactly as documented

     

    References:

    Thank you!

     

    Please 'Upvote'(Thumbs-up) and 'Accept' as answer if the reply was helpful. This will be benefitting other community members who face the same issue.

    0 comments No comments

  2. Q&A Assist
    2026-03-19T03:45:40.48+00:00
    AI generated content. This question contains an answer created with AI from Q&A Assist. Learn more

    Yes, the deployment must be created from a Realtime‑capable model family, and the error strongly indicates the deployment model does not support the Realtime API.

    Key points based on the current Realtime GA behavior:

    1. Realtime requires a Realtime model deployment
      • The client_secrets endpoint is only supported for deployments created from Realtime models:
        • gpt-4o-realtime-preview
        • gpt-4o-mini-realtime-preview
        • gpt-realtime
        • gpt-realtime-mini
        • gpt-realtime-mini (newer versions)
        • gpt-realtime-1.5
      • The session.model value in the client_secrets request must be the deployment name of one of these Realtime models, not a deployment of a standard chat/completions model.
      • If the deployment is based on a non‑Realtime model (for example, a regular gpt-4o or other chat model), the Realtime operations are not supported and an error like OperationNotSupported is expected.
    2. No extra “Realtime enablement” beyond using a Realtime model
      • There are no additional flags or configuration steps to “turn on” Realtime for a deployment.
      • The only requirement is that the deployment itself is created from a Realtime‑capable base model listed above and is in a supported region.
    3. What the error most likely means in this scenario
      • Given:
        • POST /openai/v1/realtime/client_secrets is used (GA endpoint format).
        • Authentication succeeds.
        • The deployment name exists and resolves.
      • The error message "The realtime operation does not work with the specified model." is consistent with:
        • The deployment being based on a non‑Realtime model family, even though the deployment name is valid.
    4. Practical checks
      • Confirm in Azure AI Foundry / Azure OpenAI that the deployment is explicitly created from one of the Realtime models listed above.
      • Ensure the GA endpoint format is used (no api-version in the URL) and that the deployment is in a supported region for Realtime.

    Once the deployment is recreated using a Realtime model (for example, a deployment of gpt-realtime-mini), the same client_secrets call with session.type = "realtime" and session.model = "<that deployment name>" should be accepted.


    References:

    0 comments No comments

Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.