Note
Access to this page requires authorization. You can try signing in or changing directories.
Access to this page requires authorization. You can try changing directories.
Note
This document refers to the Microsoft Foundry (classic) portal.
🔄 Switch to the Microsoft Foundry (new) documentation if you're using the new portal.
Note
This document refers to the Microsoft Foundry (new) portal.
This article describes how to manage traffic with spillover for provisioned deployments in Azure OpenAI. Spillover manages traffic fluctuations by routing overage traffic to a corresponding standard deployment. This optional capability can be set for all requests on a deployment or managed on a per-request basis, helping you reduce disruptions during traffic bursts.
Prerequisites
An Azure subscription. Create one for free.
A provisioned managed deployment and a standard deployment in the same Azure OpenAI resource
The data processing level of your standard deployment must match your provisioned deployment. For example, use a global provisioned deployment with a global standard spillover deployment.
Azure CLI installed for REST API examples, or access to the Foundry portal
The
AZURE_OPENAI_ENDPOINTenvironment variable set to your Azure OpenAI endpoint URLCognitive Services Contributor role or higher on the Azure OpenAI resource to create or modify deployments
Enable spillover for all requests on a provisioned deployment
To deploy a model with the spillover capability, go to the Foundry portal. On the left navigation menu, then select Deployments.
Select Deploy model. In the menu that appears, select Customize.
Specify one of the provisioned options as the Deployment type, for example Global Provisioned Throughput. Select Traffic spillover to enable spillover for your provisioned deployment.
Tip
- To enable spillover, your account must have at least one active pay-as-you-go deployment that matches the model and version of your current provisioned deployment.
- To see how to enable spillover for select inference requests, select the REST API tab above.
See the REST API tab to learn how to enable spillover.
When to enable spillover
To maximize the utilization of your provisioned deployment, enable spillover for all global and data zone provisioned deployments. With spillover, bursts or fluctuations in traffic can be automatically managed by the service. This capability reduces the risk of experiencing disruptions when a provisioned deployment is fully utilized. Alternatively, spillover is configurable per-request to provide flexibility across different scenarios and workloads. Spillover also works with the Foundry Agent Service.
When spillover comes into effect
When you enable spillover for a deployment or configure it for a given inference request, spillover initiates when a specific non-200 response code is received as a result of one of these scenarios:
Provisioned throughput units (PTU) are completely used, which results in a
429response code.You send a long context token request, which results in a
400error code. For example, when you usegpt 4.1series models, PTU supports only context lengths less than 128K and returns HTTP 400.Server errors occur when processing your request, which results in error code
500or503.
When a request results in one of these non-200 response codes, Azure OpenAI automatically sends the request from your provisioned deployment to your standard deployment to be processed.
Note
Even if a subset of requests is routed to the standard deployment, the service prioritizes sending requests to the provisioned deployment before sending any overage requests to the standard deployment. This prioritization might incur additional latency.
Identify spillover requests
The following HTTP response headers indicate that a specific request spilled over:
x-ms-spillover-from-<deployment-name>. This header contains the PTU deployment name. The presence of this header indicates that the request is a spillover request.x-ms-<deployment-name>. This header contains the name of the deployment that serves the request. If the request spills over, the deployment name is the name of the standard deployment.
For a request that spills over, if the standard deployment request fails for any reason, the original PTU response is used in the response to the customer. The customer sees a header x-ms-spillover-error that contains the response code of the spillover request (such as 429 or 500) so that they know the reason for the failed spillover.
Monitor spillover usage
Since the spillover capability relies on a combination of provisioned and standard deployments to manage traffic overages, monitoring can be conducted at the deployment level for each deployment. To view how many requests were processed on the primary provisioned deployment versus the spillover standard deployment, apply the splitting feature in Azure Monitor metrics to view the requests processed by each deployment and their respective status codes. Similarly, use the splitting feature to view how many tokens were processed on the primary provisioned deployment versus the spillover standard deployment for a given time period.
For more information on observability within Azure OpenAI, review the Monitor Azure OpenAI documentation.
Spillover cost
Because spillover uses a combination of provisioned and standard deployments to manage traffic fluctuations, billing for spillover involves two components:
For any requests processed by your provisioned deployment, only the hourly provisioned deployment cost applies. No additional costs are incurred for these requests.
For any requests routed to your standard deployment, the request is billed at the associated input token, cached token, and output token rates for the specified model version and deployment type.
Monitor metrics in the Azure portal
The following Azure Monitor metrics chart provides an example of the split of requests between the primary provisioned deployment and the spillover standard deployment when spillover is initiated. To create a chart, navigate to your resource in the Azure portal.
Select Monitoring > Metrics from the left navigation menu.
Add the
Azure OpenAI Requestsmetric.Select Apply splitting and apply the
ModelDeploymentNamesplit andStatusCodesplits to theAzure OpenAI Requestsmetric. This shows a chart with the200(success) and429(too many requests) response codes generated for your resource.Be sure to add the model deployments you want to view when applying the
ModelDeploymentNamesplit.The following example shows an instance where a spike in requests sent to the provisioned throughput deployment generates
429error codes. Shortly after, spillover occurs and requests begin going to the pay-as-you-go deployment used for spillover, generating200responses for that deployment.
View spillover metrics
Applying the IsSpillover split lets you view the requests to your deployment that are being redirected to your spillover deployment. Following from the previous example, you can see how the 429 responses from the primary deployment match the 200 response codes generated by the spillover deployment.