now:
HttpResponseError: Operation returned an invalid status 'Internal Server Error' Content: Model service is unavailable.
This browser is no longer supported.
Upgrade to Microsoft Edge to take advantage of the latest features, security updates, and technical support.
DeepSeek-V3 and DeepSeek-V3-0324 model stop working on non-streaming request.
Simple use the example shows timeout. Using "stream=True" then works.
import os
from azure.ai.inference import ChatCompletionsClient
from azure.ai.inference.models import SystemMessage, UserMessage
from azure.core.credentials import AzureKeyCredential
endpoint = "https://XXXXXX.services.ai.azure.com/models"
model_name = "DeepSeek-V3-0324"
client = ChatCompletionsClient(
endpoint=endpoint,
credential=AzureKeyCredential("<API_KEY>"),
)
response = client.complete(
messages=[
SystemMessage(content="You are a helpful assistant."),
UserMessage(content="I am going to Paris, what should I see?")
],
max_tokens=2048,
temperature=0.8,
top_p=0.1,
presence_penalty=0.0,
frequency_penalty=0.0,
model=model_name
)
print(response.choices[0].message.content)
now:
HttpResponseError: Operation returned an invalid status 'Internal Server Error' Content: Model service is unavailable.
Hello @Xiwei Zhou ,
I am converting the comment to answer so that it will help the community to find better solution.
Issue:
The DeepSeek-V3
and DeepSeek-V3-0324
models were returning timeouts or internal server errors for non-streaming (stream=False
) requests. However, streaming requests (stream=True
) continued to work as expected.
Solution:
This issue appears to be transient and related to temporary backend unavailability. If you encounter similar errors with non-streaming requests (such as timeouts or internal server errors), please try the following steps:
apim-request-id
from the response headers and share it with Azure support for further investigation.Please do accept
the solution and give feedback by clicking on yes
.
Thank you