[BUG] Azure Open AI Assistant API Streaming Suddenly Returning 400 Errors #48502
Labels
Client
This issue points to a problem in the data-plane of the library.
customer-reported
Issues that are reported by GitHub users external to the Azure organization.
needs-team-attention
Workflow: This issue needs attention from Azure service team or SDK team
OpenAI
question
The issue doesn't require a change to the product in order to be resolved. Most issues start as that
Service Attention
Workflow: This issue is responsible by Azure service team.
Library name and version
Azure.AI.OpenAI 2.2.0-beta.2
Describe the bug
On 2/26 around 8:22 am, the Azure Open AI Assistant API stopped allowing streaming responses from an Azure deployed Open AI gpt-4o model. Here is the error:
HTTP 400 (invalid_request_error: unsupported_model) Parameter: model Unsupported value: 'stream' does not support 'true' with this model. Supported values are: 'false'.
Model Deployment details
oia-oiaoiaoia-dev-eastus2-001-global-standard-gpt4o
(NOTE: this seems important to replicate the issue)The deployed model was previously working for quite some time. I tried deploying a new gpt-4o model and had the same issue. No code or packages have changed our end that could explain the breakage. We did try upgrading the nuget package (from
Azure.AI.OpenAI 2.2.0-beta.1
toAzure.AI.OpenAI 2.2.0-beta.2
) to see if that would resolve the issue. It did not.I was able to work around this issue by deploying a gpt-4o-mini, but obviously that has other implications. I tried some of the newly available o models as well, which as far as I can tell should support streaming. Those had the same issue as gpt-4o. Here's an example request I captured using a proxy:
Response:
I can share a curl request that consistently reproduces the issue as well as the redacted bits above with Azure support personnel.
I have since been able to deploy
gpt-4o
models that are streaming successfully.It might be a model name length issue but have seen some newly deployed models (same name length/pattern, deployment type, model version, model name, and content filter) have the issue while another did not. Even looking at the network traffic, the requests are almost identical between the working model and not working model.Edit: This does appear to be related to the name of the model. Creating a brand new gpt-4o model with:
gpt4o-ncs
- can streambai-baibaibai-dev-eastus2-001-global-standard-gpt4o
- can streamoia-oiaoiaoia-dev-eastus2-001-global-standard-gpt4o
- can NOT streamoai-oaioaioai-dev-eastus2-001-global-standard-gpt4o
- can NOT streamI tried creating a support ticket via Azure but was unsuccessful.
Expected behavior
Streaming is still supported using the Assistant API using an Azure deployed GPT-4o model.
Actual behavior
The following error is returned, along with an
HTTP 400
Reproduction Steps
Issue the following curl request (again, happy to share the redacted bits):
Environment
Model Deployment details
Model: gpt-4o
Model version: 2024-05-13 (though others as well)
Region: eastus2
Deployment Type: Global Standard
The text was updated successfully, but these errors were encountered: