Azure OpenAI Deployed Model Suddenly throw 400 for Assistant API Streaming

Yesterday morning (around 8:22am ET) an Azure deployed model consistently started returning 400 errors for the Assistant API streaming responses. The error:

HTTP 400 (invalid_request_error: unsupported_model) Parameter: model Unsupported value: 'stream' does not support 'true' with this model. Supported values are: 'false'.

Deployment details

  • gpt-4o
  • Model version 2024-05-13 (though others as well)
  • Deployed in Azure eastus2 using global standard

The deployed model was previously working for quite some time. I tried deploying a new gpt-4o model and had the same issue.

No code or packages have changed our end that could explain the breakage. We did try upgrading the nuget package to see if that would resolve the issue. It did not.

I was able to work around this issue by deploying a gpt-4o-mini, but obviously that has other implications. I tried the o models as well, which as far as I can tell should support streaming. Those had the same issue as gpt-4o.

Any suggestions on troubleshooting? I’m working to get a contact from Azure to assist.