Batch API degraded since March 4 — stuck at 0 progress, expiring after partial completion

We’ve been experiencing Batch API degradation since March 4, 2026. The issue started with abnormally slow processing and has escalated to batches making zero progress.

Timeline:

  1. March 4, 12:43 UTC — Batch batch_69a828e39a4081908428434be28a1df6 (60 requests) completed successfully. Last known good batch.
  2. March 4, 13:45 UTC — Batch batch_69a83760285c81909fb8f58a466b4309 (102 requests) started processing noticeably slower than usual. Only completed 50/102 requests before expiring after the 24-hour window. No errors reported — just ran out of time.
  3. March 5, 14:15 UTC — Resubmitted as batch batch_69a98fe7451c81908a36e4f41e902944 (102 requests). Sat at 0/102 with no progress. Manually cancelled.
  4. March 5, 19:25 UTC — Resubmitted again as batch batch_69a9d8b5a29881909e073eef0d4f68b5 (60 requests, smaller payload). Currently stuck at 0/60 completed, 0 failed. No errors, no progress.

Pattern:

  • Started as slow processing (March 4), now escalated to zero progress (March 5)
  • No error messages returned on any batch — they either silently expire or sit indefinitely at 0
  • Smaller batches (60 requests) completed fine on March 4 but are now also stuck
  • Using /v1/responses endpoint with structured output

Environment:

  • Endpoint: /v1/responses
  • Request sizes: 60–102 requests per batch
  • JSONL format validated, same format that completed successfully on March 3–4

Is anyone else seeing similar degradation since March 4? Is there a known issue with batch queue processing?

3 Likes

I have the same issue happening on my end with no progress in batch API with gpt-5-mini.
But gpt 4.1 mini is working, may be you may consider that model.

1 Like

Yeah, that pattern definitely looks concerning. Going from slow processing to zero progress without errors isn’t ideal.

I checked on our side and there aren’t any reported outages with the Batch API right now. I also ran a small test batch using the same /v1/responses flow, and it completed successfully without errors. So it doesn’t look like a system-wide failure at the moment.

A couple things that would help narrow this down:

  • Which model are you using for these batches?
  • Are the stuck batches still showing in_progress, or are they in another state?
  • Has your most recent resubmission made any progress since posting?

If it continues to sit at 0 with no movement, this may need a deeper look with internal logs. In that case, I’d recommend reaching out to support@openai.com and including the batch IDs you shared here so the team can investigate directly.

Let us know what model you’re running and whether anything’s changed on the latest batch.

2 Likes

@OpenAI_Support Also having batch processing issues. I’ve been stuck in_progress for around 6 hours. Just noticed this morning and hadn’t ran any jobs since last week. Using gpt-5-mini-2025-08-07 with around 300 jobs. Have canceled and restarted many times.

1 Like

Trying the same jobs but with gpt-5 finished in a few minutes while my gpt-5-mini jobs are still in_progress. Hope this info helps.

1 Like

@OpenAI_Support Could you please provide any updates on this issue? We are still encountering the same problem. Processing a very small batch file (around 150KB) is taking more than two hours, and in some cases even longer.

1 Like

Thanks for looking into this. Here are the details:

Model: gpt-5-mini-2025-08-07
Endpoint: /v1/responses with structured output (Zod schemas)

Current state of the latest batch:
batch_69a9d8b5a29881909e073eef0d4f68b5 — still in_progress, now at 50/60 completed (was stuck at 0/60 for hours after submission at 19:25 UTC yesterday). So it is moving, but extremely slowly.

Context — we have solid prior experience with this exact setup:

On March 3, we ran a full pipeline with the same model, same endpoint, same JSONL format, same request sizes — everything completed successfully and quickly:

  • batch_69a6bd4f83908190b69451d4ac1dadd5 — 20 requests → completed :white_check_mark:
  • batch_69a6bf3297888190b6e78c98d15310d1 — 29 requests → completed :white_check_mark:
  • batch_69a6d1e2c9908190b7561fbb3ae9bdca — 60 requests → completed :white_check_mark:
  • batch_69a6d31fa46481908bdeeac1d335d20e — 102 requests → completed :white_check_mark:
  • batch_69a6e387a6d88190ba567dd409ee5ee7 — 9 requests → completed (8/9, 1 expected failure) :white_check_mark:

All of these finished within reasonable time frames. No issues whatsoever.

Then starting March 4, same payload, same setup — degradation:

  1. batch_69a828e39a4081908428434be28a1df6 — 60 requests → completed :white_check_mark: (this was the header phase, still fine)
  2. batch_69a83760285c81909fb8f58a466b4309 — 102 requests → expired :cross_mark: (only 50/102 completed before 24h timeout — processing was noticeably slower than March 3 when the identical 102-request batch completed fine)
  3. batch_69a98fe7451c81908a36e4f41e902944 — 102 requests → sat at 0/102 with no progress, manually cancelled :cross_mark:
  4. batch_69a9d8b5a29881909e073eef0d4f68b5 — 60 requests → currently in_progress, 50/60 after ~12+ hours :cross_mark: (on March 3, a 60-request batch completed much faster)

Summary: Nothing changed on our side between March 3 and March 4 — same model, same endpoint, same JSONL format, same structured output schemas, same request sizes. The only difference is processing speed on your end. March 3 was fast and reliable, March 4 onward has been slow to stuck.

Thanks for the advice. I think I should try if the problems continue. But I am a bit concerned about the quality of results I will receive. I have prerry precise task with structured output.

1 Like

I can also confirm this issue with my batches. I also use gpt5-mini, and on March 4 it almost stopped working. Before that day, my batches were processed in 5-10 minutes. One of my overdue batches: batch_69a95f5ba3188190b12055b007c34cbe

1 Like

We are also facing the same problem and we are also using gpt-5-mini model and it stopped from march 4.

1 Like

I’m having the same issue with GPT‑5‑mini. Previously, batches processed quickly, but now even 2 of them haven’t completed in 24 hours.

@OpenAI_Support do you have any updates about this ongoing problem? Looking at the number of comments, it doesn’t look like it is a personal problem.

I am having similar issues. I started a batch of 1709 running around 7am ET last night. It sat at completed=0 until about 4am ET, then completed 1439, and has been slowly creeping up to 1688 (current progress) since then. I’ve been using the batch API for months and have never had a run go this long.

I also just started another batch of 58, and that has been at completed=0 for about half an hour, whereas usually I could expect this batch to finish in 15 minutes or less.

@OpenAI_Support, can we get someone to look into this on OpenAI’s side?

First batch: batch_69aa1efbcf4881909c550bcaa1af12f6

Second batch: batch_69ab1c9cc96c81908ea34e740f87e924

@OpenAI_Support I completed a batch processing of 122 jobs this morning on gpt-5-mini-2025-08-07. The issue appears resolved.

2 Likes

It looks like this was fixed for me as well. 20 minutes for a 102 big parsing batch.

2 Likes

Thanks for the update @Dmytro_Boiko @jason.jmp !

I will put this topic on auto-close assuming that any new issues in the future will have a different cause.

And thanks to everyone for raising this!

2 Likes

I’m having the same issues, the difference is that I’m using gpt 5 nano for this batch api. Even cancelling the batch takes some time (it hasn’t even finished cancelling at the moment). Any ideas?

2 Likes

I’m still having issues on my side, gpt-5-nano is stuck at 0 for houuuurs whereas few weeks ago it was treated in a few minutes. :frowning: :frowning:

1 Like

Second this, batch gpt-5-nano is being unusable, it renders 1000 rows in 12 hrs, and I need to process millions, this so slow I could do it manually. On the other hand standard service runs within api limits, no problem to get 500 jobs done per minute, so considering the price it’s still acceptable. But batch looks like being intentionally slowed down, every 2nd job stops at 99x/1000 for hours but keeps 24h completion window so nobody can officially complain.

2 Likes

Hey everyone, Issue is not reproducing in current testing: a /v1/responses batch progressed and completed normally.

There have also been backend improvements in this area from engineering. If anyone is still seeing batches stuck at 0 progress or expiring after partial completion, please share a recent batch ID and/or x-request-id so we can trace that specific execution path and investigate further.

2 Likes