GPT-5.5 is here! Available in the API, Codex and ChatGPT today

Introducing GPT-5.5

A new class of intelligence for real work and powering agents, built to understand complex goals, use tools, check its work, and carry more tasks through to completion. It marks a new way of getting computer work done.

GPT-5.5 gets to what you are trying to do more quickly and can handle more of the work on its own. It is particularly strong at writing and debugging code, researching online, analyzing data, creating documents and spreadsheets, operating software, and working across tools until the task is complete. Instead of managing every step closely, you can hand GPT-5.5 a messy, multi-part task and rely on it to plan, use tools, verify its work, navigate ambiguity, and keep going.

The improvements stand out most in agentic coding, computer use, knowledge work, and early scientific research, areas where progress depends on reasoning across context and taking action over time. GPT-5.5 delivers that increase in capability without giving up speed. Larger and more capable models are often slower, but GPT-5.5 matches GPT-5.4 on per-token latency in real-world serving while operating at a higher level overall. It also uses significantly fewer tokens to complete the same Codex tasks, which makes it more efficient as well as more capable.

The gains are especially clear in agentic coding, computer use, knowledge work, and early scientific research—areas where progress depends on reasoning across context and taking action over time.

The same qualities that make GPT-5.5 strong at coding also make it more effective for everyday computer-based work. It is better at understanding intent, using tools, checking results, and turning rough input into useful output. In Codex, it outperforms GPT-5.4 on documents, spreadsheets, and slide decks, and gets closer to feeling like a model that can actively use the computer alongside you.

Serving GPT‑5.5 at GPT‑5.4 latency required rethinking inference as an integrated system, not a set of isolated optimizations. Codex and GPT‑5.5 were instrumental in how we achieved our performance targets.[…] Put simply, the model helped improve the infrastructure that serves it.


GPT-5.5 is available in the Responses and Chat Completions APIs with a 1M context window.
GPT-5.5-pro is also available in the Responses API for higher-accuracy work.

14 Likes

Not yet available in the API. Any word on when?

3 Likes

API deployments require different safeguards and we are working closely with partners and customers on the safety and security requirements for serving it at scale. We’ll bring GPT‑5.5 and GPT‑5.5 Pro to the API very soon.

9 Likes
6 Likes

https://openai.com/index/gpt-5-5-system-card/

5 Likes

A straight-up price-doubling on top of a price-doubling between gpt-5.1 to gpt-5.4, on top of a price doubling to use faster yet cheaper to operate inference, reserved for service_tier:priority.

For what they directly say is the same “latency” aka compute time.

And with pricing tied to API, that’s ChatGPT-purchased Codex credits going half as far, on a platform where they give 0-day model shutoffs.

(“Remember, we don’t have enough thinking time for both you and the US Department of War”)

1 Like

Well, now you can work in parallel

which is great because it sounds like you wanted some more doubling in your doubling

^.^

AGI IS HERE BOYS LETSGO

fast mode will go off tho this model is more expensive keep that in mind!

1 Like

O lançamento do GPT-5.5 Codex representa um avanço importante na forma como interagimos com desenvolvimento assistido por IA. No entanto, há um ponto crítico que precisa ser revisto: a limitação de uso.

Atualmente, não faz sentido que, ao atingir o limite do plano, o usuário fique completamente impedido de continuar utilizando até mesmo funções básicas. Esse tipo de interrupção quebra o fluxo de trabalho, especialmente para quem depende da ferramenta de forma contínua.

Uma abordagem mais equilibrada seria adotar um modelo semelhante ao do ChatGPT: quando o limite for atingido, o sistema não deveria simplesmente bloquear o acesso, mas sim reduzir suas capacidades. No caso do Codex, ele poderia continuar operando de forma restrita — por exemplo, limitado ao contexto local do projeto, executando pequenas tarefas, revisando código ou sugerindo ajustes — sem acesso mais amplo ao sistema ou funcionalidades mais avançadas.

Isso garantiria continuidade, produtividade e uma experiência mais consistente para o usuário, ao mesmo tempo em que preserva as limitações do plano contratado.

Em resumo, a proposta não é remover limites, mas torná-los mais inteligentes e menos destrutivos.

INSANE HOW EXPENSIVE IT IS!

i usually dont write in this kind of forums, however i feel the need to say that is crazy HOW 2 PROMPTS!! JUST 2!! (BIG DATABASE THO) BUT JUST 2!!! MAN… THEY BURNED 100 BUCKS IN A MATTER OF AN HOUR… YES! MODEL MIGHT BE BETTER BUT DAM… THIS IS DEEPLY UNSUSTAINABLE AND CRAZY! and no i used not to say anything i was burning through 100 perhaps every two days, thats considerable but not insane considering the output you can generate… BUT 100 BUCKS IN AN HOUR???

1 Like

It makes building apps on top of openai API close to unprofitable. We’re building ai form filler and even 5.2 cost us a lot. I won’t even consider switching to the $30/million model because our margin will go from thin to negative.

It is now available in the API!

6 Likes

I think I need to phone my bank manager …

3 Likes

Fingers crossed for your bank account:

Looking at the price per million tokens, GPT-5.5 appears to be twice as expensive, but efficiency gains may largely offset that.

Based on Artificial Analysis output tokens to run:

  • GPT-5.4 (xhigh): 120M tokens
  • GPT-5.5 (xhigh): 75M tokens
  • GPT-5.5 (high): 45M tokens
  • GPT-5.5 (medium): 22M tokens

5 Likes

Yeah the efficiency gain is a big win! However for basic non-reasoning tasks it must surely make sense to use models like 4.1 still?

I might add I’ve been prototyping in Codex CLI with 5.5 and it absolutely flies through requirements :rocket:

2 Likes

" … so we’re working on behalf of our customers, on behalf of of everyone who wants to use these agents to ensure that there is enough. And I don’t think we’re going to get there. We’re going to do our best, but I think that we are headed to a world of compute scarcity.", Greg Brockman

This is reflected in the price, I suspect.

Testing now: The breadth of document analysis has increased dramatically. It even includes a disclaimer:

This analysis is an issue-spotting review based on the text provided. It is not a legal opinion, and it does not determine whether the Firm actually violated any law in practice.

Translating a document analysis into another language is still accurate - not much change due to the fact that nuance is unwanted.

2 Likes

The python SDK is not updated currently right?

Man I just took a second mortgage to build a website :joy::joy::joy:

2 Likes