O3 is 80% cheaper and introducing o3-pro

We’re cutting the price of o3 by 80% and introducing o3-pro in the API, which uses even more compute.

o3:
Input: $2 / 1M tokens
Output: $8 / 1M tokens

Now in effect.

We optimized our inference stack that serves o3. Same exact model—just cheaper.

o3-pro:
Input: $20 / 1M tokens
Output: $80 / 1M tokens

(87% cheaper than o1-pro!)

We recommend using background mode with o3-pro: long-running tasks will be kicked off asynchronously, preventing timeouts. https://platform.openai.com/docs/guides/background

20 Likes

Hi @edwinarbus,

Thanks for the update on the o3 model pricing — the 80% reduction is a welcome change.

However, I wanted to flag a potential billing issue I encountered. I ran a batch using o3-2025-04-16 shortly after (June 10, 2025 at 23:03 UTC) the new pricing was officially announced as “now in effect” (June 10, 2025 at 21:18 UTC), but I was still charged the old rates — $5/million for input and $20/million for output — totaling $150.00 for the batch.

I tried to contact the support team, but I was only able to reach the automated Operator and couldn’t find a way to submit a formal request about the issue.

Regards,
Kutub

1 Like

Hm! If you send a message at the bottom of https://help.openai.com/en/?q=contact while logged in, and create a case with our Support team, they’ll be able to take a look into your account. (Unfortunately in this forum we’re not able to investigate individual account issues here.)

2 Likes

Thank you for you help, but I am following the guide from https://help.openai.com/en/articles/6614161-how-can-i-contact-support and just reaching Operator again and again. There is something I am missing?

EDIT: Operator said it escalated to Support Team. Thanks.

The input token consumption of vision does not match between o3 and o3 pro. Token usage is higher, besides the higher price.

67x80 image

checkerboard_67x80

Sending nothing other than the image:

O3-Pro: 262 tokens
O3: 232 tokens

The extra 10 tokens x3 shows that there is 85 token image billing per base tile like other mainline models, not the 75 of o3 (or even 65 base + 129/tile of computer-use-preview or gpt-image-1)

(previous o3 model pricing shown)

It does not follow the placement of o3 or o1-pro in the vision pricing table:

Is this intentional or an oversight? @edwinarbus

Pricing calculator here by OpenAI says 75.

5 Likes

Will Codex (cloud app) use o3-Pro now?

Wow, fantastic news. Do we have the same pricing on azure hosted instances?

Here https://help.openai.com/en/articles/9624314-model-release-notes and here https://x.com/OpenAI/status/1932530423911096508 yesterday you announced that o3-pro is available in the model picker for Pro and Team users starting today, replacing OpenAI o1-pro.

I just paid a Team subscription €68, just to discover that I don’t have 03-pro in the list. Please refund.

2 Likes

Currently o3-pro is so slow it’s almost unusable in both ChatGPT and the API.

A simple, dumb request like this took 15ms in ChatGPT desktop app and 19ms in browser for a response.

This shouldn’t be normal right?


I think you mean “minutes”, not “ms”.

A simple input of one image to o3 pro and a small token billing was an extensive wait.

I suspect that “no flex processing discount” is because that is already being done by “inference efficiencies” also coming to this model: if you’re willing to wait for a long response, you’ll also be waiting behind other jobs transparently for fitting the API call into a queue…

I think it is rolling out.
This is to compare:

https://openai.com/chatgpt/pricing/

1 Like

What is the usage limit for o3-pro for ChatGPT Teams accounts?

The help page doesn’t say.

Also, does Deep Research now use o3-pro (or did it always)?

No, Codex uses its own model, codex-1.

4 Likes

Really rather you would have increased the context length instead, but okay.

Edit-Update: I just checked my team account and o3-pro is available. Treat the information below with a grain of salt.


I did confirm the issue and checked in with staff: The new model should be available within a week for all Team users.

The best course of action is to contact support at help.openai.com for your refund.

Idk, hope this helps!

2 Likes

Cool, have they increased the weekly limit for o3?


hahahhahah love it

One seeming benefit of o3-pro seen - at least you aren’t the one paying for hundreds of tokens of unseen decision and moderation as in other reasoning models.

O3-Pro

36 output tokens billed on 26 tokens received.

Still with apparent 85 tokens for vision base tile instead of 75 tokens.

O1-Pro

193 more output tokens billed than received:

Note the peculiar vision input billing of o1-pro, also seen in o1. An image 512x512 would be 1 tile (75 or 85 tokens) says the pricing guide. Here however, a detail:low image is always min/max 22 tokens with container overhead, and detail:high as showing is 41 tokens with its text. 512x513 is a jump to 63, 22 tokens more input. Perhaps a price break because of the stratospheric cost otherwise? At the very least, o1 is undisclosed and unpublished vision pricing formula.


Adding images has added latency of around 1-3 seconds across all other models. So still with these 15 second response times, there’s either a queue, there is unseen moderations or decisions before your billed task…or OpenAI figured out how to publish a model with 3 token-per-second generation rate.

yeah, and two days later we have to let you face scan us to use o3 in API.
No way!