Prompt caching enabled in O3-Mini?

Hi folks.

I’m a bit confused. On the prompt caching page, there is no mentioned of o3-mini, however I have seen a post in forum which confirms that prompt caching is enabled for o3-mini.

Further more, I’m using Response SDK in javascript (v4.88.0) and I m consistently seeing cache hits as 0 tokens.

I confirmed from Open AI API logs and diff checker tool that first ~3000 tokens are exactly same between 2 runs.

I think prompt caching is currently having issues even with models that are explicitly mentioned on that page (e.g. gpt-4o-mini): 4o input not being cached - #40 by bento

Hi all! Thanks for reporting. Working with the docs team to clarify this now and will let you know when we ship a change 👀

3 Likes

Thanks. But let me know if caching is support in o3-mini or not please?

The models documentation explicitly lists a price for context window cache:

Thus, one must assume that it is supported and advertised as such, and broad failure to deliver on activating discounts on compliant repetitive inputs should be considered an overbilling while such pricing is being shown.

2 Likes

Yes, caching is supported in o3-mini! Thanks for jumping in to answer this, @_j.

2 Likes