|
Prompt Token Cache Gaming to Save Money?
|
|
2
|
945
|
February 12, 2026
|
|
Caching is borked for GPT 5 models
|
|
19
|
2268
|
January 8, 2026
|
|
Understanding "prompt_cache_keys" in query efficiency
|
|
5
|
866
|
November 12, 2025
|
|
We need to talk about prompt caching
|
|
1
|
498
|
October 25, 2025
|
|
Can I cache large chunks on gpt-5-nano?, Does each cache-read request reset cache inactive time?, Does large caches affect cache overflow limits?
|
|
1
|
107
|
October 25, 2025
|
|
Prompt Caching seems not working even if long common prefix in the system prompt
|
|
3
|
442
|
September 23, 2025
|
|
Following Instructions Quality - Developer message and its position, instructions and prompt caching
|
|
1
|
165
|
September 16, 2025
|
|
Prompt caching with tools
|
|
1
|
516
|
September 15, 2025
|
|
New Realtime API voices and cache pricing
|
|
27
|
12469
|
September 2, 2025
|
|
Does structured output schema come BEFORE/AFTER system message for prompt caching?
|
|
3
|
237
|
August 3, 2025
|
|
How to use cached_tokens field to calculate cost estimation
|
|
1
|
690
|
July 31, 2025
|
|
Prompt Cache Routing + the `user` Parameter
|
|
3
|
595
|
July 31, 2025
|
|
Prompt caching doesn't seem to work regularly
|
|
4
|
809
|
July 13, 2025
|
|
How to improve caching accuracy
|
|
1
|
336
|
July 8, 2025
|
|
System prompt not regard when using web search in OpenAI – Why?
|
|
3
|
606
|
July 6, 2025
|
|
Prompt caching - how many prompts are cached?
|
|
2
|
208
|
June 20, 2025
|
|
4o input not being cached
|
|
42
|
1880
|
April 25, 2025
|
|
Is there a way to disable prompt caching in the APIs
|
|
9
|
7578
|
April 24, 2025
|
|
Responses API not using cached inputs for o3-mini
|
|
0
|
120
|
April 17, 2025
|
|
Automatic context window caching - better performance? When?
|
|
0
|
198
|
April 3, 2025
|
|
Prompt caching enabled in O3-Mini?
|
|
5
|
342
|
March 31, 2025
|
|
Does prompt caching reduce TPM?
|
|
4
|
539
|
March 9, 2025
|
|
How Prompt caching works?
|
|
17
|
9722
|
February 4, 2025
|
|
Dashboard usage vs Prompt response usage not matching
|
|
13
|
807
|
January 9, 2025
|
|
Understanding Prompt caching
|
|
0
|
410
|
January 2, 2025
|
|
Does prompt caching persist between different models?
|
|
1
|
287
|
December 23, 2024
|
|
Why don't we have prompt caching on gpt-4?
|
|
1
|
185
|
November 22, 2024
|
|
Cache not caching more than 1024 tokens (expected: increments of 128 tokens)
|
|
6
|
382
|
November 14, 2024
|
|
Gpt-4o-2024-08-06 randomly fails to cache tokens
|
|
7
|
255
|
November 12, 2024
|
|
Improving Cache Management: Handling Tool Removal in Active Conversations
|
|
0
|
59
|
November 11, 2024
|