Why are Deep Research API token counts so high?

I’m using Deep Research by API with websearch tool. It counts way more Input tokens than it should be. Even with small input prompt, I get couple millions of tokens usage. In the screenshots you can see, that I have too much cached and noncached input tokens usage which I never used. And as a consequence, it burned a lot of money in my OpenAI account. It looks like bug in Deep Research api.
Customer support is ignoring me. Please help!

Welcome to the OpenAI dev community, @Rustemov_Bauyrzhan.

The deep research models are agentic models, which per the Deep Research docs:

The o3-deep-research and o4-mini-deep-research models can find, analyze, and synthesize hundreds of sources to create a comprehensive report at the level of a research analyst.

Thus, the input costs would correspond to the web search result ingestion and analysis.

Yes, I know. But I’m talking about high input tokens consumption. And as you can see there are reasoning tokens usage calculation.
And also, over 2 mln input tokens usage with one small query - do you think it is ok?

I have been talking about this too .. here. O3-deep-research - 1 million tokens spent .. no output :( - #28 by jlvanhulst

have you found that the Deep Research though the API is more capable than Deep Research on chatgpt?