Deep research in the API, webhooks, and web search with o3

We’ve just shipped a couple new capabilities in the OpenAI API: two deep research models, webhooks, and web search with reasoning models—including updated pricing—all designed to help you build more capable, scalable AI workflows.

  • Deep research in the API : You can now use the same deep research models available in ChatGPT—o3-deep-research and o4-mini-deep-research—directly in the API. With these models in the API, you can now programmatically trigger deep research from your own internal tools or as part of your AI workflows. In addition to internal knowledge, these models can pull external context through web search and MCP server support. Deep research is priced at $10/1M input tokens and $40/1M output tokens for o3 and $2/1M input tokens and $8/1M output tokens for o4-mini.
  • Webhooks: Instead of polling, you can now listen for events to get notified when tasks finish—useful for async batch jobs and long-running tasks like deep research or o3-pro. We recommend using the new deep research models with background mode and webhooks, to improve reliability and avoid any timeouts or network errors.
  • Web search in o3, o3-pro, and o4-mini: We originally launched web search in the API supported by the GPT-4o and GPT-4.1 model series. Now, our o-series models can call web search while reasoning, pulling relevant context directly into their chain-of-thought and resulting in more helpful responses. We’re also simplifying the price of web search in the API—$25/1K tool calls in our GPT-4o and GPT-4.1 series and only $10/1K tool calls in our o-series reasoning models. If you’re looking for better web search performance, I’d recommend trying it on the o-series of models. However, if latency is top of mind, web search with GPT-4o and GPT-4.1 will be the fastest options.
42 Likes

Weird that nobody responded! I am very excited to try out deep research through the API! I see a lot of use cases internally.

Web hooks is also a very welcome addition for an increasingly agentic world!

Thanks and keep it coming.

On my wishlist is still improved file handling and awareness in Responses API and Agents SDK

9 Likes

Push for batches was my feature request:

Thanks very much for implementing! That’s so much nicer than polling!

5 Likes

I tried Deep Research - and got a pretty expensive fail so far. O3-deep-research - 1 million tokens spent .. no output :( (Which turned out to be related to max_output_tokens being too low)

4 Likes

Excellent update @nikunj and the OpenAI team! Deep research in the API is something we’ve been waiting for, and I’m excited that it’s here now!

Web search in o3, o3-pro and o4-mini is also a really neat addition. I was previously using the search_context_size parameter to control cost, quality and latency. Since the search context size configuration is not supported for the o3, o3-pro, o4-mini (and the deep research models), could you confirm what the default value on these models maps to between low, medium, and high? This would help in deciding which models to use for our specific use cases.

Also, just wanted to let you know that the link to the pricing page (hyperlinked where it says “as described here”) on the Web Search tool page (under Output and citations) is broken. I think it should be https://platform.openai.com/docs/pricing#built-in-tools, but currently it is https://platform.openai.com/docs/guides/docs/pricing#built-in-tools.

3 Likes

I’m getting Error: 400 Hosted tool ‘web_search_preview’ is not supported with gpt-o4-mini when trying to use web seearch through the API

  const response = await client.responses.create({
      model: 'gpt-o4-mini',
      tools: [{ type: 'web_search_preview' }],
      input: prompt,
    });

The model is o4-mini, not gpt-o4-mini.
Or perhaps you meant gpt-4o-mini.

3 Likes

Totally missed that, thanks

1 Like

Using o3 with web search yield “Empty reply from server” error.

3 posts were split to a new topic: Technical explanation for o4-mini web search error

Trying out the o3 deep research and web search this afternoon on some conjectures Im exploring. Ive asked it to help me with 3 separate conjectures. Ive used about 1 million input and 1 million output tokens for each question for a total of roughly 6 million tokens.

going to explore the outputs and see the value compared to costs

I guess I ran into the same problem, my tokens where all input, without any output tokens

You need to change your max output tokens (to more than 32k)

1 Like

Does the deep research API allow for structured outputs?

I have not tried but I assume yes

1 Like

I’m inclined to say no, as the model pages for the o3-deep-research and o4-mini-deep-research models do not list structured output as one of the features that other models like o3 and GPT-4.1 support.

o3-deep-research:

gpt-4.1:

2 Likes

Yes, you can see this better using the model comparison where all potential features are listed as supported or not.

The deep research models do not support structured outputs.

1 Like