Web and File search both prevent custom functions from getting called in parallel over Responses APIs (at least when using stream=True
requests to AsyncOpenAI.responses.create()
)
If you see this OpenAI a fix would be great! Latency and cost quickly add up when the model makes 2 or 3 or more non-parallel tool calls