I’ve noticed in the last few days that api calls using gpt-4o
are hanging after the first chunk is received. I don’t think this was happening previously. Whats weird is as soon as the second chunk comes back the entire payload is right behind it, similar to a non-streaming request.
Anyone see me doing something wrong?
Using Azure, latest openai nodejs version
const stream = await openai.chat.completions.create({
model: 'gpt-4o',
messages: messages,
stream: true,
});
for await (const part of stream) {
console.log(`created ==> `, part.created, Utils.getFormattedDate());
...
}
created ==> 0 2024-06-04T14:57:53.299
# see the ~18 second difference here between the first chunk and second
created ==> 1717538273 2024-06-04T14:58:11.624
created ==> 1717538273 2024-06-04T14:58:11.626