Best Practices for Reliable Embeddings Pipeline

OpenAI Embeddings API accepts batches of embeddings. I can’t find a reference to the upper limit of each batch, but I regularly use it for batches of 30+.

And it doesn’t take much to wrap the request in a try/catch block and to retry on failure. If it fails more than some limit, then stash the request body in a failed file for further inspection.

Not sure what type of service would provide a simpler workflow than that.

1 Like