I have created a batch file containing 1100 requests (i believe the limit is 50,000?) which fails with an error
Enqueued token limit reached for gpt-3.5-turbo in organization org-xxx. Limit: 200,000 enqueued tokens. Please try again once some in_progress batches have been completed.
Example of a single request:
{“custom_id”: “1186”, “method”: “POST”, “url”: “/v1/chat/completions”, “body”: {“model”: “gpt-3.5-turbo”, “max_tokens”:500, “temperature”:0.2, “messages”:[{“role”:“user”,“content”:“Write a product description in pure html format that is optimized for search engines for a product made by Apple, Apple iPhone 15 Pro 128 GB in Black Titanium with manufacturers part code MTUV3ZD/A. In the first paragraph describe the product and benefits. In a seperate paragraph list the product features, product specifications and product compatibility. Use a professional tone to attract potential customers. Do not include image links. At the end include in a small red font "Disclaimer: While every reasonable effort is made to ensure that the product specification is accurate, no guarantees for the accuracy of information are made."”}]}}
I have a max_token limit of 500 per item, so wondering why i’m hitting a token limit with what is a relatively low number of requests for a batch?
I have lowered the batch requests from 1100 to 600 which appear to have been accepted as they are now processing, but anything much higher than this fails