Hi I’m experiencing an unusual token usage situation and seeking insights.
Current Usage Status:
- Dashboard total tokens: 77,720
- Input tokens: 77,056
- Output tokens: 664
Actual API Requests (curl commands):
- First Chat Completion Request:
curl https://api.openai.com/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer [REDACTED]" \
-d '{
"model": "gpt-4o-mini",
"messages": [
{
"role": "user",
"content": [
{
"type": "text",
"text": "Analyze the image and extract a prompt ..."
},
{
"type": "image_url",
"image_url": {
"url": "..."
}
}
]
}
],
"max_tokens": 300
}'
- Second Chat Completion Request:
curl https://api.openai.com/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer [REDACTED]" \
-d '{
"model": "gpt-4o-mini",
"messages": [
{
"role": "user",
"content": [
{
"type": "text",
"text": "Analyze the image and extract a prompt ..."
},
{
"type": "image_url",
"image_url": {
"url": "..."
}
}
]
}
],
"max_tokens": 300
}'
Observations:
- Actual logged requests are significantly lower than reported usage
- Requests were primarily chat completions
- Each request averaged around 1,000 tokens
Questions for Community:
- Has anyone experienced similar unexpected token consumption?
- Are there known billing quirks with recent API updates?
- Recommendations for accurate usage tracking?
Appreciate any insights or experiences you can share.