Is it my imagination, or was my recently restricted quota of 40 prompts per 3 hours (which I never agreed to) depleted before I submitted 40 messages today?
Recommendation:
OpenAI should provide a log of the submitted prompts, so customers have visibility of their ChatGPT transactions, like a bank statement. This seems like a reasonable request since we pay for the service and get cut-off based on the number of transactions. I want the ability to review my OpenAI usage to ensure any cutoffs by Open AI are valid.
OpenAI should provide full transparency of transactions. I want the equivalent of a bank statement, including my fees (deposits). costs (prompts) and revenue (based on upcoming revenue sharing) so it’s clear what I paid and what transactions depleted my entitlement.
If a specific GPT significantly depletes my OpenAI quota, then I need visibility of this.
Currently, it feels like OpenAI are dynamically revising their offering based on their ability to handle the number of user requests. Any such changes need to be clearly communicated to users, especially if they’ve signed up expecting more.
I want to add that users who suddenly find themselves unable to continue their conversations due to intransparencies in the GPT system will be turned off.
Not just from the GPT they had just been using but from using custom GPTs at all.
The number of messages used from the 40 messages per three hours for each single message send to the Custom GPT.
If I send one message and get one reply but whatever the GPT did counts as several replies.
Pretty much what you wrote but from a different perspective.
If someone responds to a GPT once, that can potentially count as more than one of the users allocated 40 prompts per 3hrs. eg: if the GPT being used made 10 calls, that would count as 10 calls against the original users quota, without the GPT user having visibility of how many transactions the GPT requires?
Is this what you’re saying? And is this a guess or do you know for sure?
This is what users have been reporting. And comparing this to what is being billed from using a Assistant run it appears to be a reasonable assumption.
If that’s correct, it highlights the need for users of GPTs / assistants to have visibility of the number of transactions associated, so they know what they are paying for.
Otherwise, GPT subscribers will be disappointed when they find their 40 prompts unexpectedly depleted due to running a couple of GPTs.
I have been able to see that, additionally, if in the generation of a response an error is generated, it still counts the message as “used”. It has already become usual for me to reach the limit of requests when resending one that previously threw an error.
Not trying to beat a dead horse here, but this is absolutely maddening. If they INSIST on a cap, they should be very open about it. Give users a visual indicator of how much use they have left instead of just letting us hit a brick wall when we’re finally getting steam. There is an ebb and flow in interacting with AI, I’ve found, at least for my purposes. When the limit pops up suddenly it makes me never want to come back because I’ve got to build the same energy I did before.
Hi, I just joined based on your thread ha ha, thanks bro because I am getting throttled like a moty foty all the time now as a paid subscriber. So I sign The Petition of your Thread!
This is troublesome as both a user and a developer, but I have developed a development flow to help relieve the frustration.
Use 3.5 for really basic programming help.
Test my GPTs with 4.0 of course.
When I hit the cap, go over and work on improving my API, using Postman to test.
When I get my GPT usage back, test my new and improved API.
a) Other users get the same experience, at different times of the day eg: How dynamic is the cap?
b) How much the 40 prompt cap can vary based on running GPTs? And if GPT’s do vary in their impact on user cap, then what GPT can deplete the users quota fastest?
Of course, my cap is now reached, so I can’t test further.
I think you’re missing the point. The question isn’t whether it is 40 simple questions within 3 hours, but how it counts errors and long requests, and when exactly these 3 hours start. Is it from the moment the first message is written, or is it a 3-hour difference between the last message and the 40 messages before it? This is a significant distinction.
Furthermore, there’s an issue with intentional errors that GPT seems to provide, especially when using Analysis (as in my case, over 90%) It appears to be deliberate errors that aren’t treated as questions at all but simply result in error responses. When you consider the cost savings in processing power on an economy of scale, it amounts to millions of euros per day, and that’s not a contract I signed up for.
Typically, when there’s a lack of transparency in anything in life, there’s always a reason behind it that’s unlikely to please the paying party.
This was my first measurable test of the user caps. And I thought some users on this forum may find it interesting. Obviously, it’s best to start with the most basic tests, such as checking if we actually get 40 messages per 3 hours as stated. My initial test confirmed that 40 prompt cap.
My second test, found that listing questions inside an excel spreadsheet and uploading it resulted in many more questions being responded to, although I haven’t measured the limits of this.
This technology is evolving rapidly, but anyone suggesting “intentional” and “deliberate” errors should really present evidence. It’s widely recognised that LLM solutions are not foolproof, yet are improving significantly and the main AI LLM providers are open about this.
I agree, but I don’t believe OpenAI will be willing to provide any evidence, much like their lack of transparency in the 40/3 issue.
Regarding the analysis errors, it’s quite straightforward. When you instruct GPT-4 to search the web for a specific topic (let’s take “cats vs. dogs in cartoons”) and request it to create a title, description, and article without capitalizing titles (i.e., avoid using Title Case), approximately 70% of the time, it will return an error and stop. Out of the remaining 30%, roughly 95% will still use Title Case for everything, not just the title. If you replicate this process on some other ChatGPT account using VPN and different geographic locations, the errors disappear. It’s important to note that these mistakes don’t seem to be specific to particular locations or accounts but rather depend on the primary usage pattern of the account. So if you’re not receiving the error, it’s likely because you use it primarily for something else. I’m not referring to GPTs or the API .