In the old times, if you write a SYSTEM prompt and delete default USER input, a reasonable output will be generated, but now, if you do so, garbage will come out.
It does feel like in some areas it has been underperforming older chatgpt versions. I do think that they work hard on improving overall user experience but feels bad that it is somewhat give an take between use cases users.
I think they might have distilled GPT3.5âs 175B params into a smaller model.
Warning: It is just a thought! I guessed. It has a very high probability to be wrong.
You should give AlphaWave a try⌠https://GitHub/Stevenic/AlphaWave-py
I guarantee youâll get valid JSON back every time. Even with GPT-3.5-turbo
The prompt could be improved though. It is a natural language system with the reasoning ability of a 5-year old. Ideally talk to it like you would a child.
Even as an adult who does programming, itâs not entirely clear what youâre expecting. You have to give it some examples.
Also the system instructions are more like âsoftâ instructions with GPT3.5, and they seem to lose fine details like commands and formatting.
Likely wrong conclusion to think that they spend another million dollars in compute just to train on smaller tokens or parameters, when the lack of computing resources are the biggest hurdle to moving forward.
Thereâs lots of other optimizations to be done to an AI that cut the generation resources at the expense of quality.
I can confirm it is not just 3.5 but also 4 and 4-0316 in the playground for me. Very noticeable drop in memory, logic and reasoning
In the âold versionâ (if thereâs any), the model would simulate user input and outputs.
It used to work as zero-shot model, though.
Actually I havenât get GPT-4 API access until now so I have no clue howâs GPT-4 doing.
But Iâd rather spend more on better models rather than decreasing the price and make the model worse.
Youâre right, but they can just release more versions and make pricing more detailed.
Big updates to 3.5-turbo today, new version of the model releasing
Welcome to the forum!
I know this update, I thought they made old version more stupid and the new version better than the old one.
The 16k-model seemed to be better.
The new version is significantly less intelligent. Itâs depressing because I got online thinking the new engine would be better. This is what weâre stuck with.
Sad, maybe they know there are competitors so they started to make it cheaper.
I completely agree! Iâm deeply involved in cold email outreach and Iâve found that the new model seems almost programmed to start every email with âI hope this email finds you wellâ. Despite spending hours trying to find a workaround, Iâve had no luck in changing this default greeting. Interestingly, this wasnât an issue with the previous version.
They changed stuffs, without changing the model name.
Have you tried adding enforcement to the input itself for the most important parts, vs depending on the internal guidelines/rules to work? I have found that it works great. If you havenât and want to try I can give you more tips, just let me know.
Itâs very frustrating how it refuses to follow some instructions now. My theory is that the bulk of the computational power for an LLM is comprehending the request, similar to a human.
If we as humans understand the instructions, giving a response (even if itâs wrong) is relatively easy. If we are given detailed/advanced instructions, understanding them requires a lot of brain power and we will only respond / do the parts we understand.
Thatâs essentially what the engine is doing now. Itâs not using computational resources to comprehend the request, itâs simply blast out a response to whichever part is easiest for it to understand.
I imagine Microsoft and OpenAI are going to save hundreds of millions, if not billions with this change, but itâs terrible for the end-user. They should at minimum give it 10% intelligence back.
If that works well for you, maybe ClosedAI is trying to force users to consume more tokens so they make more money, evil move.
Thatâs how they would make more money hehe.