It’s very, very bad now for coding. Feels like GPT-4 Turbo level, or close.
o3-mini-high was miles ahead of the current o4-mini-high or o3.
These models seem heavily gimped, possibly due to compute constraints? They don’t really think anymore - they feel lazy and shallow. Hellish experience after o1 and o3-mini.
How does that reduce costs, if now I need to enter 10 separate requests to solve a simple task that took only 1-3 prompts with o3-mini-high?
Who’s going to give me back the time wasted debugging the nonsense these newer models output?
And then I have to submit 30 prompts just to get the simplest code working? That’s absurd.
I genuinely can’t believe that the amazing engineers at OpenAI - the same people who delivered the excellent o3-mini and o1 - are behind these current models.
It’s astonishing and mind-boggling how such a downgrade made it into production.
How did this even pass QA?
If it did pass, then… is this level of performance actually intended?
Are there any updates from OpenAI?
Any fixes coming?