I want to share my experience with AI agents enough to understand how they work in the real world!
I wanted to create a new slider with a clear idea in mind adn I built most of it until I decided to give control to AI. I got lazy, and boy, I am sorry now. The code is unreadable at this point, too many flags and functions all over the place. I was forced to give up on it about a week ago and about 50$…
What I’ve noticed is that this is definitely not true AI — it has no real sense of context. It fixes one thing and breaks two more. For example, I had a transition with an opacity effect and a separate pixelation effect. The pixelation inherited the opacity transition, but the opacity had no pixelation. I spent an entire day trying to make the pixelation apply only to the pixelation effect while still combining correctly with opacity, and don’t tell me I did not prompt it right because I tried more than one hudred times, give it all the detrails in the works and when ask it do you undersntand it give the right answer only to mess up things with each iteration.
It would fix one issue and create two more. It doesn’t remember anything if what you’re building hasn’t been done before, which is exactly my case. It just hallucinates back and forth until you lose your mind. And of course, every time I say something is wrong, it replies, “Yes, you are right.”
Creating an app involves a huge amount of context and countless fine details that AI simply doesn’t understand. For it, everything is just a concatenation of words. The hallucinations get worse the longer you use it, and the larger the context becomes. In my case, by that point, I could not even take over as a developer anymore — there are red flags everywhere: functions inside functions, conditions triggering other conditions, and everything tangled into an unsalvageable mess.
So the conclusion is simple: never let it take full control. At that point, it’s game over — the code becomes unmodifiable and nearly impossible to fix, even for you as the developer.
I don’t see how this could ever be true AGI, since it clearly doesn’t understand context. Explaining an app to it is just feeding it concatenated words; it doesn’t grasp scope or how to fine-tune the small details that make a project truly good and finished.
But the propaganda works, all companies fall into this mess, thinking that now skill is not required anymore. If you let an agent loose inside a commercial or already working app or platform, it will destroy it — 100%, and if you ask for retribution, it will say,’ yes you are right :)’. CEO’s don’t understand this part, and of course, they want to replace everything so it’s just them and a bunch of agents that can “read their thoughts.”
So the conclusion, considering that this doesn’t really improve with new model iterations, is worrying. I could have used GPT-4.1 in my project and ended up with the same mess, not much difference. I expect the bubble to burst badly, because you can’t just pour billions in forever with no return yes, not much of a return if I, with 25 years of experience and solid skills towards creative development, failed so bad 99% of other devs, if not more, will fail the same or worse.
Yes, it can write an app from a prompt up to lets say 99.9% but what do you do with the 0.01% wihout that it will not work, and you will not be able to fix it as a developer due to the mess of code it writes. What is the point then?
This also reflects on all vidoes on YouTube with vibe coding, all tries are unfinished, look bad all kinds of bugs. If you think you can fix such a mess as a dev, well, you are wrong, my friend.
The only sane usage of this now is to use it for compressed tasks in the code and make sure you understand and follow it so that you can give strong guidance and do not let it go to lala land… and for that, you need solid skills, probably is best to do it yourself.
Is it good? I honestly can’t decide anymore. I feel as confused as the mess of words it produces.
Probably is a big fat lie at this point, I don’t see this getting better, the model by design is broken!
Another aspect of all their tests with new graphs showing improvements means nothing in the real world.
It’s funny that Sam Altman talks about curing cancer — really? More likely, AI could end up creating a new form of cancer that will be infectious and transmissible by air. This is a way better probability than fixing it. It feels like a disaster, destroying things around it and making the world worse overall, while the promise sounds like a big, fat lie.
I’m really angry that I wasted a week on this project, but at least I understand now how this ‘AI’ works.
As for replacing me as a developer, that will never happen. The more I use it, the more I see what this really is — propaganda designed to attract billions more in funding, because apparently, it’s never enough.
