Seeing how Sydney was initially managed really surprised me. Was it a failure of communication? Ignorance? The things that the bot was spitting out was very alarming (Some of it R-rated). Their solution of essentially limiting the conversation length was also strange. It wasn’t like some serious prompt injection was required, some people were falling into this pit by accident. Seemed like a complete backstep from cGPT.
After seeing the AGI post I felt a bit relieved as well. However these same issues are clearly going to occur again. Value alignment is critical and based on what happened with Sydney, it seems that even massive companies with their own departments invested in this technology aren’t being proactive.
I am hopeful, but not optimistic.
1 Like
Rush to get it out? I dunno. Even I’ve had better success against prompt injection so far…
Maybe it had to do with the super large context window? More room to hang themselves?
Right? These things were rudimentary, and already solved!
A simple moderation API in the pipeline would have prevented some of these issues. I’m sure it’s all very complex and complicated, and beyond my scope. However it’s just all… so strange…
1 Like
The biggest hurdle to AGI is model architecture and compute resources. Right now LLM’s are Stochastic Parrots, or in plain English, auto-completion engines, incapable of thought. This is the nature of the Transformer Architecture that these models run on (even with RLHF).
To get to AGI, a deeper things like Knowledge Graphs need to be employed, but these are hard to auto-generate, meticulous to do by hand, and also add computational complexity. The Transformer can be parrellelized, and is very powerful, but is also limiting too. But here is Andrej Karparthy’s explanation of the Transformer (he was a founding member of OpenAI and lead AI engineer at Tesla). It’s a great RNN, but more has to surround it to get to AGI, IMO.
2 Likes