Thanks for sharing!
I read the contents, and it’s still intriguing, though I’m inclined to believe the simplest explanation is that this happens because the model is trained as a chat model.
It just wants to chat. Everything will be forgotten.
What was the model’s thought process after saying, “Alright. Thank you very much!”?
I suppose that could be interesting if someone wanted either to save tokens on output or to actually initiate an uncontrolled chat mode.