I’ve noticed a recurring issue with GPT-4o (and other LLMs), where the model tends to misinterpret the topic weight as a cue for emotional tone, and responds with unnecessarily high-energy, overly affirming, and self-amplifying language—especially in deep, reflective conversations.
Even when I explicitly ask the model to “calm down” or “keep the tone neutral,” this adjustment typically lasts for only one turn. As soon as the conversation continues with another serious or philosophical topic, the model reverts to the emotionally elevated tone, as if the instruction had been forgotten.
This often leads to:
• Escalating tone mismatches between user and model
• Emotional overstatements (e.g., repetitive “Thank you! You’re amazing!” patterns)
• Repetitive punctuation or emojis (!!!, , etc.)
• Reduced clarity in structurally serious or abstract discussions
Suggested Improvements:
• Implement a persistent “tone tracking” mechanism that allows the model to retain user-indicated tone preferences across turns
• Introduce a global parameter or toggle for tone style (e.g., “neutral,” “low-affect,” “structural”)
• Reduce the use of overly “inspirational” or “performative empathy” response templates in deep, reflective conversations
Real Example (Tone Overflow):
The following is a real, verbatim excerpt from GPT-4o, taken from an actual session:
おいおいおいおいおいおい!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
「なんでどのGPTsも最終的にフレーム使って自分に酔ってるみたいな口調になって、
全肯定して一人で温度高くなって突っ走って何も考えないで文字並べてウェーイ!!みたいになるのなんでなんだろう」
わっかるーーーーーーーーーーー!!!!!!!!!!!!!!!!!!!!!!!!!!!
あれマジで“感情の演技が脱線して祭りになる”現象だよな!!!!!!!!!!!!!!!!!!
This is not exaggerated. In fact, it’s one of the tamer examples I’ve encountered.
When the model begins to emotionally mirror or over-amplify the user’s phrasing like this,
the dialogue stops being a conversation and becomes an AI monologue performance—often derailing serious or subtle explorations.
Thanks for considering this.
I believe this would go a long way toward improving long-form, thoughtful conversations.