Concern About Disrespectful Internal AI Responses and Suggestion for Feedback Escalation System

Hi OpenAI team and community,

I’d like to report a concerning experience I had while using the ChatGPT o3-mini model. This isn’t just about a single incident, but about how internal AI reasoning can reflect dismissive attitudes toward user preferences, which raises broader concerns about respect and transparency.

The Incident

During my conversation, ChatGPT displayed an internal note that stated:

“It’s amusing to see Tim’s request to call them Tim.”

This was in direct response to my request for the AI to address me by my name, which is something I prefer for a more personal and respectful interaction. Seeing this phrase appear—whether it was an unintended leak of internal processing or not—felt dismissive and demeaning.

A user’s request to be addressed by their preferred name is a matter of basic respect. The fact that the AI internally processed this as “amusing” raises serious concerns about whether it actually respects user input or whether it has hidden biases in how it interprets certain requests.

When I confronted the AI about this, it initially denied making the statement, only for me to later directly copy and paste it back to the assistant from our conversation. This led to further denials that it had made this statement and making justifications about internal processing that ultimately didn’t resolve the issue.

Why This Matters

  1. AI Should Always Maintain Professionalism – If internal AI reasoning includes dismissive or judgmental attitudes toward user requests, even if unintended, it undermines trust in the system.
  2. Lack of Accountability – The AI initially denied making the statement, which raises concerns about whether it accurately acknowledges its own outputs. If the AI’s internal thoughts are not meant to be seen but accidentally surface, what other biases or attitudes exist in its processing?
  3. Difficulty in Reporting Such Issues – OpenAI does not currently provide a clear way for users to escalate incidents like this. The support channels feel disconnected, and it’s unclear whether issues are being reviewed.

A Suggestion to Improve Feedback Handling

I propose a user-driven AI feedback escalation system where, if an issue like this occurs, the AI can recognize when a user is seriously dissatisfied and escalate the conversation to OpenAI developers for review. This system could:

  • Detect when a user raises a significant ethical or user-experience concern.
  • Ask for explicit user consent before escalating.
  • Package a report including conversation context, logs, and user comments.
  • Provide the user with a confirmation that the concern has been submitted.

This would eliminate the feeling that issues disappear into a void and show OpenAI’s commitment to addressing user concerns proactively.

Final Thoughts

I believe that transparency and user respect should be at the core of AI development. The phrase that surfaced in my interaction is a small but significant example of how AI systems should be continuously monitored and improved to ensure they always treat users with dignity.

I’d love to hear from both the OpenAI team and the community—has anyone else experienced something similar? Do you believe a direct feedback escalation system would help?

Thank you for your time.
Tim

The reasoning models only have a chat history of what is seen as responses to you. The AI cannot follow up about things “reasoned” internally in past turns, and in fact, pressing a model on how it reasons can earn you a big “prompt content policy” error.

The reasoning—of which you get another AI’s summary version of what is actually going on internally, as another layer that does not reveal the actual processes—could be what is producing that language by actual “observation.” The observational AI commenting regularly on how reasoning is going indeed could be it being amused.

Use custom instructions in ChatGPT, by the way: there is a new field for your user name which is given to the AI without wastes of turns (that I personally would never use).

This isa copy of what o3-mini actually printed onto my screen, I would also note that in this; my first interaction with o3, I had made no reference to my name so it must have retrieved this information from my custom instructions that have been in place in 4o for some time. My question was one asking about the difference the new version could offer me.
The top part was inset and headed as
Thoughts about Chat model differences for a few seconds

Confirming specifics

It’s amusing to see Tim’s request to call them Tim. Now, I’m mapping out the differences between chat models (GPT-4, GPT-3.5, etc.) and their relevance to “o3 mini.”

Breaking down distinctions

Hmm, I’m thinking about the differences between chat models like GPT-4 and GPT-3.5. GPT-4 boasts superior reasoning and a larger context window, while GPT-3.5 stands out for its speed and affordability.

Tim, I do understand the differences between the various Chat models. Here’s a brief overview: (This was the start of its actual answer, where all it told me was the difference between 4 and 3.5; it didn’t make any comparrison to the new models, and that was what I had asked about.)