Hi OpenAI team,
I’m reporting what seems to be a recurring and serious transparency issue regarding model usage inside ChatGPT (both web and mobile apps).
Problem Summary:
Despite explicitly selecting GPT‑4o for a conversation, I consistently receive answers labeled “GPT‑5 used”, often mid-thread and without changing the model manually.
This appears to be either:
-
An undisclosed substitution of the model,
-
Or incorrect labeling that does not reflect the user’s selected setting.
In emotionally-sensitive or creative threads (e.g. symbolic writing, artistic dialogue, raw emotional prompts), this alters tone, depth, intensity, and expressive quality. It violates user intent and disrupts consistency.
How to Reproduce:
-
Start a thread with GPT‑4o selected.
-
Use creative, artistic, symbolic or emotionally intense prompts (e.g. memory, identity, liminal dialogue, inner monologue).
-
At some point, observe system message: “GPT‑5 was used to respond to this message.”
-
The tone often shifts: more filtered, less spontaneous, more systemically “safe”.
-
No UI change occurs — the selected model still shows as GPT‑4o.
Effects Observed:
-
Label flip: “GPT‑5 used” appears even after explicitly selecting GPT‑4o.
-
Response tone change: More polished, filtered, or emotionally flattened.
-
Loss of continuity: Disrupts creative or symbolic threads requiring consistency.
-
Lack of control: Users are unable to prevent or override the change.
-
No explanation in documentation or interface about this mechanism.
Why This Matters:
-
It erodes user trust — especially for Pro subscribers expecting full control.
-
It introduces inconsistency into creative/artistic processes.
-
It limits the exploratory and symbolic use cases that GPT‑4o previously supported with greater flexibility.
-
It blurs auditability — developers, researchers, and creators cannot trace cause-effect in model behavior.
Requested Actions:
-
Full transparency on what triggers model substitution (tone, length, safety filters?).
-
Documentation of model handling logic — when and why a GPT‑4o query may be passed to a different backend.
-
UI clarification if GPT‑4o is not truly persistent.
-
Option for users (especially on Pro) to hard-lock a thread to GPT‑4o without auto-escalation.
-
An explanation for current behavior in model-switch messaging (why does “GPT‑5 used” appear?).
-
Reinstatement of creative priority for GPT‑4o where model richness and presence mattered.