The GPT-5 (non-reasoning) model has recently started to frequently insert U+202F — Narrow No-Break Space instead of the standard U+0020 space.
In my environment (and several other macOS applications like T3 Chat), U+202F renders as a very narrow gap—approximately one-fifth the width of a normal space—making the generated text visually compressed and extremely difficult to read.
Symptom Examples:
- Observed in: Markdown table headers, headings, and at times, in regular prose.
- Not limited to specific threads: Old threads that never produced these characters, now does.
Technical Details:
| Field | Value |
|---|---|
| Model | GPT-5 (non-reasoning variant) |
| Token ID | 35971 (per OpenAI tokenizer) |
| Unicode | U+202F (Narrow No-Break Space) |
| OS | macOS Sequoia 15.6.1 |
| Hardware | Apple M1 Max, 32 GB RAM |
| Interface | T3 Chat / Other macOS UIs |
Request:
Could the OpenAI team clarify if this frequent substitution of U+0020 with U+202F is intentional (e.g., related to tokenization or watermarking), and if not, please flag this for a normalization fix in the GPT-5 generation pipeline. A consistent use of U+0020 for general spacing is crucial for UI readability.
Thank you,
Patrik