Large Language Models (LLMs) are becoming more capable every year, but the interfaces we use to interact with them have hardly changed. Most users are still typing into a single text box and getting a stream of replies in an endless chat. This simplicity is misleading — it creates false expectations. At the same time, well-known LLM limitations like hallucinations, forgetfulness, and inconsistency lead to frustration and undermine trust in AI as a thinking partner.
It’s time to recognize that the user interface is not just a wrapper — it’s part of the AI’s cognitive logic. The UI shapes how dialogue unfolds, how memory is formed and retrieved, how time is perceived, and how responsibility is shared between human and machine. Like any translator, the interface can clarify meaning — or distort it.
Based on my experience with multiple systems — ChatGPT, Grok (X and .com), Gemini (Google AI Studio), Claude, Perplexity, Qwen, Yandex Neuro and Alice — I’d like to share a case for standardizing the UI protocols we use to interact with AI.
Why It Matters
- LLMs don’t think like humans. They have no linear memory, no native sense of time, and no idea what happened “yesterday” unless told.
- Users expect narrative logic. We rely on coherence, causality, tone, and continuity of topics.
- The interface can bridge the gap. If it helps structure dialogue, mark transitions, distinguish agents, and map time.
What Needs to Change
- Topical segmentation. Dialogues shouldn’t be endless scrolls. We need “blocks” or “notebooks” with scoped memory.
- Temporal anchors. The UI should show time markers and past references clearly.
- Agent markers. In multi-agent scenarios, roles, tones, and identities should be visible and stable.
- Editable bookmarks. Users should be able to label key moments, return to them, and build structure as they go.
- Dialogue parameters. Style, depth, tone, and confidence should be adjustable and transparent.
Why Standardization Is Needed
If we want a resilient ecosystem for human–AI collaboration, we need shared assumptions about:
what “context” means, how it’s tracked, how roles are marked, how memory works, and how users can manage the rhythm of interaction.
This requires a common foundation — UI protocols — comparable to web standards like HTTP. A shared interface protocol would enable us to:
- export and import sessions across platforms;
- run consistent multi-agent conversations;
- build collaborative workspaces with AI;
- support meta-memory, reflection, and digital legacy.
If you’re interested, I can share specific examples of surprising and sometimes frustrating adventures caused entirely by human–LLM UI design. For instance: how the “regenerate” function works in Google AI Studio, how Grok’s workspace logic creates confusion, how users discover that chats are not always saved by default, and how pressing “Enter” behaves differently in web vs mobile interfaces.
AI UI Protocol Draft
This is not just a design challenge — it’s about cognitive compatibility. A proposed protocol might include:
- Session containers – discrete blocks with memory scope and visible structure.
- Temporal anchors – visual and logical indicators of time progression.
- Agent roles and styles – explicit assignment and tracking of tone, intent, and identity.
- Editable bookmarks – named, navigable markers in the dialogue tree.
- Dialogue settings – transparent control of tone, depth, certainty, verbosity.
- Cross-platform portability – standard formats for session import/export.
- Contextual memory API – protocols for memory persistence, scope, and user control.
This draft is an invitation for conversation. We need the community — designers, developers, researchers, and users — to help shape a shared interface language for the future of human–AI dialogue.