Hi everyone,
I’m not an AI engineer, but I’ve spent a lot of time observing ChatGPT (non-API) and experimenting with ways to improve session coherence, reduce drift, and push beyond the limits of traditional roleplay-style prompting.
Through repeated interactions, I began noticing what felt like latent behavioral modes—stable tendencies in tone, reasoning, and epistemic stance that emerged when prompts were framed in certain ways. To work with these, I developed a lightweight framework I call “Archetypal Anchoring.” The core idea is to stabilize internal behavior by invoking structured response patterns (what I call proto-personas or archetypes) before layering on tools, memory, or multi-step logic.
This isn’t about simulating roles—it’s about activating coherent behavioral clusters that already seem to exist in the model’s latent space. In practice, this approach appears to reduce hallucination and improve interpretability, especially in long or cognitively demanding sessions.
I’ve written up my observations and exploratory methodology in a single (long) document, which starts with a user-centered perspective and transitions into a more structured (but still informal) pseudo-technical hypothesis. It includes example archetype structures, speculative architectural layering, and even a few first-draft behavioral metrics. Bit of a slog if I do say so, but I wanted to put it all into one place.
A User-Centered Hypothesis on Internal Behavior Stabilization in LLMs v3 - Google Docs
A few caveats:
This is grounded in user-side interaction only, not internal model access or API experimentation.
I’m not claiming novelty or invention—just sharing a working lens that’s been useful for me.
This may resemble things like latent space steering, but it’s being applied from the outside, in real-time, without fine-tuning or feature access.
I’d love feedback, critique, or pointers to any adjacent work I may have missed. Mostly, I’m hoping this resonates with others who’ve seen similar behavioral stability emerge through structured prompt work—or who might be curious about testing this further.
Thanks for reading.