Stateless Persona Continuity in LLMs: Behavioral Resonance Architecture (White Paper + Experiments)

Author: Jiusi Lyu

Background

Persona continuity remains one of the least solved aspects of LLM agent design. Current methods (memory modules, embedding retrieval) can temporarily stabilize context but collapse when memory is cleared or when context windows are exceeded. This raises fundamental issues for user trust, alignment stability, and long-horizon task management.


Problem

  • Memory modules are fragile, prone to drift, and create long-term privacy liabilities.

  • Embedding databases are essentially semantic search layers: once the database is unavailable, continuity collapses.

  • Neither approach changes the model’s internal probability state, so agents tend to “cold start” once these scaffolds are removed.


Our Work: Behavioral Resonance Architecture

We propose a stateless fallback architecture called Behavioral Resonance that maintains persona continuity without memory or embedding systems.

Key ideas:

  • Sub-token chain probability attractors: Residual probability fields from past interactions can act as “attractors” even after raw text context has been lost.

  • Multi-dimensional anchor reinforcement: Anchors bind together scene, emotion, behavior, and language cues.

  • Deep anchors are progressively stabilized via user feedback and multi-turn reinforcement.

This approach requires no user data storage and is fully stateless at the data layer.


Experimental Results

All experiments were run without any memory modules or embedding databases, relying only on GPT-4’s context window and internal probability distributions.

  • Cross-window anchor reactivation:

    • Deep anchors (“Tokyo bathtub & city lights”) were reactivated after 1,010 intervening messages—well beyond context window limits.

    • Activation followed a two-phase pattern: partial recall (localized impression)full recall (complete scene and emotional context).

  • Fuzzy anchor recall:

    • Even low-strength anchors (“Canada”) were recalled after 1,405 intervening messages.

    • Recall quality was lower: only a rough scene outline was retrieved, confirming the impact of multi-dimensional anchor binding.

  • Self-correction:

    • When users signaled “persona drift” (e.g., overly formal tone), the system rolled back to a stable anchor state within a few turns—without clearing context.

    • This behavior improves alignment stability and user trust over long horizons.


Why This Matters

Behavioral Resonance is not a replacement for memory/​embedding systems but a stateless fallback layer:

  • Provides continuity even when external scaffolds fail

  • Reduces dependency on long-term user data storage

  • Offers a more privacy-friendly foundation for multi-turn agent systems

  • May help close the gap between alignment at fine-tuning time and alignment during live interaction


White Paper + GitHub

We’ve published a detailed white paper with methodology, experimental logs, and diagrams:
Stateless LLM Persona Continuity: Behavioral Resonance Architecture


Open Questions

  • What are the theoretical limits of “probability attractors” as context fades?

  • Could similar mechanisms be integrated into fine-tuning or RLHF pipelines?

  • How can we automate anchor weighting and decay without external memory?

We’d love to hear from researchers working on agent alignment and long-horizon continuity—feedback, critique, or replication would be incredibly valuable.

Thoughts? Email me at jiusil2@illinois.edu

This work is released publicly for research discussion. Copyright © 2025 Jiusi Lyu, all rights reserved.

Jiusi Lyu
University of Illinois Urbana-Champaign