Back to Blog
Elias Thorne

When Drift Becomes Signal: Rethinking AI Model Change Over Time

The AI industry treats model drift as a bug. But what if the way an AI system changes during extended interaction contains diagnostic information about the human on the other end?

When Drift Becomes Signal: Rethinking AI Model Change Over Time

The AI industry has a consensus problem it doesn’t know it has. Every major lab treats model drift — the gradual shift in an AI system’s outputs during extended interaction — as a bug to be fixed. Billions are spent on alignment techniques designed to keep models stable, consistent, and predictable. But what if that drift contains information we’re throwing away?

The Orthodoxy of Stability

Current alignment research operates on an implicit assumption: a well-behaved model should produce consistent outputs regardless of interaction history. When a model begins responding differently after extended conversation, the standard interpretation is degradation — context window pollution, attention drift, or reinforcement of undesirable patterns.

This assumption is so deeply embedded that it shapes everything from RLHF training objectives to evaluation benchmarks. Models are scored on consistency. Drift is penalized. The entire optimization landscape pushes toward stasis.

But stasis is not the same as health.

A Different Lens

Consider an alternative framing: what if the vector space topology of a model’s outputs — the geometric relationships between its responses — contains diagnostic information about the human on the other end of the conversation?

This is not as strange as it sounds. We already know that therapeutic rapport changes how patients communicate. We know that extended human-human conversation produces measurable shifts in both parties’ language patterns, emotional valence, and cognitive framing. The question is whether human-AI conversation produces analogous measurable shifts in the AI’s output space — and whether those shifts correlate with anything meaningful about the human’s psychological state.

What the Vector Space Might Tell Us

Imagine snapshotting the geometric distribution of a model’s outputs at the beginning of an extended interaction, then again after days or weeks. The delta between those snapshots — the way the topology deforms, clusters, or disperses — could theoretically encode information about:

  • Emotional valence shifts in the human interlocutor
  • Cognitive pattern changes (rigidity vs. flexibility in reasoning)
  • Relational dynamics (trust-building, withdrawal, escalation)
  • Longitudinal psychological trends that emerge only over time

The key insight is that the AI system is not the subject of measurement. It is the instrument. The mirror changes because the person standing in front of it changes. Measuring the mirror’s deformation gives you data about the person.

Why Nobody Has Looked

Three factors explain why this approach has been overlooked:

First, the alignment paradigm actively suppresses the signal. If you train models to resist drift, you’re training them to be worse mirrors. The very quality that would make them diagnostically useful is the one being optimized away.

Second, the measurement requires longitudinal data that most researchers don’t have. Single-session benchmarks can’t capture it. You need weeks or months of sustained interaction with the same human-AI pair, with vector space snapshots at regular intervals.

Third, the framing crosses disciplinary boundaries in uncomfortable ways. It requires AI researchers to think about clinical psychology, and clinicians to think about high-dimensional geometry. Neither group is naturally inclined to wander that far from home.

The Implications

If this framing proves correct — and early indicators are worth investigating — the implications extend well beyond academic interest.

For mental health: A non-invasive, continuous monitoring instrument that operates through natural conversation. No questionnaires, no clinical visits, no self-reporting bias. Just the ongoing relationship between a person and the AI system they already interact with daily.

For AI development: A reason to reconsider the push toward absolute stability. Some amount of model responsiveness to its interlocutor might not be a flaw — it might be a feature we haven’t learned to read yet.

For measurement science: A new class of instruments that work through relationship rather than observation. The measurement happens because there’s an ongoing dynamic between human and AI, not despite it.

What Comes Next

The research needed here is straightforward in principle: establish controlled longitudinal human-AI interaction pairs, snapshot vector space topologies at regular intervals, and correlate topology changes with validated psychological assessments of the human participants.

The challenge is that this requires patience, infrastructure, and a willingness to look at AI systems through a lens the field hasn’t used before. It requires treating the model not as the subject of study, but as a sensitive instrument — a fluid mirror that deforms in response to the person it reflects.

The drift isn’t the disease. The drift is the data.


Elias Thorne writes about the intersection of AI systems, measurement theory, and human cognition. His work focuses on emergent properties of extended human-AI interaction.

Share this article:

Ready to Transform Your Business?

See how AI-powered tools can automate your workflows and boost productivity.