A model from 1968

In 1968, Richard Atkinson and Richard Shiffrin published a paper that would define how psychologists think about human memory for the next half century. Their multi-store model proposed that memory is not a single system but a pipeline of three distinct stores, each with different capacities, durations, and encoding mechanisms.

The model was elegant in its simplicity: information enters through a sensory register, gets promoted to short-term memory if it receives attention, and eventually consolidates into long-term memory through rehearsal and repetition. At each stage, information can be lost. That loss is not a failure — it is the system working as designed.

When we set out to build a memory system for AI agents, we did not start with database schemas or vector embeddings. We started with this model. And it mapped perfectly.

The three stores

Sensory RegisterSession buffer — raw capture, high volume
↓ attention gate
Short-Term Memory7-day half-life — recently important
↓ rehearsal & promotion
Long-Term Memory60-day half-life — consolidated knowledge

Sensory Register: everything the agent sees

In the human model, the sensory register captures raw perceptual input for fractions of a second. Most of it is discarded immediately. Only what you pay attention to moves forward.

In NEXO Brain, the sensory register is the session buffer. Every conversation produces a stream of information: user statements, agent responses, decisions made, errors encountered. The session buffer captures all of it as structured events. At session end, an extraction process identifies what matters — corrections, factual claims, preferences, decisions — and promotes those items to short-term memory. The rest is discarded.

This is the first layer of filtering. Without it, memory fills up with noise: greetings, thinking-out-loud, and context that was relevant only in the moment.

Short-Term Memory: the working set

Atkinson and Shiffrin described short-term memory as having limited capacity (roughly 7 items, per George Miller's famous paper) and rapid decay. Information stays here only if it is actively rehearsed.

NEXO's STM holds memories with a 7-day half-life. A memory created today starts at full strength (1.0). After 7 days without access, its strength drops to 0.5. After 14 days, 0.25. Eventually it falls below the retrieval threshold and becomes dormant.

But here is the key: every time a memory is retrieved, its strength resets. A fact you look up every day never decays. A decision you made once and never reference fades naturally. This mirrors how human short-term memory works: rehearsal keeps information alive, neglect lets it go.

Long-Term Memory: consolidated knowledge

Long-term memory in the Atkinson-Shiffrin model has theoretically unlimited capacity and much slower decay. Getting information there requires effort — repetition, meaningful encoding, or strong emotional association.

In NEXO, memories are promoted to LTM when they demonstrate sustained importance: repeated access, high relevance scores, or explicit pinning by the user. LTM memories have a 60-day half-life, giving them roughly 8x the persistence of STM. They are the agent's durable knowledge base: your deployment process, your team members' names, the architectural decisions behind your codebase.

Why forgetting is a feature

The Atkinson-Shiffrin model is as much about forgetting as it is about remembering. Information is lost at every stage. This was not a limitation the researchers lamented — it was a design principle they documented.

The same logic applies to AI agents. An agent that remembers everything is not smarter — it is slower and noisier. Every query returns more results, many of them outdated. The context window fills with stale information. Decision quality degrades because the signal-to-noise ratio drops.

NEXO implements forgetting through Ebbinghaus forgetting curves, the mathematical model Hermann Ebbinghaus published in 1885 describing how memory strength decays exponentially over time. The formula is straightforward:

strength = e^(-t / half_life)

Where t is time since last access and half_life depends on the memory store (7 days for STM, 60 for LTM). Pinned memories are exempt from decay entirely.

Rehearsal and reinforcement

In human cognition, rehearsal is the mechanism that moves information from short-term to long-term memory. Repeating a phone number keeps it in STM. Understanding why the number matters — because it belongs to your doctor — encodes it into LTM.

NEXO models rehearsal in two ways. Access reinforcement resets the decay clock every time a memory is retrieved. Spreading activation strengthens connections between memories that are frequently retrieved together, building an associative network that mirrors how human knowledge is structured.

In practice, this means the agent naturally develops clusters of related knowledge. Memories about your deployment process link to memories about your staging environment, which link to memories about your team's code review workflow. Retrieving any node in the cluster activates its neighbors, producing richer, more contextual responses.

From theory to production

We have been running this architecture in production for months. Here is what we have observed:

Atkinson and Shiffrin could not have anticipated that their model would be used to build memory for artificial agents. But the principles they identified — attention gating, limited capacity, rehearsal, and decay — turn out to be exactly what AI agents need. The problem of managing knowledge over time is not new. Cognitive science solved it decades ago. We just had to listen.