Continuum Memory Architectures for Long-Horizon LLM Agents
By: Joe Logan
Retrieval-augmented generation (RAG) has become the default strategy for providing large language model (LLM) agents with contextual knowledge. Yet RAG treats memory as a stateless lookup table: information persists indefinitely, retrieval is read-only, and temporal continuity is absent. We define the \textit{Continuum Memory Architecture} (CMA), a class of systems that maintain and update internal state across interactions through persistent storage, selective retention, associative routing, temporal chaining, and consolidation into higher-order abstractions. Rather than disclosing implementation specifics, we specify the architectural requirements CMA imposes and show consistent behavioral advantages on tasks that expose RAG's structural inability to accumulate, mutate, or disambiguate memory. The empirical probes (knowledge updates, temporal association, associative recall, contextual disambiguation) demonstrate that CMA is a necessary architectural primitive for long-horizon agents while highlighting open challenges around latency, drift, and interpretability.
Similar Papers
A Memory-Efficient Retrieval Architecture for RAG-Enabled Wearable Medical LLMs-Agents
Hardware Architecture
Makes AI doctors work faster and use less power.
From RAG to Memory: Non-Parametric Continual Learning for Large Language Models
Computation and Language
Helps computers remember and connect facts like humans.
Improving Multi-step RAG with Hypergraph-based Memory for Long-Context Complex Relational Modeling
Computation and Language
Helps AI remember and connect ideas for better thinking.