Learning How to Remember: A Meta-Cognitive Management Method for Structured and Transferable Agent Memory
By: Sirui Liang , Pengfei Cao , Jian Zhao and more
Potential Business Impact:
Teaches computers to remember and learn better.
Large language model (LLM) agents increasingly rely on accumulated memory to solve long-horizon decision-making tasks. However, most existing approaches store memory in fixed representations and reuse it at a single or implicit level of abstraction, which limits generalization and often leads to negative transfer when distribution shift. This paper proposes the Meta-Cognitive Memory Abstraction method (MCMA), which treats memory abstraction as a learnable cognitive skill rather than a fixed design choice. MCMA decouples task execution from memory management by combining a frozen task model with a learned memory copilot. The memory copilot is trained using direct preference optimization, it determines how memories should be structured, abstracted, and reused. Memories are further organized into a hierarchy of abstraction levels, enabling selective reuse based on task similarity. When no memory is transferable, MCMA transfers the ability to abstract and manage memory by transferring the memory copilot. Experiments on ALFWorld, ScienceWorld, and BabyAI demonstrate substantial improvements in performance, out-of-distribution generalization, and cross-task transfer over several baselines.
Similar Papers
From Experience to Strategy: Empowering LLM Agents with Trainable Graph Memory
Computation and Language
Helps AI remember past lessons to solve problems better.
Agentic Memory: Learning Unified Long-Term and Short-Term Memory Management for Large Language Model Agents
Computation and Language
Helps computers remember more for longer tasks.
Learning Hierarchical Procedural Memory for LLM Agents through Bayesian Selection and Contrastive Refinement
Machine Learning (CS)
Teaches computers to learn new tasks faster.