LLMs Can't Play Hangman: On the Necessity of a Private Working Memory for Language Agents
By: Davide Baldelli , Ali Parviz , Amal Zouaq and more
Potential Business Impact:
Gives AI a secret notebook to remember things.
As LLMs move from text completion toward autonomous agents, they remain constrained by the standard chat interface, which lacks private working memory. This raises a fundamental question: can agents reliably perform interactive tasks that depend on hidden state? We define Private State Interactive Tasks (PSITs), which require agents to generate and maintain hidden information while producing consistent public responses. We show theoretically that any agent restricted to the public conversation history cannot simultaneously preserve secrecy and consistency in PSITs, yielding an impossibility theorem. To empirically validate this limitation, we introduce a self-consistency testing protocol that evaluates whether agents can maintain a hidden secret across forked dialogue branches. Standard chat-based LLMs and retrieval-based memory baselines fail this test regardless of scale, demonstrating that semantic retrieval does not enable true state maintenance. To address this, we propose a novel architecture incorporating an explicit private working memory; we demonstrate that this mechanism restores consistency, establishing private state as a necessary component for interactive language agents.
Similar Papers
Procedural Memory Is Not All You Need: Bridging Cognitive Gaps in LLM-Based Agents
Artificial Intelligence
Makes AI smarter in tricky, changing situations.
LLMs and their Limited Theory of Mind: Evaluating Mental State Annotations in Situated Dialogue
Computation and Language
Helps teams spot misunderstandings in their talks.
Language Models Do Not Have Human-Like Working Memory
Computation and Language
Computers forget things, making them make mistakes.