Score: 2

Dynamic Long Short-Term Memory Based Memory Storage For Long Horizon LLM Interaction

Published: July 3, 2025 | arXiv ID: 2507.03042v1

By: Yuyang Lou, Charles Li

BigTech Affiliations: University of Washington

Potential Business Impact:

Helps computers remember what you like.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Memory storage for Large Language models (LLMs) is becoming an increasingly active area of research, particularly for enabling personalization across long conversations. We propose Pref-LSTM, a dynamic and lightweight framework that combines a BERT-based classifier with a LSTM memory module that generates memory embedding which then is soft-prompt injected into a frozen LLM. We synthetically curate a dataset of preference and non-preference conversation turns to train our BERT-based classifier. Although our LSTM-based memory encoder did not yield strong results, we find that the BERT-based classifier performs reliably in identifying explicit and implicit user preferences. Our research demonstrates the viability of using preference filtering with LSTM gating principals as an efficient path towards scalable user preference modeling, without extensive overhead and fine-tuning.

Country of Origin
πŸ‡ΊπŸ‡Έ United States

Page Count
8 pages

Category
Computer Science:
Computation and Language