Towards Effective Model Editing for LLM Personalization
By: Baixiang Huang , Limeng Cui , Jiapeng Liu and more
Potential Business Impact:
Teaches AI to remember what you like.
Personalization is becoming indispensable for LLMs to align with individual user preferences and needs. Yet current approaches are often computationally expensive, data-intensive, susceptible to catastrophic forgetting, and prone to performance degradation in multi-turn interactions or when handling implicit queries. To address these challenges, we conceptualize personalization as a model editing task and introduce Personalization Editing, a framework that applies localized edits guided by clustered preference representations. This design enables precise preference-aligned updates while preserving overall model capabilities. In addition, existing personalization benchmarks frequently rely on persona-based dialogs between LLMs rather than user-LLM interactions, or focus primarily on stylistic imitation while neglecting information-seeking tasks that require accurate recall of user-specific preferences. We introduce User Preference Question Answering (UPQA), a short-answer QA dataset constructed from in-situ user queries with varying levels of difficulty. Unlike prior benchmarks, UPQA directly evaluates a model's ability to recall and apply specific user preferences. Across experimental settings, Personalization Editing achieves higher editing accuracy and greater computational efficiency than fine-tuning, while outperforming prompting-based baselines in multi-turn conversations and implicit preference questions settings.
Similar Papers
A Survey on Personalized Alignment -- The Missing Piece for Large Language Models in Real-World Applications
Computation and Language
Teaches AI to be helpful and kind, your way.
CUPID: Evaluating Personalized and Contextualized Alignment of LLMs from Interactions
Computation and Language
Tests AI chatbots on remembering your shifting likes
CUPID: Evaluating Personalized and Contextualized Alignment of LLMs from Interactions
Computation and Language
Tests AI's grasp of shifting user likes across chats