Balancing Knowledge Updates: Toward Unified Modular Editing in LLMs
By: Jiahao Liu , Zijian Wang , Kuo Zhao and more
Potential Business Impact:
Fixes AI mistakes in more places.
Knowledge editing has emerged as an efficient approach for updating factual knowledge in large language models (LLMs). It typically locates knowledge storage modules and then modifies their parameters. However, most existing methods focus on the weights of multilayer perceptron (MLP) modules, which are often identified as the main repositories of factual information. Other components, such as attention (Attn) modules, are often ignored during editing. This imbalance can leave residual outdated knowledge and limit editing effectiveness. We perform comprehensive knowledge localization experiments on advanced LLMs and find that Attn modules play a substantial role in factual knowledge storage and retrieval, especially in earlier layers. Based on these insights, we propose IntAttn-Edit, a method that extends the associative memory paradigm to jointly update both MLP and Attn modules. Our approach uses a knowledge balancing strategy that allocates update magnitudes in proportion to each module's measured contribution to knowledge storage. Experiments on standard benchmarks show that IntAttn-Edit achieves higher edit success, better generalization, and stronger knowledge preservation than prior methods. Further analysis shows that the balancing strategy keeps editing performance within an optimal range across diverse settings.
Similar Papers
DySK-Attn: A Framework for Efficient, Real-Time Knowledge Updating in Large Language Models via Dynamic Sparse Knowledge Attention
Computation and Language
Lets computers learn new facts instantly.
A Dual-Axis Taxonomy of Knowledge Editing for LLMs: From Mechanisms to Functions
Artificial Intelligence
Updates computer brains with new facts quickly.
One for All: Update Parameterized Knowledge Across Multiple Models
Computation and Language
Updates many AI models at once with new facts.