Exploiting Contextual Knowledge in LLMs through V-usable Information based Layer Enhancement
By: Xiaowei Yuan , Zhao Yang , Ziyang Huang and more
Potential Business Impact:
Helps computers remember and use information better.
Large Language Models (LLMs) have demonstrated remarkable capabilities in various tasks, yet they often struggle with context-faithfulness generations that properly reflect contextual knowledge. While existing approaches focus on enhancing the decoding strategies, they ignore the fundamental mechanism of how contextual information is processed within LLMs' internal states. As a result, LLMs remain limited in their ability to fully leverage contextual knowledge. In this paper, we propose Context-aware Layer Enhancement (CaLE), a novel intervention method that enhances the utilization of contextual knowledge within LLMs' internal representations. By employing V-usable information analysis, CaLE strategically amplifies the growth of contextual information at an optimal layer, thereby enriching representations in the final layer. Our experiments demonstrate that CaLE effectively improves context-faithful generation in Question-Answering tasks, particularly in scenarios involving unknown or conflicting contextual knowledge.
Similar Papers
Semantic Mastery: Enhancing LLMs with Advanced Natural Language Understanding
Computation and Language
Makes AI understand and talk like people.
On the Power of Context-Enhanced Learning in LLMs
Machine Learning (CS)
Teaches computers to learn much faster and safer.
Enhancing LLM Knowledge Learning through Generalization
Computation and Language
Helps computers remember new facts without forgetting old ones.