Score: 2

Attention Consistency for LLMs Explanation

Published: September 21, 2025 | arXiv ID: 2509.17178v1

By: Tian Lan , Jinyuan Xu , Xue He and more

Potential Business Impact:

Shows how smart computer programs make choices.

Business Areas:
Semantic Search Internet Services

Understanding the decision-making processes of large language models (LLMs) is essential for their trustworthy development and deployment. However, current interpretability methods often face challenges such as low resolution and high computational cost. To address these limitations, we propose the \textbf{Multi-Layer Attention Consistency Score (MACS)}, a novel, lightweight, and easily deployable heuristic for estimating the importance of input tokens in decoder-based models. MACS measures contributions of input tokens based on the consistency of maximal attention. Empirical evaluations demonstrate that MACS achieves a favorable trade-off between interpretability quality and computational efficiency, showing faithfulness comparable to complex techniques with a 22\% decrease in VRAM usage and 30\% reduction in latency.

Repos / Data Links

Page Count
15 pages

Category
Computer Science:
Computation and Language