Score: 0

AttenDence: Maximizing Attention Confidence for Test Time Adaptation

Published: November 24, 2025 | arXiv ID: 2511.18925v1

By: Yash Mali

Potential Business Impact:

Helps AI see better when pictures change.

Business Areas:
Image Recognition Data and Analytics, Software

Test-time adaptation (TTA) enables models to adapt to distribution shifts at inference time. While entropy minimization over the output distribution has proven effective for TTA, transformers offer an additional unsupervised learning signal through their attention mechanisms. We propose minimizing the entropy of attention distributions from the CLS token to image patches as a novel TTA objective.This approach encourages the model to attend more confidently to relevant image regions under distribution shift and is effective even when only a single test image is available. We demonstrate that attention entropy minimization improves robustness across diverse corruption types while not hurting performance on clean data on a single sample stream of images at test time.

Country of Origin
🇨🇦 Canada

Page Count
5 pages

Category
Computer Science:
CV and Pattern Recognition