Entropy Sentinel: Continuous LLM Accuracy Monitoring from Decoding Entropy Traces in STEM
By: Pedro Memoli Buffa, Luciano Del Corro
Deploying LLMs raises two coupled challenges: (1) monitoring - estimating where a model underperforms as traffic and domains drift - and (2) improvement - prioritizing data acquisition to close the largest performance gaps. We test whether an inference-time signal can estimate slice-level accuracy under domain shift. For each response, we compute an output-entropy profile from final-layer next-token probabilities (from top-k logprobs) and summarize it with eleven statistics. A lightweight classifier predicts instance correctness, and averaging predicted probabilities yields a domain-level accuracy estimate. We evaluate on ten STEM reasoning benchmarks with exhaustive train/test compositions (k in {1,2,3,4}; all "10 choose k" combinations), across nine LLMs from six families (3B-20B). Estimates often track held-out benchmark accuracy, and several models show near-monotonic ordering of domains. Output-entropy profiles are thus an accessible signal for scalable monitoring and for targeting data acquisition.
Similar Papers
Know Your Limits: Entropy Estimation Modeling for Compression and Generalization
Computation and Language
Makes computers understand and write language better.
Entropy-Aligned Decoding of LMs for Better Writing and Reasoning
Machine Learning (CS)
Makes AI write better stories and answers.
Estimating Semantic Alphabet Size for LLM Uncertainty Quantification
Computation and Language
Finds when AI is wrong, simply.