When AI Settles Down: Late-Stage Stability as a Signature of AI-Generated Text Detection
By: Ke Sun , Guangsheng Bao , Han Cui and more
Potential Business Impact:
Finds fake writing by spotting AI's predictable patterns.
Zero-shot detection methods for AI-generated text typically aggregate token-level statistics across entire sequences, overlooking the temporal dynamics inherent to autoregressive generation. We analyze over 120k text samples and reveal Late-Stage Volatility Decay: AI-generated text exhibits rapidly stabilizing log probability fluctuations as generation progresses, while human writing maintains higher variability throughout. This divergence peaks in the second half of sequences, where AI-generated text shows 24--32\% lower volatility. Based on this finding, we propose two simple features: Derivative Dispersion and Local Volatility, which computed exclusively from late-stage statistics. Without perturbation sampling or additional model access, our method achieves state-of-the-art performance on EvoBench and MAGE benchmarks and demonstrates strong complementarity with existing global methods.
Similar Papers
AI-Generated Text is Non-Stationary: Detection via Temporal Tomography
Computation and Language
Finds fake writing by looking at how it changes.
Model-Agnostic Sentiment Distribution Stability Analysis for Robust LLM-Generated Texts Detection
Computation and Language
Finds fake writing by checking emotions.
Can You Detect the Difference?
Computation and Language
Finds AI writing that tricks human detectors.