Black-box Detection of LLM-generated Text Using Generalized Jensen-Shannon Divergence
By: Shuangyi Chen, Ashish Khisti
Potential Business Impact:
Finds fake writing made by computers.
We study black-box detection of machine-generated text under practical constraints: the scoring model (proxy LM) may mismatch the unknown source model, and per-input contrastive generation is costly. We propose SurpMark, a reference-based detector that summarizes a passage by the dynamics of its token surprisals. SurpMark quantizes surprisals into interpretable states, estimates a state-transition matrix for the test text, and scores it via a generalized Jensen-Shannon (GJS) gap between the test transitions and two fixed references (human vs. machine) built once from historical corpora. We prove a principled discretization criterion and establish the asymptotic normality of the decision statistic. Empirically, across multiple datasets, source models, and scenarios, SurpMark consistently matches or surpasses baselines; our experiments corroborate the statistic's asymptotic normality, and ablations validate the effectiveness of the proposed discretization.
Similar Papers
Surprisal reveals diversity gaps in image captioning and different scorers change the story
Computation and Language
Makes AI describe pictures more like people.
Zero-Shot Statistical Tests for LLM-Generated Text Detection using Finite Sample Concentration Inequalities
Machine Learning (Stat)
Finds if computers wrote text, not people.
DivScore: Zero-Shot Detection of LLM-Generated Text in Specialized Domains
Computation and Language
Finds fake doctor and lawyer writing online.