Optimal Watermark Generation under Type I and Type II Errors
By: Hengzhi He , Shirong Xu , Alexander Nemecek and more
Watermarking has recently emerged as a crucial tool for protecting the intellectual property of generative models and for distinguishing AI-generated content from human-generated data. Despite its practical success, most existing watermarking schemes are empirically driven and lack a theoretical understanding of the fundamental trade-off between detection power and generation fidelity. To address this gap, we formulate watermarking as a statistical hypothesis testing problem between a null distribution and its watermarked counterpart. Under explicit constraints on false-positive and false-negative rates, we derive a tight lower bound on the achievable fidelity loss, measured by a general $f$-divergence, and characterize the optimal watermarked distribution that attains this bound. We further develop a corresponding sampling rule that provides an optimal mechanism for inserting watermarks with minimal fidelity distortion. Our result establishes a simple yet broadly applicable principle linking hypothesis testing, information divergence, and watermark generation.
Similar Papers
On the Information-Theoretic Fragility of Robust Watermarking under Diffusion Editing
Cryptography and Security
Breaks hidden codes in pictures using AI.
Secure and Robust Watermarking for AI-generated Images: A Comprehensive Survey
Cryptography and Security
Marks AI pictures to show they're fake.
Optimal Detection for Language Watermarks with Pseudorandom Collision
Machine Learning (CS)
Finds hidden messages in computer writing.