Learning to Watermark: A Selective Watermarking Framework for Large Language Models via Multi-Objective Optimization
By: Chenrui Wang , Junyi Shu , Billy Chiu and more
Potential Business Impact:
Makes AI writing sound natural, not fake.
The rapid development of LLMs has raised concerns about their potential misuse, leading to various watermarking schemes that typically offer high detectability. However, existing watermarking techniques often face trade-off between watermark detectability and generated text quality. In this paper, we introduce Learning to Watermark (LTW), a novel selective watermarking framework that leverages multi-objective optimization to effectively balance these competing goals. LTW features a lightweight network that adaptively decides when to apply the watermark by analyzing sentence embeddings, token entropy, and current watermarking ratio. Training of the network involves two specifically constructed loss functions that guide the model toward Pareto-optimal solutions, thereby harmonizing watermark detectability and text quality. By integrating LTW with two baseline watermarking methods, our experimental evaluations demonstrate that LTW significantly enhances text quality without compromising detectability. Our selective watermarking approach offers a new perspective for designing watermarks for LLMs and a way to preserve high text quality for watermarks. The code is publicly available at: https://github.com/fattyray/learning-to-watermark
Similar Papers
Yet Another Watermark for Large Language Models
Cryptography and Security
Marks computer writing so you know it's real.
Yet Another Watermark for Large Language Models
Cryptography and Security
Marks AI writing so you know it's from a machine.
WaterSearch: A Quality-Aware Search-based Watermarking Framework for Large Language Models
Computation and Language
Makes computer writing sound more natural.