DualGuard: Dual-stream Large Language Model Watermarking Defense against Paraphrase and Spoofing Attack
By: Hao Li , Yubing Ren , Yanan Cao and more
Potential Business Impact:
Protects AI writing from being faked or changed.
With the rapid development of cloud-based services, large language models (LLMs) have become increasingly accessible through various web platforms. However, this accessibility has also led to growing risks of model abuse. LLM watermarking has emerged as an effective approach to mitigate such misuse and protect intellectual property. Existing watermarking algorithms, however, primarily focus on defending against paraphrase attacks while overlooking piggyback spoofing attacks, which can inject harmful content, compromise watermark reliability, and undermine trust in attribution. To address this limitation, we propose DualGuard, the first watermarking algorithm capable of defending against both paraphrase and spoofing attacks. DualGuard employs the adaptive dual-stream watermarking mechanism, in which two complementary watermark signals are dynamically injected based on the semantic content. This design enables DualGuard not only to detect but also to trace spoofing attacks, thereby ensuring reliable and trustworthy watermark detection. Extensive experiments conducted across multiple datasets and language models demonstrate that DualGuard achieves excellent detectability, robustness, traceability, and text quality, effectively advancing the state of LLM watermarking for real-world applications.
Similar Papers
Unified attacks to large language model watermarks: spoofing and scrubbing in unauthorized knowledge distillation
Computation and Language
Makes AI models reveal if they copied others.
Defending LLM Watermarking Against Spoofing Attacks with Contrastive Representation Learning
Cryptography and Security
Stops bad people from changing AI text meaning.
Watermarks for Embeddings-as-a-Service Large Language Models
Computation and Language
Protects AI text tools from being copied.