Integrity Shield A System for Ethical AI Use & Authorship Transparency in Assessments
By: Ashish Raj Shekhar , Shiven Agarwal , Priyanuj Bordoloi and more
Potential Business Impact:
Stops AI from cheating on school tests.
Large Language Models (LLMs) can now solve entire exams directly from uploaded PDF assessments, raising urgent concerns about academic integrity and the reliability of grades and credentials. Existing watermarking techniques either operate at the token level or assume control over the model's decoding process, making them ineffective when students query proprietary black-box systems with instructor-provided documents. We present Integrity Shield, a document-layer watermarking system that embeds schema-aware, item-level watermarks into assessment PDFs while keeping their human-visible appearance unchanged. These watermarks consistently prevent MLLMs from answering shielded exam PDFs and encode stable, item-level signatures that can be reliably recovered from model or student responses. Across 30 exams spanning STEM, humanities, and medical reasoning, Integrity Shield achieves exceptionally high prevention (91-94% exam-level blocking) and strong detection reliability (89-93% signature retrieval) across four commercial MLLMs. Our demo showcases an interactive interface where instructors upload an exam, preview watermark behavior, and inspect pre/post AI performance & authorship evidence.
Similar Papers
EditMark: Watermarking Large Language Models based on Model Editing
Cryptography and Security
Marks AI writing to prove it's yours.
StealthInk: A Multi-bit and Stealthy Watermark for Large Language Models
Cryptography and Security
Marks AI writing so you know who wrote it.
Yet Another Watermark for Large Language Models
Cryptography and Security
Marks computer writing so you know it's real.