Score: 1

AWM: Accurate Weight-Matrix Fingerprint for Large Language Models

Published: October 8, 2025 | arXiv ID: 2510.06738v1

By: Boyi Zeng , Lin Chen , Ziwei He and more

Potential Business Impact:

Finds if AI was copied from another AI.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Protecting the intellectual property of large language models (LLMs) is crucial, given the substantial resources required for their training. Consequently, there is an urgent need for both model owners and third parties to determine whether a suspect LLM is trained from scratch or derived from an existing base model. However, the intensive post-training processes that models typically undergo-such as supervised fine-tuning, extensive continued pretraining, reinforcement learning, multi-modal extension, pruning, and upcycling-pose significant challenges to reliable identification. In this work, we propose a training-free fingerprinting method based on weight matrices. We leverage the Linear Assignment Problem (LAP) and an unbiased Centered Kernel Alignment (CKA) similarity to neutralize the effects of parameter manipulations, yielding a highly robust and high-fidelity similarity metric. On a comprehensive testbed of 60 positive and 90 negative model pairs, our method demonstrates exceptional robustness against all six aforementioned post-training categories while exhibiting a near-zero risk of false positives. By achieving perfect scores on all classification metrics, our approach establishes a strong basis for reliable model lineage verification. Moreover, the entire computation completes within 30s on an NVIDIA 3090 GPU. The code is available at https://github.com/LUMIA-Group/AWM.

Country of Origin
🇨🇳 China

Repos / Data Links

Page Count
23 pages

Category
Computer Science:
Computation and Language