Zero-Knowledge Proof Based Verifiable Inference of Models
By: Yunxiao Wang
Potential Business Impact:
Lets you check AI answers without seeing its secrets.
Recent advances in artificial intelligence (AI), particularly deep learning, have led to widespread adoption across various applications. Yet, a fundamental challenge persists: how can we verify the correctness of AI model inference when model owners cannot (or will not) reveal their parameters? These parameters represent enormous training costs and valuable intellectual property, making transparent verification difficult. In this paper, we introduce a zero-knowledge framework capable of verifying deep learning inference without exposing model internal parameters. Built on recursively composed zero-knowledge proofs and requiring no trusted setup, our framework supports both linear and nonlinear neural network layers, including matrix multiplication, normalization, softmax, and SiLU. Leveraging the Fiat-Shamir heuristic, we obtain a succinct non-interactive argument of knowledge (zkSNARK) with constant-size proofs. To demonstrate the practicality of our approach, we translate the DeepSeek model into a fully SNARK-verifiable version named ZK-DeepSeek and show experimentally that our framework delivers both efficiency and flexibility in real-world AI verification workloads.
Similar Papers
Enhancing Trust in AI Marketplaces: Evaluating On-Chain Verification of Personalized AI models using zk-SNARKs
Cryptography and Security
Lets computers prove AI models work safely.
TeleSparse: Practical Privacy-Preserving Verification of Deep Neural Networks
Machine Learning (CS)
Checks AI decisions without seeing private data.
JSTprove: Pioneering Verifiable AI for a Trustless Future
Cryptography and Security
Lets AI prove it's right without showing secrets.