FaceCloak: Learning to Protect Face Templates
By: Sudipta Banerjee , Anubhav Jain , Chinmay Hegde and more
Potential Business Impact:
Hides faces so computers can't copy them.
Generative models can reconstruct face images from encoded representations (templates) bearing remarkable likeness to the original face raising security and privacy concerns. We present FaceCloak, a neural network framework that protects face templates by generating smart, renewable binary cloaks. Our method proactively thwarts inversion attacks by cloaking face templates with unique disruptors synthesized from a single face template on the fly while provably retaining biometric utility and unlinkability. Our cloaked templates can suppress sensitive attributes while generalizing to novel feature extraction schemes and outperforms leading baselines in terms of biometric matching and resiliency to reconstruction attacks. FaceCloak-based matching is extremely fast (inference time cost=0.28ms) and light-weight (0.57MB).
Similar Papers
A Deep Learning Approach for Facial Attribute Manipulation and Reconstruction in Surveillance and Reconnaissance
CV and Pattern Recognition
Makes face recognition work better for everyone.
FaceShield: Explainable Face Anti-Spoofing with Multimodal Large Language Models
CV and Pattern Recognition
Keeps face scanners safe from fake faces.
Big Brother is Watching: Proactive Deepfake Detection via Learnable Hidden Face
CV and Pattern Recognition
Hides a secret face to catch fake videos.