Vulnerabilities of Audio-Based Biometric Authentication Systems Against Deepfake Speech Synthesis
By: Mengze Hong , Di Jiang , Zeying Xie and more
Potential Business Impact:
Fake voices trick voice locks easily.
As audio deepfakes transition from research artifacts to widely available commercial tools, robust biometric authentication faces pressing security threats in high-stakes industries. This paper presents a systematic empirical evaluation of state-of-the-art speaker authentication systems based on a large-scale speech synthesis dataset, revealing two major security vulnerabilities: 1) modern voice cloning models trained on very small samples can easily bypass commercial speaker verification systems; and 2) anti-spoofing detectors struggle to generalize across different methods of audio synthesis, leading to a significant gap between in-domain performance and real-world robustness. These findings call for a reconsideration of security measures and stress the need for architectural innovations, adaptive defenses, and the transition towards multi-factor authentication.
Similar Papers
Physics-Guided Deepfake Detection for Voice Authentication Systems
Sound
Protects voice locks from fake voices and hacking.
A Survey of Threats Against Voice Authentication and Anti-Spoofing Systems
Cryptography and Security
Makes voice locks safer from fake voices.
A Survey of Threats Against Voice Authentication and Anti-Spoofing Systems
Cryptography and Security
Makes voice locks harder to trick.