Decoding Deception: Understanding Automatic Speech Recognition Vulnerabilities in Evasion and Poisoning Attacks
By: Aravindhan G, Yuvaraj Govindarajulu, Parin Shah
Potential Business Impact:
Makes voice assistants hear wrong words easily.
Recent studies have demonstrated the vulnerability of Automatic Speech Recognition systems to adversarial examples, which can deceive these systems into misinterpreting input speech commands. While previous research has primarily focused on white-box attacks with constrained optimizations, and transferability based black-box attacks against commercial Automatic Speech Recognition devices, this paper explores cost efficient white-box attack and non transferability black-box adversarial attacks on Automatic Speech Recognition systems, drawing insights from approaches such as Fast Gradient Sign Method and Zeroth-Order Optimization. Further, the novelty of the paper includes how poisoning attack can degrade the performances of state-of-the-art models leading to misinterpretation of audio signals. Through experimentation and analysis, we illustrate how hybrid models can generate subtle yet impactful adversarial examples with very little perturbation having Signal Noise Ratio of 35dB that can be generated within a minute. These vulnerabilities of state-of-the-art open source model have practical security implications, and emphasize the need for adversarial security.
Similar Papers
Are Modern Speech Enhancement Systems Vulnerable to Adversarial Attacks?
Audio and Speech Processing
Makes voices say different things with hidden sounds.
Exploiting Vulnerabilities in Speech Translation Systems through Targeted Adversarial Attacks
Sound
Tricks voice translators to say wrong things.
Impact of Phonetics on Speaker Identity in Adversarial Voice Attack
Sound
Makes voice assistants understand speech better.