Robust Deep Reinforcement Learning in Robotics via Adaptive Gradient-Masked Adversarial Attacks
By: Zongyuan Zhang , Tianyang Duan , Zheng Lin and more
Potential Business Impact:
Tricks robots into making bad choices.
Deep reinforcement learning (DRL) has emerged as a promising approach for robotic control, but its realworld deployment remains challenging due to its vulnerability to environmental perturbations. Existing white-box adversarial attack methods, adapted from supervised learning, fail to effectively target DRL agents as they overlook temporal dynamics and indiscriminately perturb all state dimensions, limiting their impact on long-term rewards. To address these challenges, we propose the Adaptive Gradient-Masked Reinforcement (AGMR) Attack, a white-box attack method that combines DRL with a gradient-based soft masking mechanism to dynamically identify critical state dimensions and optimize adversarial policies. AGMR selectively allocates perturbations to the most impactful state features and incorporates a dynamic adjustment mechanism to balance exploration and exploitation during training. Extensive experiments demonstrate that AGMR outperforms state-of-the-art adversarial attack methods in degrading the performance of the victim agent and enhances the victim agent's robustness through adversarial defense mechanisms.
Similar Papers
State-Aware Perturbation Optimization for Robust Deep Reinforcement Learning
Machine Learning (CS)
Makes robots safer by fooling them with tricky inputs.
Towards Robust Deep Reinforcement Learning against Environmental State Perturbation
Machine Learning (CS)
Makes robots learn better even when things change.
Adversarial Agents: Black-Box Evasion Attacks with Reinforcement Learning
Cryptography and Security
Teaches computers to trick other computers.