Large Language Model-Based Reward Design for Deep Reinforcement Learning-Driven Autonomous Cyber Defense
By: Sayak Mukherjee , Samrat Chatterjee , Emilie Purvine and more
Potential Business Impact:
Teaches computers to defend against cyberattacks.
Designing rewards for autonomous cyber attack and defense learning agents in a complex, dynamic environment is a challenging task for subject matter experts. We propose a large language model (LLM)-based reward design approach to generate autonomous cyber defense policies in a deep reinforcement learning (DRL)-driven experimental simulation environment. Multiple attack and defense agent personas were crafted, reflecting heterogeneity in agent actions, to generate LLM-guided reward designs where the LLM was first provided with contextual cyber simulation environment information. These reward structures were then utilized within a DRL-driven attack-defense simulation environment to learn an ensemble of cyber defense policies. Our results suggest that LLM-guided reward designs can lead to effective defense strategies against diverse adversarial behaviors.
Similar Papers
Large Language Model Integration with Reinforcement Learning to Augment Decision-Making in Autonomous Cyber Operations
Cryptography and Security
Teaches computers to fight cyber threats faster.
Large Language Models are Autonomous Cyber Defenders
Artificial Intelligence
Helps AI teams fight computer attackers together.
Agent-R1: Training Powerful LLM Agents with End-to-End Reinforcement Learning
Computation and Language
Teaches AI to learn and solve problems better.