Graph Contextual Reinforcement Learning for Efficient Directed Controller Synthesis
By: Toshihide Ubukata , Enhong Mu , Takuto Yamauchi and more
Controller synthesis is a formal method approach for automatically generating Labeled Transition System (LTS) controllers that satisfy specified properties. The efficiency of the synthesis process, however, is critically dependent on exploration policies. These policies often rely on fixed rules or strategies learned through reinforcement learning (RL) that consider only a limited set of current features. To address this limitation, this paper introduces GCRL, an approach that enhances RL-based methods by integrating Graph Neural Networks (GNNs). GCRL encodes the history of LTS exploration into a graph structure, allowing it to capture a broader, non-current-based context. In a comparative experiment against state-of-the-art methods, GCRL exhibited superior learning efficiency and generalization across four out of five benchmark domains, except one particular domain characterized by high symmetry and strictly local interactions.
Similar Papers
Power Grid Control with Graph-Based Distributed Reinforcement Learning
Machine Learning (CS)
Helps power grids run better with smart computers.
Goal Reaching with Eikonal-Constrained Hierarchical Quasimetric Reinforcement Learning
Machine Learning (CS)
Teaches robots to reach goals without mistakes.
GoRL: An Algorithm-Agnostic Framework for Online Reinforcement Learning with Generative Policies
Machine Learning (CS)
Lets robots learn complex moves safely and quickly.