Inferring Causal Graph Temporal Logic Formulas to Expedite Reinforcement Learning in Temporally Extended Tasks
By: Hadi Partovi Aria, Zhe Xu
Potential Business Impact:
Learns how changes spread to make smart decisions.
Decision-making tasks often unfold on graphs with spatial-temporal dynamics. Black-box reinforcement learning often overlooks how local changes spread through network structure, limiting sample efficiency and interpretability. We present GTL-CIRL, a closed-loop framework that simultaneously learns policies and mines Causal Graph Temporal Logic (Causal GTL) specifications. The method shapes rewards with robustness, collects counterexamples when effects fail, and uses Gaussian Process (GP) driven Bayesian optimization to refine parameterized cause templates. The GP models capture spatial and temporal correlations in the system dynamics, enabling efficient exploration of complex parameter spaces. Case studies in gene and power networks show faster learning and clearer, verifiable behavior compared to standard RL baselines.
Similar Papers
Expediting Reinforcement Learning by Incorporating Knowledge About Temporal Causality in the Environment
Machine Learning (CS)
Teaches robots to learn tasks faster and better.
Graph Contextual Reinforcement Learning for Efficient Directed Controller Synthesis
Artificial Intelligence
Helps robots learn faster by remembering past tries.
Unifying Causal Reinforcement Learning: Survey, Taxonomy, Algorithms and Applications
Artificial Intelligence
Makes smart computers learn better and explain why.