Score: 0

Inferring Causal Graph Temporal Logic Formulas to Expedite Reinforcement Learning in Temporally Extended Tasks

Published: January 6, 2026 | arXiv ID: 2601.02666v1

By: Hadi Partovi Aria, Zhe Xu

Potential Business Impact:

Learns how changes spread to make smart decisions.

Business Areas:
Machine Learning Artificial Intelligence, Data and Analytics, Software

Decision-making tasks often unfold on graphs with spatial-temporal dynamics. Black-box reinforcement learning often overlooks how local changes spread through network structure, limiting sample efficiency and interpretability. We present GTL-CIRL, a closed-loop framework that simultaneously learns policies and mines Causal Graph Temporal Logic (Causal GTL) specifications. The method shapes rewards with robustness, collects counterexamples when effects fail, and uses Gaussian Process (GP) driven Bayesian optimization to refine parameterized cause templates. The GP models capture spatial and temporal correlations in the system dynamics, enabling efficient exploration of complex parameter spaces. Case studies in gene and power networks show faster learning and clearer, verifiable behavior compared to standard RL baselines.

Country of Origin
🇺🇸 United States

Page Count
4 pages

Category
Computer Science:
Artificial Intelligence