Data-Driven Abstraction and Synthesis for Stochastic Systems with Unknown Dynamics
By: Mahdi Nazeri , Thom Badings , Anne-Kathrin Schmuck and more
Potential Business Impact:
Teaches robots to learn and follow rules.
We study the automated abstraction-based synthesis of correct-by-construction control policies for stochastic dynamical systems with unknown dynamics. Our approach is to learn an abstraction from sampled data, which is represented in the form of a finite Markov decision process (MDP). In this paper, we present a data-driven technique for constructing finite-state interval MDP (IMDP) abstractions of stochastic systems with unknown nonlinear dynamics. As a distinguishing and novel feature, our technique only requires (1) noisy state-input-state observations and (2) an upper bound on the system's Lipschitz constant. Combined with standard model-checking techniques, our IMDP abstractions enable the synthesis of policies that satisfy probabilistic temporal properties (such as "reach-while-avoid") with a predefined confidence. Our experimental results show the effectiveness and robustness of our approach.
Similar Papers
Data-Driven Yet Formal Policy Synthesis for Stochastic Nonlinear Dynamical Systems
Systems and Control
Teaches robots to control tricky machines reliably.
Probabilistic Alternating Simulations for Policy Synthesis in Uncertain Stochastic Dynamical Systems
Systems and Control
Makes robots handle tricky, unpredictable moves.
Rollout-Based Approximate Dynamic Programming for MDPs with Information-Theoretic Constraints
Systems and Control
Helps computers make better choices with less data.