Reinforcement Learning Using known Invariances
By: Alexandru Cioba , Aya Kayal , Laura Toni and more
Potential Business Impact:
Teaches robots faster by using their shape.
In many real-world reinforcement learning (RL) problems, the environment exhibits inherent symmetries that can be exploited to improve learning efficiency. This paper develops a theoretical and algorithmic framework for incorporating known group symmetries into kernel-based RL. We propose a symmetry-aware variant of optimistic least-squares value iteration (LSVI), which leverages invariant kernels to encode invariance in both rewards and transition dynamics. Our analysis establishes new bounds on the maximum information gain and covering numbers for invariant RKHSs, explicitly quantifying the sample efficiency gains from symmetry. Empirical results on a customized Frozen Lake environment and a 2D placement design problem confirm the theoretical improvements, demonstrating that symmetry-aware RL achieves significantly better performance than their standard kernel counterparts. These findings highlight the value of structural priors in designing more sample-efficient reinforcement learning algorithms.
Similar Papers
Partially Equivariant Reinforcement Learning in Symmetry-Breaking Environments
Machine Learning (CS)
Teaches robots to learn faster, even with imperfect symmetry.
Symmetry-Guided Multi-Agent Inverse Reinforcement Learnin
Robotics
Robots learn better with less practice.
Symmetry-Guided Multi-Agent Inverse Reinforcement Learning
Robotics
Robots learn better with fewer examples.