Frictional Q-Learning
By: Hyunwoo Kim, Hyo Kyung Lee
Potential Business Impact:
Teaches robots to learn new skills safely.
We draw an analogy between static friction in classical mechanics and extrapolation error in off-policy RL, and use it to formulate a constraint that prevents the policy from drifting toward unsupported actions. In this study, we present Frictional Q-learning, a deep reinforcement learning algorithm for continuous control, which extends batch-constrained reinforcement learning. Our algorithm constrains the agent's action space to encourage behavior similar to that in the replay buffer, while maintaining a distance from the manifold of the orthonormal action space. The constraint preserves the simplicity of batch-constrained, and provides an intuitive physical interpretation of extrapolation error. Empirically, we further demonstrate that our algorithm is robustly trained and achieves competitive performance across standard continuous control benchmarks.
Similar Papers
Automatic Reward Shaping from Confounded Offline Data
Artificial Intelligence
Makes AI learn safely from bad past game experiences.
Imagination-Limited Q-Learning for Offline Reinforcement Learning
Machine Learning (CS)
Teaches robots to learn from past mistakes.
Adaptive Neighborhood-Constrained Q Learning for Offline Reinforcement Learning
Machine Learning (CS)
Helps robots learn from past mistakes safely.