Moderate Actor-Critic Methods: Controlling Overestimation Bias via Expectile Loss
By: Ukjo Hwang, Songnam Hong
Potential Business Impact:
Fixes computer learning mistakes for better results.
Overestimation is a fundamental characteristic of model-free reinforcement learning (MF-RL), arising from the principles of temporal difference learning and the approximation of the Q-function. To address this challenge, we propose a novel moderate target in the Q-function update, formulated as a convex optimization of an overestimated Q-function and its lower bound. Our primary contribution lies in the efficient estimation of this lower bound through the lower expectile of the Q-value distribution conditioned on a state. Notably, our moderate target integrates seamlessly into state-of-the-art (SOTA) MF-RL algorithms, including Deep Deterministic Policy Gradient (DDPG) and Soft Actor Critic (SAC). Experimental results validate the effectiveness of our moderate target in mitigating overestimation bias in DDPG, SAC, and distributional RL algorithms.
Similar Papers
Mitigating Estimation Bias with Representation Learning in TD Error-Driven Regularization
Machine Learning (CS)
Teaches robots to learn better by balancing risks.
Efficient $Q$-Learning and Actor-Critic Methods for Robust Average Reward Reinforcement Learning
Machine Learning (CS)
Teaches computers to make good choices even with bad info.
Risk-sensitive Reinforcement Learning Based on Convex Scoring Functions
Mathematical Finance
Teaches computers to trade money safely and smartly.