An Empirical Study on the Effectiveness of Incorporating Offline RL As Online RL Subroutines
By: Jianhai Su, Jinzhu Luo, Qi Zhang
Potential Business Impact:
Teaches robots to learn faster from past mistakes.
We take the novel perspective of incorporating offline RL algorithms as subroutines of tabula rasa online RL. This is feasible because an online learning agent can repurpose its historical interactions as offline dataset. We formalize this idea into a framework that accommodates several variants of offline RL incorporation such as final policy recommendation and online fine-tuning. We further introduce convenient techniques to improve its effectiveness in enhancing online learning efficiency. Our extensive and systematic empirical analyses show that 1) the effectiveness of the proposed framework depends strongly on the nature of the task, 2) our proposed techniques greatly enhance its effectiveness, and 3) existing online fine-tuning methods are overall ineffective, calling for more research therein.
Similar Papers
A Tutorial: An Intuitive Explanation of Offline Reinforcement Learning Theory
Machine Learning (CS)
Teaches computers to learn from old data.
Using Non-Expert Data to Robustify Imitation Learning via Offline Reinforcement Learning
Robotics
Teaches robots to learn from bad examples.
Using Non-Expert Data to Robustify Imitation Learning via Offline Reinforcement Learning
Robotics
Teaches robots to learn from mistakes and play.