The Impact of On-Policy Parallelized Data Collection on Deep Reinforcement Learning Networks
By: Walter Mayor , Johan Obando-Ceron , Aaron Courville and more
Potential Business Impact:
Makes robots learn faster by collecting more data.
The use of parallel actors for data collection has been an effective technique used in reinforcement learning (RL) algorithms. The manner in which data is collected in these algorithms, controlled via the number of parallel environments and the rollout length, induces a form of bias-variance trade-off; the number of training passes over the collected data, on the other hand, must strike a balance between sample efficiency and overfitting. We conduct an empirical analysis of these trade-offs on PPO, one of the most popular RL algorithms that uses parallel actors, and establish connections to network plasticity and, more generally, optimization stability. We examine its impact on network architectures, as well as the hyper-parameter sensitivity when scaling data. Our analyses indicate that larger dataset sizes can increase final performance across a variety of settings, and that scaling parallel environments is more effective than increasing rollout lengths. These findings highlight the critical role of data collection strategies in improving agent performance.
Similar Papers
Enhancing Diversity in Parallel Agents: A Maximum State Entropy Exploration Story
Machine Learning (CS)
Makes AI learn faster by having different helpers.
Staggered Environment Resets Improve Massively Parallel On-Policy Reinforcement Learning
Machine Learning (CS)
Makes robots learn faster and better.
Offline vs. Online Learning in Model-based RL: Lessons for Data Collection Strategies
Machine Learning (CS)
Helps robots learn better by mixing old and new experiences.