Learning controllable dynamics through informative exploration
By: Peter N. Loxley, Friedrich T. Sommer
Potential Business Impact:
Finds best places to learn about new worlds.
Environments with controllable dynamics are usually understood in terms of explicit models. However, such models are not always available, but may sometimes be learned by exploring an environment. In this work, we investigate using an information measure called "predicted information gain" to determine the most informative regions of an environment to explore next. Applying methods from reinforcement learning allows good suboptimal exploring policies to be found, and leads to reliable estimates of the underlying controllable dynamics. This approach is demonstrated by comparing with several myopic exploration approaches.
Similar Papers
An Optimal Policy for Learning Controllable Dynamics by Exploration
Machine Learning (CS)
Helps robots learn to make smart choices faster.
Information Gain Is Not All You Need
Robotics
Robot explores more with less travel.
Learning from Reward-Free Offline Data: A Case for Planning with Latent Dynamics Models
Machine Learning (CS)
Helps robots learn new tasks with less data.