Environment Agnostic Goal-Conditioning, A Study of Reward-Free Autonomous Learning
By: Hampus Åström, Elin Anna Topp, Jacek Malec
Potential Business Impact:
Lets robots learn any task without rewards.
In this paper we study how transforming regular reinforcement learning environments into goal-conditioned environments can let agents learn to solve tasks autonomously and reward-free. We show that an agent can learn to solve tasks by selecting its own goals in an environment-agnostic way, at training times comparable to externally guided reinforcement learning. Our method is independent of the underlying off-policy learning algorithm. Since our method is environment-agnostic, the agent does not value any goals higher than others, leading to instability in performance for individual goals. However, in our experiments, we show that the average goal success rate improves and stabilizes. An agent trained with this method can be instructed to seek any observations made in the environment, enabling generic training of agents prior to specific use cases.
Similar Papers
Why Goal-Conditioned Reinforcement Learning Works: Relation to Dual Control
Machine Learning (CS)
Teaches robots to reach any goal.
Autonomous Learning From Success and Failure: Goal-Conditioned Supervised Learning with Negative Feedback
Machine Learning (CS)
Helps robots learn from mistakes, not just wins.
Self-Supervised Goal-Reaching Results in Multi-Agent Cooperation and Exploration
Machine Learning (CS)
Robots learn to work together to reach goals.