Context Representation via Action-Free Transformer encoder-decoder for Meta Reinforcement Learning
By: Amir M. Soufi Enayati, Homayoun Honari, Homayoun Najjaran
Reinforcement learning (RL) enables robots to operate in uncertain environments, but standard approaches often struggle with poor generalization to unseen tasks. Context-adaptive meta reinforcement learning addresses these limitations by conditioning on the task representation, yet they mostly rely on complete action information in the experience making task inference tightly coupled to a specific policy. This paper introduces Context Representation via Action Free Transformer encoder decoder (CRAFT), a belief model that infers task representations solely from sequences of states and rewards. By removing the dependence on actions, CRAFT decouples task inference from policy optimization, supports modular training, and leverages amortized variational inference for scalable belief updates. Built on a transformer encoder decoder with rotary positional embeddings, the model captures long range temporal dependencies and robustly encodes both parametric and non-parametric task variations. Experiments on the MetaWorld ML-10 robotic manipulation benchmark show that CRAFT achieves faster adaptation, improved generalization, and more effective exploration compared to context adaptive meta--RL baselines. These findings highlight the potential of action-free inference as a foundation for scalable RL in robotic control.
Similar Papers
Heuristic Transformer: Belief Augmented In-Context Reinforcement Learning
Machine Learning (CS)
Teaches robots to learn new tasks faster.
CRAFT: Coaching Reinforcement Learning Autonomously using Foundation Models for Multi-Robot Coordination Tasks
Robotics
Teaches robots to work together on hard jobs.
Task-Oriented Grasping Using Reinforcement Learning with a Contextual Reward Machine
Robotics
Teaches robots to grab things better and faster.