Exploration Through Introspection: A Self-Aware Reward Model
By: Michael Petrowski, Milica Gašić
Potential Business Impact:
AI learns to feel "pain" to learn better.
Understanding how artificial agents model internal mental states is central to advancing Theory of Mind in AI. Evidence points to a unified system for self- and other-awareness. We explore this self-awareness by having reinforcement learning agents infer their own internal states in gridworld environments. Specifically, we introduce an introspective exploration component that is inspired by biological pain as a learning signal by utilizing a hidden Markov model to infer "pain-belief" from online observations. This signal is integrated into a subjective reward function to study how self-awareness affects the agent's learning abilities. Further, we use this computational framework to investigate the difference in performance between normal and chronic pain perception models. Results show that introspective agents in general significantly outperform standard baseline agents and can replicate complex human-like behaviors.
Similar Papers
Learning To Explore With Predictive World Model Via Self-Supervised Learning
Machine Learning (CS)
Teaches robots to learn games by themselves.
Agent Learning via Early Experience
Artificial Intelligence
Computers learn better by trying things themselves.
Agent Learning via Early Experience
Artificial Intelligence
Computers learn to do better by trying things.