On Decision-Making Agents and Higher-Order Causal Processes
By: Matt Wilson
Potential Business Impact:
Makes AI agents learn and remember better.
We establish a precise correspondence between decision-making agents in partially observable Markov decision processes (POMDPs) and one-input process functions, the classical limit of higher-order quantum operations. In this identification an agent's policy and memory update combine into a process function w that interacts with a POMDP environment via the link product. This suggests a dual interpretation: in the physics view, the process function acts as the environment into which local operations (agent interventions) are inserted, whereas in the AI view it encodes the agent and the inserted functions represent environments. We extend this perspective to multi-agent systems by identifying observation-independent decentralized POMDPs as natural domains for multi-input process functions.
Similar Papers
AI-Driven Optimization under Uncertainty for Mineral Processing Operations
Systems and Control
AI helps dig up needed metals faster.
Scalable Solution Methods for Dec-POMDPs with Deterministic Dynamics
Artificial Intelligence
Helps robots plan paths together without crashing.
Learning Attentive Neural Processes for Planning with Pushing Actions
Robotics
Robots learn to push blocks to exact spots.