Align While Search: Belief-Guided Exploratory Inference for World-Grounded Embodied Agents
By: Seohui Bae , Jeonghye Kim , Youngchul Sung and more
In this paper, we propose a test-time adaptive agent that performs exploratory inference through posterior-guided belief refinement without relying on gradient-based updates or additional training for LLM agent operating under partial observability. Our agent maintains an external structured belief over the environment state, iteratively updates it via action-conditioned observations, and selects actions by maximizing predicted information gain over the belief space. We estimate information gain using a lightweight LLM-based surrogate and assess world alignment through a novel reward that quantifies the consistency between posterior belief and ground-truth environment configuration. Experiments show that our method outperforms inference-time scaling baselines such as prompt-augmented or retrieval-enhanced LLMs, in aligning with latent world states with significantly lower integration overhead.
Similar Papers
CoBel-World: Harnessing LLM Reasoning to Build a Collaborative Belief World for Optimizing Embodied Multi-Agent Collaboration
Artificial Intelligence
Helps AI teams work together better by guessing what others think.
Agent2World: Learning to Generate Symbolic World Models via Adaptive Multi-Agent Feedback
Artificial Intelligence
Teaches computers how the world works.
Emergence: Overcoming Privileged Information Bias in Asymmetric Embodied Agents via Active Querying
Artificial Intelligence
Teaches robots to ask questions for better teamwork.