AsyncVoice Agent: Real-Time Explanation for LLM Planning and Reasoning
By: Yueqian Lin , Zhengmian Hu , Jayakumar Subramanian and more
Potential Business Impact:
Lets you talk to AI while it thinks.
Effective human-AI collaboration on complex reasoning tasks requires that users understand and interact with the model's process, not just receive an output. However, the monolithic text from methods like Chain-of-Thought (CoT) prevents this, as current interfaces lack real-time verbalization and robust user barge-in. We present AsyncVoice Agent, a system whose asynchronous architecture decouples a streaming LLM backend from a conversational voice frontend. This design allows narration and inference to run in parallel, empowering users to interrupt, query, and steer the model's reasoning process at any time. Objective benchmarks show this approach reduces interaction latency by more than 600x compared to monolithic baselines while ensuring high fidelity and competitive task accuracy. By enabling a two-way dialogue with a model's thought process, AsyncVoice Agent offers a new paradigm for building more effective, steerable, and trustworthy human-AI systems for high-stakes tasks.
Similar Papers
Asynchronous Reasoning: Training-Free Interactive Thinking LLMs
Machine Learning (CS)
Lets AI think and talk at the same time.
Spoken Conversational Agents with Large Language Models
Computation and Language
Lets computers understand and talk like people.
Adaptive Reasoning Executor: A Collaborative Agent System for Efficient Reasoning
Artificial Intelligence
Smarter AI answers questions faster, cheaper.