Single LLM Debate, MoLaCE: Mixture of Latent Concept Experts Against Confirmation Bias
By: Hazel Kim, Philip Torr
Potential Business Impact:
Helps AI think for itself, not just agree.
Large language models (LLMs) are highly vulnerable to input confirmation bias. When a prompt implies a preferred answer, models often reinforce that bias rather than explore alternatives. This phenomenon remains underexplored, yet it is already harmful in base models and poses an even greater risk in multi-agent debate, where echo chambers reinforce bias instead of correction. We introduce Mixture of Latent Concept Experts (MoLaCE), a lightweight inference-time framework that addresses confirmation bias by mixing experts instantiated as different activation strengths over latent concepts that shape model responses. Our key insight is that, due to the compositional nature of language, differently phrased prompts reweight latent concepts in prompt-specific ways that affect factual correctness, so no single fixed intervention can be applied universally across inputs. This design enables a single LLM to emulate the benefits of debate internally while remaining computationally efficient and scalable. It can also be integrated into multi-agent debate frameworks to diversify perspectives and reduce correlated errors. We empirically show that it consistently reduces confirmation bias, improves robustness, and matches or surpasses multi-agent debate while requiring only a fraction of the computation.
Similar Papers
Multiple LLM Agents Debate for Equitable Cultural Alignment
Computation and Language
Helps AI understand different cultures better.
Multi-Agent Debate for LLM Judges with Adaptive Stability Detection
Artificial Intelligence
Debating computers make better judgments than voting ones.
Experts are all you need: A Composable Framework for Large Language Model Inference
Machine Learning (CS)
Makes AI smarter and faster by teamwork.