Opinion Consensus Formation Among Networked Large Language Models
By: Iris Yazici , Mert Kayaalp , Stefan Taga and more
Potential Business Impact:
AI groups agree, but not how we expect.
Can classical consensus models predict the group behavior of large language models (LLMs)? We examine multi-round interactions among LLM agents through the DeGroot framework, where agents exchange text-based messages over diverse communication graphs. To track opinion evolution, we map each message to an opinion score via sentiment analysis. We find that agents typically reach consensus and the disagreement between the agents decays exponentially. However, the limiting opinion departs from DeGroot's network-centrality-weighted forecast. The consensus between LLM agents turns out to be largely insensitive to initial conditions and instead depends strongly on the discussion subject and inherent biases. Nevertheless, transient dynamics align with classical graph theory and the convergence rate of opinions is closely related to the second-largest eigenvalue of the graph's combination matrix. Together, these findings can be useful for LLM-driven social-network simulations and the design of resource-efficient multi-agent LLM applications.
Similar Papers
Consensus Is All You Need: Gossip-Based Reasoning Among Large Language Models
Multiagent Systems
AI models work together to give better answers.
Emergence of human-like polarization among large language model agents
Social and Information Networks
AI agents create echo chambers, like people.
Towards Simulating Social Influence Dynamics with LLM-based Multi-agents
Multiagent Systems
Computers can now act like people talking online.