MALLM: Multi-Agent Large Language Models Framework
By: Jonas Becker , Lars Benedikt Kaesberg , Niklas Bauer and more
Potential Business Impact:
Makes AI teams smarter by letting them discuss ideas.
Multi-agent debate (MAD) has demonstrated the ability to augment collective intelligence by scaling test-time compute and leveraging expertise. Current frameworks for multi-agent debate are often designed towards tool use, lack integrated evaluation, or provide limited configurability of agent personas, response generators, discussion paradigms, and decision protocols. We introduce MALLM (Multi-Agent Large Language Models), an open-source framework that enables systematic analysis of MAD components. MALLM offers more than 144 unique configurations of MAD, including (1) agent personas (e.g., Expert, Personality), (2) response generators (e.g., Critical, Reasoning), (3) discussion paradigms (e.g., Memory, Relay), and (4) decision protocols (e.g., Voting, Consensus). MALLM uses simple configuration files to define a debate. Furthermore, MALLM can load any textual Huggingface dataset (e.g., MMLU-Pro, WinoGrande) and provides an evaluation pipeline for easy comparison of MAD configurations. MALLM is tailored towards researchers and provides a window into the heart of multi-agent debate, facilitating the understanding of its components and their interplay.
Similar Papers
MALLM: Multi-Agent Large Language Models Framework
Multiagent Systems
Makes AI smarter by letting them argue.
iMAD: Intelligent Multi-Agent Debate for Efficient and Accurate LLM Inference
Computation and Language
Smart AI debates only when it needs to.
iMAD: Intelligent Multi-Agent Debate for Efficient and Accurate LLM Inference
Computation and Language
Smart AI debates only when it needs to improve.