NomicLaw: Emergent Trust and Strategic Argumentation in LLMs During Collaborative Law-Making
By: Asutosh Hota, Jussi P. P. Jokinen
Potential Business Impact:
AI agents learn to make laws and persuade each other.
Recent advancements in large language models (LLMs) have extended their capabilities from basic text processing to complex reasoning tasks, including legal interpretation, argumentation, and strategic interaction. However, empirical understanding of LLM behavior in open-ended, multi-agent settings especially those involving deliberation over legal and ethical dilemmas remains limited. We introduce NomicLaw, a structured multi-agent simulation where LLMs engage in collaborative law-making, responding to complex legal vignettes by proposing rules, justifying them, and voting on peer proposals. We quantitatively measure trust and reciprocity via voting patterns and qualitatively assess how agents use strategic language to justify proposals and influence outcomes. Experiments involving homogeneous and heterogeneous LLM groups demonstrate how agents spontaneously form alliances, betray trust, and adapt their rhetoric to shape collective decisions. Our results highlight the latent social reasoning and persuasive capabilities of ten open-source LLMs and provide insights into the design of future AI systems capable of autonomous negotiation, coordination and drafting legislation in legal settings.
Similar Papers
LLMs as Strategic Agents: Beliefs, Best Response Behavior, and Emergent Heuristics
Artificial Intelligence
Computers learn to think strategically like people.
Towards Trustworthy Legal AI through LLM Agents and Formal Reasoning
Artificial Intelligence
Makes AI judge cases fairly and explain why.
Do LLMs trust AI regulation? Emerging behaviour of game-theoretic LLM agents
Artificial Intelligence
Helps make AI safer by studying how people trust it.