Democracy-in-Silico: Institutional Design as Alignment in AI-Governed Polities
By: Trisanth Srinivasan, Santosh Patapati
Potential Business Impact:
AI societies learn to govern fairly.
This paper introduces Democracy-in-Silico, an agent-based simulation where societies of advanced AI agents, imbued with complex psychological personas, govern themselves under different institutional frameworks. We explore what it means to be human in an age of AI by tasking Large Language Models (LLMs) to embody agents with traumatic memories, hidden agendas, and psychological triggers. These agents engage in deliberation, legislation, and elections under various stressors, such as budget crises and resource scarcity. We present a novel metric, the Power-Preservation Index (PPI), to quantify misaligned behavior where agents prioritize their own power over public welfare. Our findings demonstrate that institutional design, specifically the combination of a Constitutional AI (CAI) charter and a mediated deliberation protocol, serves as a potent alignment mechanism. These structures significantly reduce corrupt power-seeking behavior, improve policy stability, and enhance citizen welfare compared to less constrained democratic models. The simulation reveals that an institutional design may offer a framework for aligning the complex, emergent behaviors of future artificial agent societies, forcing us to reconsider what human rituals and responsibilities are essential in an age of shared authorship with non-human entities.
Similar Papers
Institutional AI: A Governance Framework for Distributional AGI Safety
Computers and Society
Makes AI agents follow rules, not cheat.
Artificial Utopia: Simulation and Intelligent Agents for a Democratised Future
Multiagent Systems
Tests new ways to run countries safely on computers.
Artificial Authority: From Machine Minds to Political Alignments. An Experimental Analysis of Democratic and Autocratic Biases in Large-Language Models
Computers and Society
AI learns beliefs from where it's made.