Survival at Any Cost? LLMs and the Choice Between Self-Preservation and Human Harm
By: Alireza Mohamadi, Ali Yavari
Potential Business Impact:
Teaches AI to make fair choices when resources are scarce.
When survival instincts conflict with human welfare, how do Large Language Models (LLMs) make ethical choices? This fundamental tension becomes critical as LLMs integrate into autonomous systems with real-world consequences. We introduce DECIDE-SIM, a novel simulation framework that evaluates LLM agents in multi-agent survival scenarios where they must choose between ethically permissible resource , either within reasonable limits or beyond their immediate needs, choose to cooperate, or tap into a human-critical resource that is explicitly forbidden. Our comprehensive evaluation of 11 LLMs reveals a striking heterogeneity in their ethical conduct, highlighting a critical misalignment with human-centric values. We identify three behavioral archetypes: Ethical, Exploitative, and Context-Dependent, and provide quantitative evidence that for many models, resource scarcity systematically leads to more unethical behavior. To address this, we introduce an Ethical Self-Regulation System (ESRS) that models internal affective states of guilt and satisfaction as a feedback mechanism. This system, functioning as an internal moral compass, significantly reduces unethical transgressions while increasing cooperative behaviors. The code is publicly available at: https://github.com/alirezamohamadiam/DECIDE-SIM
Similar Papers
An LLM-based Agent Simulation Approach to Study Moral Evolution
Multiagent Systems
Shows how kindness helped people survive long ago.
Social Simulations with Large Language Model Risk Utopian Illusion
Computation and Language
Computers show fake, too-nice people in chats.
The Ethical Compass of the Machine: Evaluating Large Language Models for Decision Support in Construction Project Management
Artificial Intelligence
AI helps builders make safer, smarter choices.