Score: 1

Exploring Persona-dependent LLM Alignment for the Moral Machine Experiment

Published: April 15, 2025 | arXiv ID: 2504.10886v1

By: Jiseon Kim , Jea Kwon , Luiz Felipe Vecchietti and more

Potential Business Impact:

AI makes different moral choices based on who it pretends to be.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Deploying large language models (LLMs) with agency in real-world applications raises critical questions about how these models will behave. In particular, how will their decisions align with humans when faced with moral dilemmas? This study examines the alignment between LLM-driven decisions and human judgment in various contexts of the moral machine experiment, including personas reflecting different sociodemographics. We find that the moral decisions of LLMs vary substantially by persona, showing greater shifts in moral decisions for critical tasks than humans. Our data also indicate an interesting partisan sorting phenomenon, where political persona predominates the direction and degree of LLM decisions. We discuss the ethical implications and risks associated with deploying these models in applications that involve moral decisions.

Country of Origin
🇰🇷 Korea, Republic of

Repos / Data Links

Page Count
13 pages

Category
Computer Science:
Computers and Society