Effects of personality steering on cooperative behavior in Large Language Model agents
By: Mizuki Sakai , Mizuki Yokoyama , Wakaba Tateishi and more
Potential Business Impact:
Makes AI more cooperative by giving it a personality.
Large language models (LLMs) are increasingly used as autonomous agents in strategic and social interactions. Although recent studies suggest that assigning personality traits to LLMs can influence their behavior, how personality steering affects cooperation under controlled conditions remains unclear. In this study, we examine the effects of personality steering on cooperative behavior in LLM agents using repeated Prisoner's Dilemma games. Based on the Big Five framework, we first measure basic personality profiles of three models, GPT-3.5-turbo, GPT-4o, and GPT-5, using the Big Five Inventory. We then compare behavior under baseline and personality-informed conditions, and further analyze the effects of independently manipulating each personality dimension to extreme values. Our results show that agreeableness is the dominant factor promoting cooperation across all models, while other personality traits have limited impact. Explicit personality information increases cooperation but can also raise vulnerability to exploitation, particularly in earlier-generation models. In contrast, later-generation models exhibit more selective cooperation. These findings indicate that personality steering acts as a behavioral bias rather than a deterministic control mechanism.
Similar Papers
Identifying Cooperative Personalities in Multi-agent Contexts through Personality Steering with Representation Engineering
Artificial Intelligence
Gives AI personalities to help them work together.
Linear Personality Probing and Steering in LLMs: A Big Five Study
Computation and Language
Changes AI personality to be more trustworthy.
NetworkGames: Simulating Cooperation in Network Games with Personality-driven LLM Agents
Physics and Society
Helps computers learn how people act together.