Emotional Cognitive Modeling Framework with Desire-Driven Objective Optimization for LLM-empowered Agent in Social Simulation
By: Qun Ma , Xiao Xue , Xuwen Zhang and more
Potential Business Impact:
Makes computer characters act more like real people.
The advent of large language models (LLMs) has enabled agents to represent virtual humans in societal simulations, facilitating diverse interactions within complex social systems. However, existing LLM-based agents exhibit severe limitations in affective cognition: They fail to simulate the bounded rationality essential for bridging virtual and real-world services; They lack empirically validated integration mechanisms embedding emotions within agent decision architectures. This paper constructs an emotional cognition framework incorporating desire generation and objective management, designed to achieve emotion alignment between LLM-based agents and humans, modeling the complete decision-making process of LLM-based agents, encompassing state evolution, desire generation, objective optimization, decision generation, and action execution. This study implements the proposed framework within our proprietary multi-agent interaction environment. Experimental results demonstrate that agents governed by our framework not only exhibit behaviors congruent with their emotional states but also, in comparative assessments against other agent types, demonstrate superior ecological validity and generate decision outcomes that significantly more closely approximate human behavioral patterns.
Similar Papers
Simulating Human-Like Learning Dynamics with LLM-Empowered Agents
Artificial Intelligence
Shows how AI learns like people, good and bad.
Modeling Layered Consciousness with Multi-Agent Large Language Models
Computation and Language
Makes AI understand feelings and act more human.
EvoEmo: Towards Evolved Emotional Policies for LLM Agents in Multi-Turn Negotiation
Artificial Intelligence
Makes AI better at talking and getting what it wants.