Beyond Context to Cognitive Appraisal: Emotion Reasoning as a Theory of Mind Benchmark for Large Language Models
By: Gerard Christopher Yeo, Kokil Jaidka
Potential Business Impact:
Helps computers understand feelings from hidden clues.
Datasets used for emotion recognition tasks typically contain overt cues that can be used in predicting the emotions expressed in a text. However, one challenge is that texts sometimes contain covert contextual cues that are rich in affective semantics, which warrant higher-order reasoning abilities to infer emotional states, not simply the emotions conveyed. This study advances beyond surface-level perceptual features to investigate how large language models (LLMs) reason about others' emotional states using contextual information, within a Theory-of-Mind (ToM) framework. Grounded in Cognitive Appraisal Theory, we curate a specialized ToM evaluation dataset1 to assess both forward reasoning - from context to emotion- and backward reasoning - from emotion to inferred context. We showed that LLMs can reason to a certain extent, although they are poor at associating situational outcomes and appraisals with specific emotions. Our work highlights the need for psychological theories in the training and evaluation of LLMs in the context of emotion reasoning.
Similar Papers
Do Machines Think Emotionally? Cognitive Appraisal Analysis of Large Language Models
Computation and Language
Teaches computers to understand feelings like people.
Beyond Classification: Towards Speech Emotion Reasoning with Multitask AudioLLMs
Computation and Language
Helps computers understand feelings in voices.
Why We Feel: Breaking Boundaries in Emotional Reasoning with Multimodal Large Language Models
Artificial Intelligence
Helps AI understand *why* people feel emotions.