Theory of Mind in Large Language Models: Assessment and Enhancement
By: Ruirui Chen , Weifeng Jiang , Chengwei Qin and more
Potential Business Impact:
Helps computers understand what people are thinking.
Theory of Mind (ToM)-the ability to reason about the mental states of oneself and others-is a cornerstone of human social intelligence. As Large Language Models (LLMs) become increasingly integrated into daily life, understanding their ability to interpret and respond to human mental states is crucial for enabling effective interactions. In this paper, we review LLMs' ToM capabilities by analyzing both evaluation benchmarks and enhancement strategies. For evaluation, we focus on recently proposed and widely used story-based benchmarks. For enhancement, we provide an in-depth analysis of recent methods aimed at improving LLMs' ToM abilities. Furthermore, we outline promising directions for future research to further advance these capabilities and better adapt LLMs to more realistic and diverse scenarios. Our survey serves as a valuable resource for researchers interested in evaluating and advancing LLMs' ToM capabilities.
Similar Papers
XToM: Exploring the Multilingual Theory of Mind for Large Language Models
Computation and Language
Computers understand feelings in different languages.
Do Theory of Mind Benchmarks Need Explicit Human-like Reasoning in Language Models?
Computation and Language
Computers can guess what others think, but maybe not really.
Large Language Models as Theory of Mind Aware Generative Agents with Counterfactual Reflection
Computation and Language
Lets computers understand what people think and feel.