Emotions Where Art Thou: Understanding and Characterizing the Emotional Latent Space of Large Language Models
By: Benjamin Reichman, Adar Avsian, Larry Heck
Potential Business Impact:
Teaches computers to understand and change feelings.
This work investigates how large language models (LLMs) internally represent emotion by analyzing the geometry of their hidden-state space. The paper identifies a low-dimensional emotional manifold and shows that emotional representations are directionally encoded, distributed across layers, and aligned with interpretable dimensions. These structures are stable across depth and generalize to eight real-world emotion datasets spanning five languages. Cross-domain alignment yields low error and strong linear probe performance, indicating a universal emotional subspace. Within this space, internal emotion perception can be steered while preserving semantics using a learned intervention module, with especially strong control for basic emotions across languages. These findings reveal a consistent and manipulable affective geometry in LLMs and offer insight into how they internalize and process emotion.
Similar Papers
Decoding Neural Emotion Patterns through Large Language Model Embeddings
Computation and Language
Maps words to brain parts showing feelings.
Bridging the behavior-neural gap: A multimodal AI reveals the brain's geometry of emotion more accurately than human self-reports
Human-Computer Interaction
AI understands feelings better than people's words.
Fluent but Unfeeling: The Emotional Blind Spots of Language Models
Computation and Language
Helps computers understand feelings more like people.