Score: 0

Surprisal and Metaphor Novelty: Moderate Correlations and Divergent Scaling Effects

Published: January 5, 2026 | arXiv ID: 2601.02015v2

By: Omar Momen , Emilie Sitter , Berenike Herrmann and more

Potential Business Impact:

Helps computers understand new, creative word uses.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Novel metaphor comprehension involves complex semantic processes and linguistic creativity, making it an interesting task for studying language models (LMs). This study investigates whether surprisal, a probabilistic measure of predictability in LMs, correlates with different metaphor novelty datasets. We analyse surprisal from 16 LM variants on corpus-based and synthetic metaphor novelty datasets. We explore a cloze-style surprisal method that conditions on full-sentence context. Results show that LMs yield significant moderate correlations with scores/labels of metaphor novelty. We further identify divergent scaling patterns: on corpus-based data, correlation strength decreases with model size (inverse scaling effect), whereas on synthetic data it increases (Quality-Power Hypothesis). We conclude that while surprisal can partially account for annotations of metaphor novelty, it remains a limited metric of linguistic creativity.

Country of Origin
🇩🇪 Germany

Page Count
15 pages

Category
Computer Science:
Computation and Language