A Shared Geometry of Difficulty in Multilingual Language Models
By: Stefano Civelli , Pietro Bernardelle , Nicolò Brunello and more
Potential Business Impact:
Helps computers understand how hard problems are.
Predicting problem-difficulty in large language models (LLMs) refers to estimating how difficult a task is according to the model itself, typically by training linear probes on its internal representations. In this work, we study the multilingual geometry of problem-difficulty in LLMs by training linear probes using the AMC subset of the Easy2Hard benchmark, translated into 21 languages. We found that difficulty-related signals emerge at two distinct stages of the model internals, corresponding to shallow (early-layers) and deep (later-layers) internal representations, that exhibit functionally different behaviors. Probes trained on deep representations achieve high accuracy when evaluated on the same language but exhibit poor cross-lingual generalization. In contrast, probes trained on shallow representations generalize substantially better across languages, despite achieving lower within-language performance. Together, these results suggest that LLMs first form a language-agnostic representation of problem difficulty, which subsequently becomes language-specific. This closely aligns with existing findings in LLM interpretability showing that models tend to operate in an abstract conceptual space before producing language-specific outputs. We demonstrate that this two-stage representational process extends beyond semantic content to high-level meta-cognitive properties such as problem-difficulty estimation.
Similar Papers
LLMs Encode How Difficult Problems Are
Computation and Language
Makes computers understand hard problems better.
Probing the Difficulty Perception Mechanism of Large Language Models
Computation and Language
Computers learn to tell if math problems are hard.
Revisiting Generalization Across Difficulty Levels: It's Not So Easy
Computation and Language
Teaches computers to learn from easy and hard lessons.