How Order-Sensitive Are LLMs? OrderProbe for Deterministic Structural Reconstruction
By: Yingjie He , Zhaolu Kang , Kehan Jiang and more
Large language models (LLMs) excel at semantic understanding, yet their ability to reconstruct internal structure from scrambled inputs remains underexplored. Sentence-level restoration is ill-posed for automated evaluation because multiple valid word orders often exist. We introduce OrderProbe, a deterministic benchmark for structural reconstruction using fixed four-character expressions in Chinese, Japanese, and Korean, which have a unique canonical order and thus support exact-match scoring. We further propose a diagnostic framework that evaluates models beyond recovery accuracy, including semantic fidelity, logical validity, consistency, robustness sensitivity, and information density. Experiments on twelve widely used LLMs show that structural reconstruction remains difficult even for frontier systems: zero-shot recovery frequently falls below 35%. We also observe a consistent dissociation between semantic recall and structural planning, suggesting that structural robustness is not an automatic byproduct of semantic competence.
Similar Papers
From Brute Force to Semantic Insight: Performance-Guided Data Transformation Design with LLMs
CV and Pattern Recognition
Helps computers write better code automatically.
When Words Change the Model: Sensitivity of LLMs for Constraint Programming Modelling
Artificial Intelligence
Computers struggle to solve problems with new words.
When Words Change the Model: Sensitivity of LLMs for Constraint Programming Modelling
Artificial Intelligence
Computers struggle to solve problems with new words.