NoReGeo: Non-Reasoning Geometry Benchmark
By: Irina Abdullaeva , Anton Vasiliuk , Elizaveta Goncharova and more
We present NoReGeo, a novel benchmark designed to evaluate the intrinsic geometric understanding of large language models (LLMs) without relying on reasoning or algebraic computation. Unlike existing benchmarks that primarily assess models' proficiency in reasoning-based geometry-where solutions are derived using algebraic methods-NoReGeo focuses on evaluating whether LLMs can inherently encode spatial relationships and recognize geometric properties directly. Our benchmark comprises 2,500 trivial geometric problems spanning 25 categories, each carefully crafted to be solvable purely through native geometric understanding, assuming known object locations. We assess a range of state-of-the-art models on NoReGeo, including frontier models like GPT-4, observing that even the most advanced systems achieve an overall maximum of 65% accuracy in binary classification tasks. Further, our ablation experiments demonstrate that such geometric understanding does not emerge through fine-tuning alone, indicating that effective training for geometric comprehension requires a specialized approach from the outset. Our findings highlight a significant gap in current LLMs' ability to natively grasp geometric concepts, providing a foundation for future research toward models with true geometric cognition.
Similar Papers
GeoGramBench: Benchmarking the Geometric Program Reasoning in Modern LLMs
Artificial Intelligence
Teaches computers to understand drawings from code.
GeoBench: Rethinking Multimodal Geometric Problem-Solving via Hierarchical Evaluation
CV and Pattern Recognition
Tests how well computers solve shape puzzles.
GeoSense: Evaluating Identification and Application of Geometric Principles in Multimodal Reasoning
Computation and Language
Helps computers understand and solve geometry problems.