MaRVL-QA: A Benchmark for Mathematical Reasoning over Visual Landscapes
By: Nilay Pande , Sahiti Yerramilli , Jayant Sravan Tamarapalli and more
Potential Business Impact:
Teaches computers to understand math graphs.
A key frontier for Multimodal Large Language Models (MLLMs) is the ability to perform deep mathematical and spatial reasoning directly from images, moving beyond their established success in semantic description. Mathematical surface plots provide a rigorous testbed for this capability, as they isolate the task of reasoning from the semantic noise common in natural images. To measure progress on this frontier, we introduce MaRVL-QA (Mathematical Reasoning over Visual Landscapes), a new benchmark designed to quantitatively evaluate these core reasoning skills. The benchmark comprises two novel tasks: Topological Counting, identifying and enumerating features like local maxima; and Transformation Recognition, recognizing applied geometric transformations. Generated from a curated library of functions with rigorous ambiguity filtering, our evaluation on MaRVL-QA reveals that even state-of-the-art MLLMs struggle significantly, often resorting to superficial heuristics instead of robust spatial reasoning. MaRVL-QA provides a challenging new tool for the research community to measure progress, expose model limitations, and guide the development of MLLMs with more profound reasoning abilities.
Similar Papers
Multimodal Mathematical Reasoning Embedded in Aerial Vehicle Imagery: Benchmarking, Analysis, and Exploration
CV and Pattern Recognition
Tests if drones can do math from pictures.
GeoPQA: Bridging the Visual Perception Gap in MLLMs for Geometric Reasoning
Computation and Language
Helps AI understand pictures for better problem-solving.
VisuLogic: A Benchmark for Evaluating Visual Reasoning in Multi-modal Large Language Models
CV and Pattern Recognition
Tests if computers can truly "see" and understand.