FinanceReasoning: Benchmarking Financial Numerical Reasoning More Credible, Comprehensive and Challenging
By: Zichen Tang , Haihong E , Ziyan Ma and more
Potential Business Impact:
Teaches computers to solve tricky money math problems.
We introduce FinanceReasoning, a novel benchmark designed to evaluate the reasoning capabilities of large reasoning models (LRMs) in financial numerical reasoning problems. Compared to existing benchmarks, our work provides three key advancements. (1) Credibility: We update 15.6% of the questions from four public datasets, annotating 908 new questions with detailed Python solutions and rigorously refining evaluation standards. This enables an accurate assessment of the reasoning improvements of LRMs. (2) Comprehensiveness: FinanceReasoning covers 67.8% of financial concepts and formulas, significantly surpassing existing datasets. Additionally, we construct 3,133 Python-formatted functions, which enhances LRMs' financial reasoning capabilities through refined knowledge (e.g., 83.2% $\rightarrow$ 91.6% for GPT-4o). (3) Challenge: Models are required to apply multiple financial formulas for precise numerical reasoning on 238 Hard problems. The best-performing model (i.e., OpenAI o1 with PoT) achieves 89.1% accuracy, yet LRMs still face challenges in numerical precision. We demonstrate that combining Reasoner and Programmer models can effectively enhance LRMs' performance (e.g., 83.2% $\rightarrow$ 87.8% for DeepSeek-R1). Our work paves the way for future research on evaluating and improving LRMs in domain-specific complex reasoning tasks.
Similar Papers
FinMMR: Make Financial Numerical Reasoning More Multimodal, Comprehensive, and Challenging
CV and Pattern Recognition
Tests computers on money math with pictures.
Understanding Financial Reasoning in AI: A Multimodal Benchmark and Error Learning Approach
Artificial Intelligence
AI learns to understand money charts and numbers.
ReasoningWeekly: A General Knowledge and Verbal Reasoning Challenge for Large Language Models
Artificial Intelligence
Tests AI with puzzles anyone can understand.