Score: 1

Finance Language Model Evaluation (FLaME)

Published: June 18, 2025 | arXiv ID: 2506.15846v1

By: Glenn Matlin , Mika Okamoto , Huzaifa Pardawala and more

Potential Business Impact:

Tests computers on finance knowledge better.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Language Models (LMs) have demonstrated impressive capabilities with core Natural Language Processing (NLP) tasks. The effectiveness of LMs for highly specialized knowledge-intensive tasks in finance remains difficult to assess due to major gaps in the methodologies of existing evaluation frameworks, which have caused an erroneous belief in a far lower bound of LMs' performance on common Finance NLP (FinNLP) tasks. To demonstrate the potential of LMs for these FinNLP tasks, we present the first holistic benchmarking suite for Financial Language Model Evaluation (FLaME). We are the first research paper to comprehensively study LMs against 'reasoning-reinforced' LMs, with an empirical study of 23 foundation LMs over 20 core NLP tasks in finance. We open-source our framework software along with all data and results.

Country of Origin
🇺🇸 United States

Repos / Data Links

Page Count
48 pages

Category
Computer Science:
Computation and Language