Score: 1

IsaMini: Redesigned Isabelle Proof Language for Machine Learning

Published: July 25, 2025 | arXiv ID: 2507.18885v2

By: Qiyuan Xu , Renxi Wang , Haonan Li and more

Potential Business Impact:

Helps computers prove math problems faster.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Neural Theorem Proving (NTP) employs deep learning methods, particularly Large Language Models (LLMs), to automate formal proofs in proof assistants. This approach holds promise for reducing the dramatic labor costs or computation costs required in proof engineering, which is fundamental to formal verification and other software engineering methods. The paper explores the potential of improving NTP by redesigning the proof language, given that LLMs' capabilities depend highly on representations. We introduce \emph{MiniLang}, a redesigned proof language for Isabelle/HOL incorporating an improved version of Sledgehammer. Experiments show MiniLang benefits two fine-tuned LLMs by improving the success rate on the PISA benchmark by up to 29\% in comparison to generation of Isar proof script. The success rate under one attempt (so-called \emph{pass@1}) reaches 69.1\%, exceeding the previous Baldur's pass@64 (65.7\%); The pass@8 reaches 79.2\%, exceeding the state-of-the-art on PISA (71.0\%) achieved by Magnushammer.

Country of Origin
πŸ‡ΈπŸ‡¬ Singapore

Page Count
13 pages

Category
Computer Science:
Programming Languages