Advancing Research via Human-AI Interactive Theorem Proving
By: Chenyi Li , Zhijian Lai , Dong An and more
We investigate how large language models can be used as research tools in scientific computing while preserving mathematical rigor. We propose a human-in-the-loop workflow for interactive theorem proving and discovery with LLMs. Human experts retain control over problem formulation and admissible assumptions, while the model searches for proofs or contradictions, proposes candidate properties and theorems, and helps construct structures and parameters that satisfy explicit constraints, supported by numerical experiments and simple verification checks. Experts treat these outputs as raw material, further refine them, and organize the results into precise statements and rigorous proofs. We instantiate this workflow in a case study on the connection between manifold optimization and Grover's quantum search algorithm, where the pipeline helps identify invariant subspaces, explore Grover-compatible retractions, and obtain convergence guarantees for the retraction-based gradient method. The framework provides a practical template for integrating large language models into frontier mathematical research, enabling faster exploration of proof space and algorithm design while maintaining transparent reasoning responsibilities. Although illustrated on manifold optimization problems in quantum computing, the principles extend to other core areas of scientific computing.
Similar Papers
Mathematics with large language models as provers and verifiers
Computation and Language
AI solves hard math problems by working together.
Mathematics with large language models as provers and verifiers
Computation and Language
AI solves hard math problems with teamwork.
Mathematics with large language models as provers and verifiers
Computation and Language
AI solves hard math problems by working together.