Score: 1

CoT-Evo: Evolutionary Distillation of Chain-of-Thought for Scientific Reasoning

Published: October 15, 2025 | arXiv ID: 2510.13166v1

By: Kehua Feng , Keyan Ding , Zhihui Zhu and more

Potential Business Impact:

Teaches computers to solve science problems better.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

While chain-of-thought (CoT) distillation from advanced large language models (LLMs) has proven effective in general reasoning tasks, it struggles in scientific domains where even advanced models often produce incorrect or superficial reasoning due to high complexity and specialized knowledge requirements. Directly distilling from such flawed outputs results in low-quality training data and limits the performance of smaller student models. To overcome this, we propose CoT-Evo, an evolutionary CoT distillation framework. It begins by constructing a diverse pool of reasoning trajectories from multiple LLM thinkers, enriches them with automatically retrieved domain knowledge, and iteratively refines the trajectories using novelty-driven selection, reflective recombination and mutation. The refinement is guided by a fitness function that evaluates answer correctness, coherence, and effective knowledge utilization. This results in a high-quality CoT dataset tailored for scientific reasoning. We employ this evolved dataset to fine-tune a compact model, which achieves state-of-the-art performance on scientific reasoning benchmarks. Our work establishes a scalable approach to synthesizing high-fidelity scientific reasoning data from diverse and fallible LLMs.

Page Count
28 pages

Category
Computer Science:
Computation and Language