Dissecting Logical Reasoning in LLMs: A Fine-Grained Evaluation and Supervision Study
By: Yujun Zhou , Jiayi Ye , Zipeng Ling and more
Potential Business Impact:
Teaches computers to think step-by-step.
Logical reasoning is a core capability for many applications of large language models (LLMs), yet existing benchmarks often rely solely on final-answer accuracy, failing to capture the quality and structure of the reasoning process. We propose FineLogic, a fine-grained evaluation framework that assesses logical reasoning across three dimensions: overall benchmark accuracy, stepwise soundness, and representation-level alignment. In addition, to better understand how reasoning capabilities emerge, we conduct a comprehensive study on the effects of supervision format during fine-tuning. We construct four supervision styles (one natural language and three symbolic variants) and train LLMs under each. Our findings reveal that natural language supervision yields strong generalization even on out-of-distribution and long-context tasks, while symbolic reasoning styles promote more structurally sound and atomic inference chains. Further, our representation-level probing shows that fine-tuning primarily improves reasoning behaviors through step-by-step generation, rather than enhancing shortcut prediction or internalized correctness. Together, our framework and analysis provide a more rigorous and interpretable lens for evaluating and improving logical reasoning in LLMs.
Similar Papers
Logical Reasoning in Large Language Models: A Survey
Artificial Intelligence
Makes AI better at solving puzzles and thinking logically.
Evaluating Mathematical Reasoning Across Large Language Models: A Fine-Grained Approach
Machine Learning (CS)
Makes AI better at solving math problems.
LogReasoner: Empowering LLMs with Expert-like Coarse-to-Fine Reasoning for Log Analysis Tasks
Artificial Intelligence
Helps computers find computer problems like experts.