Strengthening Programming Comprehension in Large Language Models through Code Generation
By: Xiaoning Ren , Qiang Hu , Wei Ma and more
Potential Business Impact:
Teaches computers to understand code better.
Large language models (LLMs) have recently shown impressive results on diverse code-related tasks, benefiting from large-scale training and instruction tuning. However, studies reveal that their grasp of fundamental programming concepts, such as data flow and control flow, remains shallow, leading to fragile performance when code requires deeper reasoning. This limitation restricts the practical adoption of LLMs in real-world software development. To address this issue, this work introduces a counterfactual code augmentation framework combined with concept-aware tuning, designed to guide LLMs toward stronger conceptual understanding. Comprehensive evaluation across multiple models and benchmarks demonstrates the effectiveness of the proposed approach.
Similar Papers
Cross-Task Benchmarking and Evaluation of General-Purpose and Code-Specific Large Language Models
Software Engineering
Makes computers better at understanding language and code.
Augmenting the Generality and Performance of Large Language Models for Software Engineering
Software Engineering
Helps computers understand and create software ideas.
On Code-Induced Reasoning in LLMs
Computation and Language
Code's structure helps computers think better than its meaning.