CodeSimpleQA: Scaling Factuality in Code Large Language Models
By: Jian Yang , Wei Zhang , Yizhi Li and more
Potential Business Impact:
Tests if AI truly knows how to code.
Large language models (LLMs) have made significant strides in code generation, achieving impressive capabilities in synthesizing code snippets from natural language instructions. However, a critical challenge remains in ensuring LLMs generate factually accurate responses about programming concepts, technical implementations, etc. Most previous code-related benchmarks focus on code execution correctness, overlooking the factual accuracy of programming knowledge. To address this gap, we present CodeSimpleQA, a comprehensive bilingual benchmark designed to evaluate the factual accuracy of code LLMs in answering code-related questions, which contains carefully curated question-answer pairs in both English and Chinese, covering diverse programming languages and major computer science domains. Further, we create CodeSimpleQA-Instruct, a large-scale instruction corpus with 66M samples, and develop a post-training framework combining supervised fine-tuning and reinforcement learning. Our comprehensive evaluation of diverse LLMs reveals that even frontier LLMs struggle with code factuality. Our proposed framework demonstrates substantial improvements over the base model, underscoring the critical importance of factuality-aware alignment in developing reliable code LLMs.
Similar Papers
CCFQA: A Benchmark for Cross-Lingual and Cross-Modal Speech and Text Factuality Evaluation
Computation and Language
Helps computers answer questions in many languages.
Facts Do Care About Your Language: Assessing Answer Quality of Multilingual LLMs
Computation and Language
Makes learning tools more truthful for all languages.
The Curious Case of Factual (Mis)Alignment between LLMs' Short- and Long-Form Answers
Computation and Language
Makes AI answer facts correctly, even in stories.