Learning by Teaching: Engaging Students as Instructors of Large Language Models in Computer Science Education
By: Xinming Yang, Haasil Pujara, Jun Li
Potential Business Impact:
Students teach computers, learn better.
While Large Language Models (LLMs) are often used as virtual tutors in computer science (CS) education, this approach can foster passive learning and over-reliance. This paper presents a novel pedagogical paradigm that inverts this model: students act as instructors who must teach an LLM to solve problems. To facilitate this, we developed strategies for designing questions with engineered knowledge gaps that only a student can bridge, and we introduce Socrates, a system for deploying this method with minimal overhead. We evaluated our approach in an undergraduate course and found that this active-learning method led to statistically significant improvements in student performance compared to historical cohorts. Our work demonstrates a practical, cost-effective framework for using LLMs to deepen student engagement and mastery.
Similar Papers
Investigating Student Interaction Patterns with Large Language Model-Powered Course Assistants in Computer Science Courses
Computers and Society
Helps students get homework help anytime, anywhere.
Prompting the Professoriate: A Qualitative Study of Instructor Perspectives on LLMs in Data Science Education
Computers and Society
Teachers use AI to help students learn better.
Integrating Large Language Models and Evaluating Student Outcomes in an Introductory Computer Science Course
Computers and Society
Helps students learn computer coding with AI.