Mark My Works Autograder for Programming Courses
By: Yiding Qiu, Seyed Mahdi Azimi, Artem Lensky
Large programming courses struggle to provide timely, detailed feedback on student code. We developed Mark My Works, a local autograding system that combines traditional unit testing with LLM-generated explanations. The system uses role-based prompts to analyze submissions, critique code quality, and generate pedagogical feedback while maintaining transparency in its reasoning process. We piloted the system in a 191-student engineering course, comparing AI-generated assessments with human grading on 79 submissions. While AI scores showed no linear correlation with human scores (r = -0.177, p = 0.124), both systems exhibited similar left-skewed distributions, suggesting they recognize comparable quality hierarchies despite different scoring philosophies. The AI system demonstrated more conservative scoring (mean: 59.95 vs 80.53 human) but generated significantly more detailed technical feedback.
Similar Papers
Autograder+: A Multi-Faceted AI Framework for Rich Pedagogical Feedback in Programming Education
Artificial Intelligence
Helps teachers give better feedback on student code.
Automated Feedback Generation for Undergraduate Mathematics: Development and Evaluation of an AI Teaching Assistant
Computers and Society
Teaches math by understanding student's own words.
CoGrader: Transforming Instructors' Assessment of Project Reports through Collaborative LLM Integration
Human-Computer Interaction
Helps teachers grade projects fairly and faster.