Beyond the Score: Uncertainty-Calibrated LLMs for Automated Essay Assessment
By: Ahmed Karim, Qiao Wang, Zheng Yuan
Potential Business Impact:
Helps computers grade essays with confidence.
Automated Essay Scoring (AES) systems now reach near human agreement on some public benchmarks, yet real-world adoption, especially in high-stakes examinations, remains limited. A principal obstacle is that most models output a single score without any accompanying measure of confidence or explanation. We address this gap with conformal prediction, a distribution-free wrapper that equips any classifier with set-valued outputs and formal coverage guarantees. Two open-source large language models (Llama-3 8B and Qwen-2.5 3B) are fine-tuned on three diverse corpora (ASAP, TOEFL11, Cambridge-FCE) and calibrated at a 90 percent risk level. Reliability is assessed with UAcc, an uncertainty-aware accuracy that rewards models for being both correct and concise. To our knowledge, this is the first work to combine conformal prediction and UAcc for essay scoring. The calibrated models consistently meet the coverage target while keeping prediction sets compact, indicating that open-source, mid-sized LLMs can already support teacher-in-the-loop AES; we discuss scaling and broader user studies as future work.
Similar Papers
Agreement Between Large Language Models and Human Raters in Essay Scoring: A Research Synthesis
Computation and Language
Helps computers grade essays as well as people.
LCES: Zero-shot Automated Essay Scoring via Pairwise Comparisons Using Large Language Models
Computation and Language
Helps computers grade essays more like humans.
Automated Refinement of Essay Scoring Rubrics for Language Models via Reflect-and-Revise
Computation and Language
Teaches computers to grade essays like humans.