Reconstructing Item Characteristic Curves using Fine-Tuned Large Language Models
By: Christopher Ormerod
Potential Business Impact:
Makes tests better without needing real students.
Traditional methods for determining assessment item parameters, such as difficulty and discrimination, rely heavily on expensive field testing to collect student performance data for Item Response Theory (IRT) calibration. This study introduces a novel approach that implicitly models these psychometric properties by fine-tuning Large Language Models (LLMs) to simulate student responses across a spectrum of latent abilities. Leveraging the Qwen-3 dense model series and Low-Rank Adaptation (LoRA), we train models to generate responses to multiple choice questions conditioned on discrete ability descriptors. We reconstruct the probability of a correct response as a function of student ability, effectively generating synthetic Item Characteristic Curves (ICCs) to estimate IRT parameters. Evaluation on a dataset of Grade 6 English Language Arts (ELA) items and the BEA 2024 Shared Task dataset demonstrates that this method competes with or outperforms baseline approaches. This simulation-based technique seems particularly effective at modeling item discrimination.
Similar Papers
Estimating Item Difficulty Using Large Language Models and Tree-Based Machine Learning Algorithms
Computers and Society
Helps guess how hard school questions are.
Can LLMs Reliably Simulate Real Students' Abilities in Mathematics and Reading Comprehension?
Computation and Language
Makes AI tutors act like real students.
Latency-Response Theory Model: Evaluating Large Language Models via Response Accuracy and Chain-of-Thought Length
Methodology
Tests AI thinking speed and accuracy better.