Score: 1

Tell Me Who Your Students Are: GPT Can Generate Valid Multiple-Choice Questions When Students' (Mis)Understanding Is Hinted

Published: May 9, 2025 | arXiv ID: 2505.05815v2

By: Machi Shimmei , Masaki Uto , Yuichiroh Matsubayashi and more

Potential Business Impact:

Makes AI create better test questions.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

The primary goal of this study is to develop and evaluate an innovative prompting technique, AnaQuest, for generating multiple-choice questions (MCQs) using a pre-trained large language model. In AnaQuest, the choice items are sentence-level assertions about complex concepts. The technique integrates formative and summative assessments. In the formative phase, students answer open-ended questions for target concepts in free text. For summative assessment, AnaQuest analyzes these responses to generate both correct and incorrect assertions. To evaluate the validity of the generated MCQs, Item Response Theory (IRT) was applied to compare item characteristics between MCQs generated by AnaQuest, a baseline ChatGPT prompt, and human-crafted items. An empirical study found that expert instructors rated MCQs generated by both AI models to be as valid as those created by human instructors. However, IRT-based analysis revealed that AnaQuest-generated questions - particularly those with incorrect assertions (foils) - more closely resembled human-crafted items in terms of difficulty and discrimination than those produced by ChatGPT.

Country of Origin
🇦🇪 🇯🇵 🇺🇸 Japan, United States, United Arab Emirates

Page Count
9 pages

Category
Computer Science:
Computation and Language