From Course to Skill: Evaluating LLM Performance in Curricular Analytics
By: Zhen Xu , Xinjin Li , Yingqi Huan and more
Potential Business Impact:
Helps schools improve classes using smart computer analysis.
Curricular analytics (CA) -- systematic analysis of curricula data to inform program and course refinement -- becomes an increasingly valuable tool to help institutions align academic offerings with evolving societal and economic demands. Large language models (LLMs) are promising for handling large-scale, unstructured curriculum data, but it remains uncertain how reliably LLMs can perform CA tasks. In this paper, we systematically evaluate four text alignment strategies based on LLMs or traditional NLP methods for skill extraction, a core task in CA. Using a stratified sample of 400 curriculum documents of different types and a human-LLM collaborative evaluation framework, we find that retrieval-augmented generation (RAG) is the top-performing strategy across all types of curriculum documents, while zero-shot prompting performs worse than traditional NLP methods in most cases. Our findings highlight the promise of LLMs in analyzing brief and abstract curriculum documents, but also reveal that their performance can vary significantly depending on model selection and prompting strategies. This underscores the importance of carefully evaluating the performance of LLM-based strategies before large-scale deployment.
Similar Papers
Evaluating 21st-Century Competencies in Postsecondary Curricula with Large Language Models: Performance Benchmarking and Reasoning-Based Prompting Strategies
Computers and Society
AI helps schools teach modern skills better.
Investigating Student Interaction Patterns with Large Language Model-Powered Course Assistants in Computer Science Courses
Computers and Society
Helps students get homework help anytime, anywhere.
A Rigorous Evaluation of LLM Data Generation Strategies for Low-Resource Languages
Computation and Language
Makes small AI learn languages better with smart text.