Towards an Efficient, Customizable, and Accessible AI Tutor
By: Juan Segundo Hevia, Facundo Arredondo, Vishesh Kumar
Potential Business Impact:
AI tutors work without internet for all students.
The integration of large language models (LLMs) into education offers significant potential to enhance accessibility and engagement, yet their high computational demands limit usability in low-resource settings, exacerbating educational inequities. To address this, we propose an offline Retrieval-Augmented Generation (RAG) pipeline that pairs a small language model (SLM) with a robust retrieval mechanism, enabling factual, contextually relevant responses without internet connectivity. We evaluate the efficacy of this pipeline using domain-specific educational content, focusing on biology coursework. Our analysis highlights key challenges: smaller models, such as SmolLM, struggle to effectively leverage extended contexts provided by the RAG pipeline, particularly when noisy or irrelevant chunks are included. To improve performance, we propose exploring advanced chunking techniques, alternative small or quantized versions of larger models, and moving beyond traditional metrics like MMLU to a holistic evaluation framework assessing free-form response. This work demonstrates the feasibility of deploying AI tutors in constrained environments, laying the groundwork for equitable, offline, and device-based educational tools.
Similar Papers
A RAG-Based Institutional Assistant
Computation and Language
Helps AI find correct school answers from books.
Aligning LLMs for the Classroom with Knowledge-Based Retrieval -- A Comparative RAG Study
Artificial Intelligence
Makes AI answers for school more truthful.
Small Models, Big Support: A Local LLM Framework for Teacher-Centric Content Creation and Assessment using RAG and CAG
Computers and Society
Helps teachers make custom lessons safely and cheaply.