Hierarchical Pedagogical Oversight: A Multi-Agent Adversarial Framework for Reliable AI Tutoring
By: Saisab Sadhu, Ashim Dhor
Large Language Models (LLMs) are increasingly deployed as automated tutors to address educator shortages; however, they often fail at pedagogical reasoning, frequently validating incorrect student solutions (sycophancy) or providing overly direct answers that hinder learning. We introduce Hierarchical Pedagogical Oversight (HPO), a framework that adapts structured adversarial synthesis to educational assessment. Unlike cooperative multi-agent systems that often drift toward superficial consensus, HPO enforces a dialectical separation of concerns: specialist agents first distill dialogue context, which then grounds a moderated, five-act debate between opposing pedagogical critics. We evaluate this framework on the MRBench dataset of 1,214 middle-school mathematics dialogues. Our 8B-parameter model achieves a Macro F1 of 0.845, outperforming GPT-4o (0.812) by 3.3% while using 20 times fewer parameters. These results establish adversarial reasoning as a critical mechanism for deploying reliable, low-compute pedagogical oversight in resource-constrained environments.
Similar Papers
How to Train a Leader: Hierarchical Reasoning in Multi-Agent LLMs
Multiagent Systems
Trains one smart AI to lead others.
Curriculum Guided Massive Multi Agent System Solving For Robust Long Horizon Tasks
Computation and Language
Helps robots solve hard, long problems together.
Heterogeneous Adversarial Play in Interactive Environments
Artificial Intelligence
Teaches computers new skills by playing games.