When Models Know More Than They Can Explain: Quantifying Knowledge Transfer in Human-AI Collaboration
By: Quan Shi , Carlos E. Jimenez , Shunyu Yao and more
Potential Business Impact:
AI learns to explain its thinking to people.
Recent advancements in AI reasoning have driven substantial improvements across diverse tasks. A critical open question is whether these improvements also yields better knowledge transfer: the ability of models to communicate reasoning in ways humans can understand, apply, and learn from. To investigate this, we introduce Knowledge Integration and Transfer Evaluation (KITE), a conceptual and experimental framework for Human-AI knowledge transfer capabilities and conduct the first large-scale human study (N=118) explicitly designed to measure it. In our two-phase setup, humans first ideate with an AI on problem-solving strategies, then independently implement solutions, isolating model explanations' influence on human understanding. Our findings reveal that although model benchmark performance correlates with collaborative outcomes, this relationship is notably inconsistent, featuring significant outliers, indicating that knowledge transfer requires dedicated optimization. Our analysis identifies behavioral and strategic factors mediating successful knowledge transfer. We release our code, dataset, and evaluation framework to support future work on communicatively aligned models.
Similar Papers
From Developer Pairs to AI Copilots: A Comparative Study on Knowledge Transfer
Software Engineering
AI helps coders learn, but they trust it too much.
Query-based Knowledge Transfer for Heterogeneous Learning Environments
Machine Learning (CS)
Lets computers learn from others without seeing their private data.
Modeling AI-Human Collaboration as a Multi-Agent Adaptation
Multiagent Systems
AI helps people work better on certain tasks.