Psychological Factors Influencing University Students Trust in AI-Based Learning Assistants
By: Ezgi Dağtekin, Ercan Erkalkan
Artificial intelligence (AI) based learning assistants and chatbots are increasingly integrated into higher education. While these tools are often evaluated in terms of technical performance, their successful and ethical use also depends on psychological factors such as trust, perceived risk, technology anxiety, and students general attitudes toward AI. This paper adopts a psychology oriented perspective to examine how university students form trust in AI based learning assistants. Drawing on recent literature in mental health, human AI interaction, and trust in automation, we propose a conceptual framework that organizes psychological predictors of trust into four groups: cognitive appraisals, affective reactions, social relational factors, and contextual moderators. A narrative review approach synthesizes empirical findings and derives research questions and hypotheses for future studies. The paper highlights that trust in AI is a psychological process shaped by individual differences and learning environments, with practical implications for instructors, administrators, and designers of educational AI systems.
Similar Papers
Understanding Human-AI Trust in Education
Computers and Society
Helps students trust AI tutors correctly.
Exploring Teenagers' Trust in Al Chatbots: An Empirical Study of Chinese Middle-School Students
Human-Computer Interaction
Teens trust AI more when they are strong inside.
Trust in AI emerges from distrust in humans: A machine learning study on decision-making guidance
Human-Computer Interaction
People trust computers more when they don't trust people.