Accessible and Pedagogically-Grounded Explainability for Human-Robot Interaction: A Framework Based on UDL and Symbolic Interfaces
By: Francisco J. Rodríguez Lera , Raquel Fernández Hernández , Sonia Lopez González and more
Potential Business Impact:
Helps robots explain themselves to everyone.
This paper presents a novel framework for accessible and pedagogically-grounded robot explainability, designed to support human-robot interaction (HRI) with users who have diverse cognitive, communicative, or learning needs. We combine principles from Universal Design for Learning (UDL) and Universal Design (UD) with symbolic communication strategies to facilitate the alignment of mental models between humans and robots. Our approach employs Asterics Grid and ARASAAC pictograms as a multimodal, interpretable front-end, integrated with a lightweight HTTP-to-ROS 2 bridge that enables real-time interaction and explanation triggering. We emphasize that explainability is not a one-way function but a bidirectional process, where human understanding and robot transparency must co-evolve. We further argue that in educational or assistive contexts, the role of a human mediator (e.g., a teacher) may be essential to support shared understanding. We validate our framework with examples of multimodal explanation boards and discuss how it can be extended to different scenarios in education, assistive robotics, and inclusive AI.
Similar Papers
Enhancing Explainability with Multimodal Context Representations for Smarter Robots
Human-Computer Interaction
Robots understand what you say and see.
Trust Through Transparency: Explainable Social Navigation for Autonomous Mobile Robots via Vision-Language Models
Robotics
Robots explain their actions so you trust them.
Towards Cognitive Collaborative Robots: Semantic-Level Integration and Explainable Control for Human-Centric Cooperation
Robotics
Robots learn to work safely and understand people.