Building Knowledge from Interactions: An LLM-Based Architecture for Adaptive Tutoring and Social Reasoning
By: Luca Garello , Giulia Belgiovine , Gabriele Russo and more
Potential Business Impact:
Robots learn to teach and remember like humans.
Integrating robotics into everyday scenarios like tutoring or physical training requires robots capable of adaptive, socially engaging, and goal-oriented interactions. While Large Language Models show promise in human-like communication, their standalone use is hindered by memory constraints and contextual incoherence. This work presents a multimodal, cognitively inspired framework that enhances LLM-based autonomous decision-making in social and task-oriented Human-Robot Interaction. Specifically, we develop an LLM-based agent for a robot trainer, balancing social conversation with task guidance and goal-driven motivation. To further enhance autonomy and personalization, we introduce a memory system for selecting, storing and retrieving experiences, facilitating generalized reasoning based on knowledge built across different interactions. A preliminary HRI user study and offline experiments with a synthetic dataset validate our approach, demonstrating the system's ability to manage complex interactions, autonomously drive training tasks, and build and retrieve contextual memories, advancing socially intelligent robotics.
Similar Papers
Gaze-supported Large Language Model Framework for Bi-directional Human-Robot Interaction
Robotics
Robots understand you better by watching and listening.
Multi-Agent Systems for Robotic Autonomy with LLMs
Robotics
Builds robots that can do jobs by themselves.
iLearnRobot: An Interactive Learning-Based Multi-Modal Robot with Continuous Improvement
Human-Computer Interaction
Robots learn from talking to people to get better.