SemanticScanpath: Combining Gaze and Speech for Situated Human-Robot Interaction Using LLMs
By: Elisabeth Menendez , Michael Gienger , Santiago Martínez and more
Potential Business Impact:
Robots understand what you mean by looking.
Large Language Models (LLMs) have substantially improved the conversational capabilities of social robots. Nevertheless, for an intuitive and fluent human-robot interaction, robots should be able to ground the conversation by relating ambiguous or underspecified spoken utterances to the current physical situation and to the intents expressed non verbally by the user, for example by using referential gaze. Here we propose a representation integrating speech and gaze to enable LLMs to obtain higher situated awareness and correctly resolve ambiguous requests. Our approach relies on a text-based semantic translation of the scanpath produced by the user along with the verbal requests and demonstrates LLM's capabilities to reason about gaze behavior, robustly ignoring spurious glances or irrelevant objects. We validate the system across multiple tasks and two scenarios, showing its generality and accuracy, and demonstrate its implementation on a robotic platform, closing the loop from request interpretation to execution.
Similar Papers
GazeLLM: Multimodal LLMs incorporating Human Visual Attention
Human-Computer Interaction
Lets computers understand videos by watching eyes.
Agreeing to Interact in Human-Robot Interaction using Large Language Models and Vision Language Models
Human-Computer Interaction
Helps robots know when to start talking to people.
Large Language Models and 3D Vision for Intelligent Robotic Perception and Autonomy
Robotics
Robots understand and act on spoken commands.