NaviSense: A Multimodal Assistive Mobile application for Object Retrieval by Persons with Visual Impairment
By: Ajay Narayanan Sridhar , Fuli Qiao , Nelson Daniel Troncoso Aldas and more
Potential Business Impact:
Helps blind people find things using sound.
People with visual impairments often face significant challenges in locating and retrieving objects in their surroundings. Existing assistive technologies present a trade-off: systems that offer precise guidance typically require pre-scanning or support only fixed object categories, while those with open-world object recognition lack spatial feedback for reaching the object. To address this gap, we introduce 'NaviSense', a mobile assistive system that combines conversational AI, vision-language models, augmented reality (AR), and LiDAR to support open-world object detection with real-time audio-haptic guidance. Users specify objects via natural language and receive continuous spatial feedback to navigate toward the target without needing prior setup. Designed with insights from a formative study and evaluated with 12 blind and low-vision participants, NaviSense significantly reduced object retrieval time and was preferred over existing tools, demonstrating the value of integrating open-world perception with precise, accessible guidance.
Similar Papers
MR.NAVI: Mixed-Reality Navigation Assistant for the Visually Impaired
CV and Pattern Recognition
Helps blind people navigate safely with sound.
User-Centered Insights into Assistive Navigation Technologies for Individuals with Visual Impairment
Human-Computer Interaction
Helps blind people navigate better with smart tools.
Scene Awareness While Using Multiple Navigation Aids in AR Search
Human-Computer Interaction
AR navigation tools can hurt your memory.