While commendable progress has been made in user-centric research on mobile assistive systems for blind and low-vision (BLV) individuals, references that directly inform robot navigation design remain rare.
To bridge this gap, we conducted a comprehensive human study involving interviews with 26 guide dog handlers, 4 white cane users, 9 guide dog trainers, and 1 O&M trainer, along with 15+ hours of observing guide dog-assisted walking. After de-identification, we open-sourced the GuideData dataset to promote human-centered development.
Building on insights from this formative study, we developed GuideNav, a vision-only, teach-and-repeat navigation system. Inspired by how guide dogs are trained and assist their handlers, GuideNav autonomously repeats a path demonstrated by a sighted person using a robot.
The system constructs a topological representation of the taught route, integrates visual place recognition with temporal filtering, and employs a relative pose estimator to compute navigation actions — all without relying on costly, heavy, power-hungry sensors such as LiDAR.
In field tests, GuideNav consistently achieved kilometer-scale route following across five outdoor environments, maintaining reliability despite noticeable scene variations between teach and repeat runs.
A user study with 3 guide dog handlers and 1 guide dog trainer further confirmed the system's feasibility, marking (to our knowledge) the first demonstration of a quadruped mobile system retrieving a path in a manner comparable to guide dogs.
Teach Phase: A sighted person walks the desired route while the robot captures images and builds a topological map representation of the environment.
Repeat Phase: The robot uses visual place recognition (VPR) with temporal filtering to localize itself within the taught map, then employs a relative pose estimator to compute navigation actions and autonomously follow the demonstrated path.
Our vision-only approach eliminates the need for expensive sensors like LiDAR, making the system more practical and accessible for real-world deployment.
@article{hwang2025guidenav,
title={GuideNav: User-Informed Development of a Vision-Only Robotic Navigation Assistant For Blind Travelers},
author={Hwang, Hochul and Yang, Soowan and Monon, Jahir Sadik and Giudice, Nicholas A and Lee, Sunghoon Ivan and Biswas, Joydeep and Kim, Donghyun},
booktitle={2026 21st ACM/IEEE International Conference on Human-Robot Interaction (HRI)},
year={2026}
}