The cane uses a LIDAR sensor to measure distance to nearby obstacles and then directs users around those areas.
A new, affordable smart cane guides people with visual impairments safely and efficiently through their environments, say researchers.
Most know the white cane as a simple but crucial tool that assists people with visual impairments in making their way through the world.
Using tools from autonomous vehicles, the research team built the augmented cane, which helps people detect and identify obstacles, move easily around those objects, and follow routes both indoors and out.
The augmented cane is not the first smart cane. Research sensor canes can be heavy and expensive—weighing up to 50 pounds with a cost of around $6,000. Currently available sensor canes are technologically limited, only detecting objects right in front of the user.
The new cane sports cutting-edge sensors, weighs only three pounds, can be built at home from off-the-shelf parts, and free, open-source software, and costs $400.
The researchers hope their device will be an affordable and useful option for the more than 250 million people with impaired vision worldwide.
“We wanted something more user-friendly than just a white cane with sensors,” says Patrick Slade, a graduate research assistant in the Stanford University Intelligent Systems Laboratory and first author of the study in the journal Science Robotics. “Something that cannot only tell you there’s an object in your way, but tell you what that object is and then help you navigate around it.”
Robotic Cane Equipped with GPS and Gyroscopes
The augmented cane is equipped with a LIDAR sensor. LIDAR is the laser-based technology used in some self-driving cars and aircraft that measures the distance to nearby obstacles. The cane has additional sensors including GPS, accelerometers, magnetometers, and gyroscopes, like those on a smartphone, that monitor the user’s position, speed, direction, and so forth.
The cane makes decisions using artificial intelligence-based way finding and robotics algorithms like simultaneous localization and mapping (SLAM) and visual servoing—steering the user toward an object in an image.
Mounted at the tip of the cane is the pièce de résistance—a motorized, omnidirectional wheel that maintains contact with the ground. This wheel leads the user with impaired vision by gently tugging and nudging, left and right, around impediments. Equipped with built-in GPS and mapping capabilities, the augmented cane can even guide its user to precise locations—like a favorite store in the mall or a local coffee shop.
In real-world tests with users that volunteered through the Palo Alto Vista Center for the Blind and Visually Impaired, the researchers put the augmented cane in the hands of people with visual impairments as well as sighted people wearing blindfolds. They were then asked to complete everyday navigation challenges—walking hallways, avoiding obstacles, and traversing outdoor waypoints.
“We want the humans to be in control but provide them with the right level of gentle guidance to get them where they want to go as safely and efficiently as possible,” says senior author Mykel Kochenderfer, an associate professor of aeronautics and astronautics and an expert in aircraft collision-avoidance systems.
In that regard, the augmented cane excelled. It increased the walking speed for participants with impaired vision by roughly 20% over the white cane alone. For sighted people wearing blindfolds, the results were more impressive, increasing their speed by more than a third. An increased walking speed is related to better quality of life, Slade notes, so the hope is that the device could improve the quality of life of its users.
Make It Yourself
The researchers are open-sourcing every aspect of the project. The paper comes with a downloadable parts list and DIY solder-at-home instructions.
“We wanted to optimize this project for ease of replication and cost. Anyone can go and download all the code, bill of materials, and electronic schematics, all for free,” Kochenderfer says.
“Solder it up at home. Run our code. It’s pretty cool,” Slade adds.
But Kochenderfer notes the cane is still a research prototype. “A lot of significant engineering and experiments are necessary before it is ready for everyday use,” he says, adding that he and the team would welcome partners in industry who could streamline the design and scale up production to make the augmented cane even more affordable.
Next steps for the team include refinements to their prototype and developing a model that uses an everyday smartphone as the processor, an advance that could improve functionality, broaden access to the technology, and further drive down costs.
The National Science Foundation, Stanford Graduate Fellowship, and the Stanford Institute for Human-Centered AI funded the work.