AI-Powered Wearable Improves Navigation for the Visually Impaired
A prototype system using AI, audio, and haptic feedback shows a 25% performance boost over traditional canes in trials but requires further refinement for everyday use.
- The wearable device uses AI to process live video from a camera mounted on glasses, identifying obstacles and guiding users with audio and haptic feedback.
- Participants in trials experienced a 25% improvement in navigation performance, including walking distance and time, compared to conventional canes.
- The system integrates multimodal feedback, delivering audio cues every 250 milliseconds and vibrational alerts through artificial skin patches for obstacle detection.
- Testing included indoor maze trials and real-world scenarios, such as navigating city streets and cluttered environments, with successful results.
- Researchers emphasize the need for further refinements to ensure the prototype's safety, reliability, and practicality for daily use.