Particle.news

Download on the App Store

AI-Powered Wearable Improves Navigation for the Visually Impaired

A prototype system using AI, audio, and haptic feedback shows a 25% performance boost over traditional canes in trials but requires further refinement for everyday use.

  • The wearable device uses AI to process live video from a camera mounted on glasses, identifying obstacles and guiding users with audio and haptic feedback.
  • Participants in trials experienced a 25% improvement in navigation performance, including walking distance and time, compared to conventional canes.
  • The system integrates multimodal feedback, delivering audio cues every 250 milliseconds and vibrational alerts through artificial skin patches for obstacle detection.
  • Testing included indoor maze trials and real-world scenarios, such as navigating city streets and cluttered environments, with successful results.
  • Researchers emphasize the need for further refinements to ensure the prototype's safety, reliability, and practicality for daily use.
Hero image