This paper introduces a novel wearable assistive device designed to enhance the daily lives of visually impaired people (VIP) by providing continuous, real-time object finding. The proposed device integrates a miniature camera and a system on module (SoM) computing unit. Deep learning-based algorithms, specifically Faster R-CNNs, run on the SoM targeting to detect and recognize pertinent objects in the images provided by the camera. An integrated speaker conveys audio sentences letting the user know the presence of object(s) in the visual space. Experimental results conducted in real-world conditions show an 86% mean average precision for object recognition and a 215 ms mean computing time for real-time processing. The proposed image-to-speech device offers a practical and efficient solution to assist VIP in finding everyday objects.
Real-Time Object Finding for the Visually Impaired Using an Image-to-Speech Wearable Device
R. de Fazio;P. ViscontiUltimo
2025-01-01
Abstract
This paper introduces a novel wearable assistive device designed to enhance the daily lives of visually impaired people (VIP) by providing continuous, real-time object finding. The proposed device integrates a miniature camera and a system on module (SoM) computing unit. Deep learning-based algorithms, specifically Faster R-CNNs, run on the SoM targeting to detect and recognize pertinent objects in the images provided by the camera. An integrated speaker conveys audio sentences letting the user know the presence of object(s) in the visual space. Experimental results conducted in real-world conditions show an 86% mean average precision for object recognition and a 215 ms mean computing time for real-time processing. The proposed image-to-speech device offers a practical and efficient solution to assist VIP in finding everyday objects.I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.


