| Literature DB >> 32783045 |
Kyungjun Lee1, Jonggi Hong1, Simone Pimento1, Ebrima Jarjue1, Hernisa Kacorri1.
Abstract
For people with visual impairments, photography is essential in identifying objects through remote sighted help and image recognition apps. This is especially the case for teachable object recognizers, where recognition models are trained on user's photos. Here, we propose real-time feedback for communicating the location of an object of interest in the camera frame. Our audio-haptic feedback is powered by a deep learning model that estimates the object center location based on its proximity to the user's hand. To evaluate our approach, we conducted a user study in the lab, where participants with visual impairments (N = 9) used our feedback to train and test their object recognizer in vanilla and cluttered environments. We found that very few photos did not include the object (2% in the vanilla and 8% in the cluttered) and the recognition performance was promising even for participants with no prior camera experience. Participants tended to trust the feedback even though they know it can be wrong. Our cluster analysis indicates that better feedback is associated with photos that include the entire object. Our results provide insights into factors that can degrade feedback and recognition performance in teachable interfaces.Entities:
Keywords: hand; object recognition; sonification; visual impairments
Year: 2019 PMID: 32783045 PMCID: PMC7415326 DOI: 10.1145/3308561.3353799
Source DB: PubMed Journal: ASSETS