| Literature DB >> 35103758 |
Jaap de Ruyter van Steveninck1,2,3, Tom van Gestel2,4, Paula Koenders2,5, Guus van der Ham2,6, Floris Vereecken2,7, Umut Güçlü1,8, Marcel van Gerven1,9, Yagmur Güçlütürk1,10, Richard van Wezel2,11,12.
Abstract
Neuroprosthetic implants are a promising technology for restoring some form of vision in people with visual impairments via electrical neurostimulation in the visual pathway. Although an artificially generated prosthetic percept is relatively limited compared with normal vision, it may provide some elementary perception of the surroundings, re-enabling daily living functionality. For mobility in particular, various studies have investigated the benefits of visual neuroprosthetics in a simulated prosthetic vision paradigm with varying outcomes. The previous literature suggests that scene simplification via image processing, and particularly contour extraction, may potentially improve the mobility performance in a virtual environment. In the current simulation study with sighted participants, we explore both the theoretically attainable benefits of strict scene simplification in an indoor environment by controlling the environmental complexity, as well as the practically achieved improvement with a deep learning-based surface boundary detection implementation compared with traditional edge detection. A simulated electrode resolution of 26 × 26 was found to provide sufficient information for mobility in a simple environment. Our results suggest that, for a lower number of implanted electrodes, the removal of background textures and within-surface gradients may be beneficial in theory. However, the deep learning-based implementation for surface boundary detection did not improve mobility performance in the current study. Furthermore, our findings indicate that, for a greater number of electrodes, the removal of within-surface gradients and background textures may deteriorate, rather than improve, mobility. Therefore, finding a balanced amount of scene simplification requires a careful tradeoff between informativity and interpretability that may depend on the number of implanted electrodes.Entities:
Mesh:
Year: 2022 PMID: 35103758 PMCID: PMC8819280 DOI: 10.1167/jov.22.2.1
Source DB: PubMed Journal: J Vis ISSN: 1534-7362 Impact factor: 2.240
Summary of participant characteristics (n = 20).
| Characteristics | Median | Interquartile range |
|---|---|---|
| Age, years | 21 | 20.8–23.3 |
| Height, m | 1.84 | 1.75–1.87 |
Figure 1.Photos of the complex (left) and plain (right) obstacle course. Both environments contained identical cardboard boxes. In the complex environment, additional visual gradients are created with wallpaper and tape.
Figure 2.Overview of the obstacle course setup. The yellow boxes indicate large obstacles and the green boxes indicate small obstacles. Dashed lines indicate alternative box locations in other random route permutations. Out of all possible route layouts, a selection of seven routes of similar difficulty (based on the shortest path length around the obstacles) were used, as well as their mirrored versions.
Figure 3.Visualization of the image processing steps. (A) Input image. (B) Blurred image, using Gaussian smooting. (C) Edge mask, produced using the Canny algorithm. (D) Simulated phosphene vision, based on the Canny edge mask. (E) Surface normals prediction by the SharpNet deep learning model. (F) Surface boundary prediction by by the SharpNet deep learning model. (G) Surface boundary mask produced using the SharpNet predictions. (H) Simulated phosphene vision, based on surface boundary mask.
Figure 4.Image processing and phosphene simulation in the plain environment. (A) Input image. (B) Edge mask, produced using the Canny algorithm. (C) Surface boundary mask produced using the SharpNet predictions. (D–F) Comparison of different simulated phosphene resoutions (10 × 10, 26 × 26, and 50 × 50 phosphenes, respectively), with activations based on the Canny edge mask.
Overview of study conditions and corresponding number of trials.
| Camera vision | CED-based SPV | SharpNet-based SPV |
|---|---|---|
| Two trials per session (one for each visual complexity) | Two trials per session for each of the following six phosphene resolutions (one for each visual complexity): 10 × 10 phosphenes 18 × 18 phosphenes 26 × 26 phosphenes 34 × 34 phosphenes 42 × 42 phosphenes 50 × 50 phosphenes | Two trials per session for each of the following two phosphene resolutions (one for each visual complexity): 26 × 26 phosphenes 42 × 42 phosphenes |
| Total (two sessions): | Total (two sessions): | Total (two sessions): |
Descriptive statistics of the overall results and the control condition with camera view. Std. = Standard deviation.
| Overall | Control condition | |||
|---|---|---|---|---|
| Mean | Std. | Mean | Std. | |
| Trial duration (s) | 31.02 | 13.79 | 16.74 | 4.438 |
| No. of collisions | 0.879 | 1.526 | 0 | 0 |
| Subjective rating | 6.130 | 2.331 | 9.363 | 0.660 |
Figure 5.Mobility performance with CED-based SPV. The simulated number of phosphenes is plotted against standardized trial duration (left), standardized number of collisions (middle), and standardized subjective rating (right). Scene complexity is controlled by comparing a simple environment with plain cardboard boxes against a complex environment with additional background and surface textures. The dashed line indicates the average result for the control condition without SPV (i.e., normal camera vision). Asterisk (*) indicates p < 0.0125, double asterisk (**) indicates p < 0.0025.
p Values for Wilcoxon signed rank test for evaluation of the effect of scene complexity with CED. With a Bonferroni correction of α for six planned comparisons, findings are considered significant if p < 0.0083.
| Resolution | 10 × 10 | 18 × 18 | 26 × 26 | 34 × 34 | 42 × 42 | 50 × 50 |
|---|---|---|---|---|---|---|
| Trial duration |
|
| 0.044 | 0.765 | 0.145 |
|
| No. of collisions |
|
|
| 0.433 | 0.889 | 0.345 |
| Subjective rating | 0.039 |
| 0.124 | 0.078 | 0.221 | 0.039 |
p values for Wilcoxon signed-rank test for evaluation of the effect of image processing method. With a Bonferroni correction of α for four planned comparisons, findings are considered significant if p < 0.0125.
| Simple scene | Complex scene | |||
|---|---|---|---|---|
| Resolution | 26 × 26 | 42 × 42 | 26 × 26 | 42 × 42 |
| Trial duration |
|
| 0.478 |
|
| No. of collisions | 0.055 | 0.084 | 0.331 | 0.169 |
| Subjective rating |
|
| 0.520 |
|
Figure 6.Results of the SharpNet trials with deep-learning based surface boundary prediction versus edge detection with the Canny algorithm. Standardized trial duration (left), standardized number of collisions (middle), and standardized subjective rating (right) are plotted for two phosphene resolutions and two levels of environmental complexity. The complex environment contained additional background and surface textures where the simple environment consisted of plain cardboard boxes. Asterisk (*) indicates p < 0.0125, double asterisk (**) indicates p < 0.0025.