| Literature DB >> 33980874 |
Jessica S Oliveira1, Felipe O Franco2,3, Mirian C Revers2, Andréia F Silva2, Joana Portolese2, Helena Brentani2,3, Ariane Machado-Lima1,3, Fátima L S Nunes4.
Abstract
An advantage of using eye tracking for diagnosis is that it is non-invasive and can be performed in individuals with different functional levels and ages. Computer/aided diagnosis using eye tracking data is commonly based on eye fixation points in some regions of interest (ROI) in an image. However, besides the need for every ROI demarcation in each image or video frame used in the experiment, the diversity of visual features contained in each ROI may compromise the characterization of visual attention in each group (case or control) and consequent diagnosis accuracy. Although some approaches use eye tracking signals for aiding diagnosis, it is still a challenge to identify frames of interest when videos are used as stimuli and to select relevant characteristics extracted from the videos. This is mainly observed in applications for autism spectrum disorder (ASD) diagnosis. To address these issues, the present paper proposes: (1) a computational method, integrating concepts of Visual Attention Model, Image Processing and Artificial Intelligence techniques for learning a model for each group (case and control) using eye tracking data, and (2) a supervised classifier that, using the learned models, performs the diagnosis. Although this approach is not disorder-specific, it was tested in the context of ASD diagnosis, obtaining an average of precision, recall and specificity of 90%, 69% and 93%, respectively.Entities:
Mesh:
Year: 2021 PMID: 33980874 PMCID: PMC8115570 DOI: 10.1038/s41598-021-89023-8
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Features selected by genetic algorithm for each category.
| Features | # of ASD features | # of TD features |
|---|---|---|
| Steerable pyramids | 3 | 4 |
| Saliency toolbox: color, intensity, orientation and skin | 4 | 4 |
| RGB color | 0 | 1 |
| Horizon line | 1 | 1 |
| Presence of face | 1 | 1 |
| Presence of people | 1 | 1 |
| Distance to the frame center | 1 | 0 |
| Motion value | 1 | 1 |
| Presence of biological movement | 1 | 0 |
| Presence of geometrical movement | 1 | 1 |
| Distance to the side-specific scene center | 1 | 1 |
| Total | 15 | 15 |
Figure 1ROC Curves for Neural Networks with the features selected by the Genetic Algorithm. The 5 lines are the results of each of the 5-fold cross-validation rounds (this figure was built with MatLab 2015a version 8.5- www.mathworks.com/products/matlab.html[25]).
Comparison of results of the evaluated approaches.
| Classification algorithm | Feature selection algorithm | Average AUC (standard deviation) |
|---|---|---|
| SVM | None | 0.775 (0.027) |
| SVM | Genetic algorithm | 0.695 (0.023) |
| SVM | Relief | 0.695 (0.042) |
| ANN | None | 0.818 (0.053) |
| ANN | Genetic algorithm | 0.822 (0.015) |
| ANN | Relief | 0.782 (0.026) |
Comparison of results among related work.
| Reference | Dataset | Average AUC |
|---|---|---|
| Chevallier et al.[ | 81 children (6–17 years) | 0.71 |
| Pierce et al.[ | 334 children (1–3 years) | 0.71 |
| Shi et al.[ | 33 children (4–6 years) | 0.86 |
| This work | 106 children (3–18 years) | 0.82 |
Figure 2Overview of the entire process of the proposed model (this figure was built with XPaint version 2.9.10- https://directory.fsf.org/wiki/Xpaint[27]).
Figure 3Example of frames of the video used as visual stimuli for training the Visual Attention Models (this figure was built in XPaint version 2.9.10- https://directory.fsf.org/wiki/Xpaint[27]).
Figure 4Example of fixation maps for a video frame that contains a scene of biological movement on the left side and a scene of geometric movement on the right side. The frame used for generating these maps are similar to frames B and C in Fig. 3 (this figure was built with MatLab 2015a version 8.5-www.mathworks.com/products/matlab.html[25]).