| Literature DB >> 35126078 |
Mohammad-Parsa Hosseini1,2, Madison Beary1, Alex Hadsell1, Ryan Messersmith1, Hamid Soltanian-Zadeh3.
Abstract
In this paper, we introduce a deep learning model to classify children as either healthy or potentially having autism with 94.6% accuracy using Deep Learning. Patients with autism struggle with social skills, repetitive behaviors, and communication, both verbal and non-verbal. Although the disease is considered to be genetic, the highest rates of accurate diagnosis occur when the child is tested on behavioral characteristics and facial features. Patients have a common pattern of distinct facial deformities, allowing researchers to analyze only an image of the child to determine if the child has the disease. While there are other techniques and models used for facial analysis and autism classification on their own, our proposal bridges these two ideas allowing classification in a cheaper, more efficient method. Our deep learning model uses MobileNet and two dense layers to perform feature extraction and image classification. The model is trained and tested using 3,014 images, evenly split between children with autism and children without it; 90% of the data is used for training and 10% is used for testing. Based on our accuracy, we propose that the diagnosis of autism can be done effectively using only a picture. Additionally, there may be other diseases that are similarly diagnosable.Entities:
Keywords: autism; children; deep learning; diagnosis; facial image analysis
Year: 2022 PMID: 35126078 PMCID: PMC8811190 DOI: 10.3389/fncom.2021.789998
Source DB: PubMed Journal: Front Comput Neurosci ISSN: 1662-5188 Impact factor: 2.380
Figure 1On the (Left) is a child with autism, and on the (Right) is a child without autism, in order to compare some facial features.
Depthwise separable vs. full convolution MobileNet (Jeatrakul and Wong, 2009; Howard et al., 2017).
|
|
|
|
|
|---|---|---|---|
| Conv. MobileNet | 71.7 | 4,866 | 29.3 |
| MobileNet | 70.6 | 569 | 4.2 |
Narrow vs. shallow MobileNet and MobileNet width multiplier (Jeatrakul and Wong, 2009; Howard et al., 2017; Hosseini et al., 2020a).
|
|
|
|
|
|---|---|---|---|
| Shallow MobileNet | 65.3 | 307 | 2.9 |
| 1.0 MobileNet-224 | 70.6 | 569 | 4.2 |
| 0.75 MobileNet | 68.4 | 325 | 2.6 |
| 0.5 MobileNet-224 | 63.7 | 149 | 1.3 |
| 0.25 MobileNet-224 | 50.6 | 41 | 0.5 |
MobileNet resolution and MobileNet comparison to popular models.
|
|
|
|
|
|---|---|---|---|
| 1.0 MobileNet-224 | 70.6 | 569 | 4.2 |
| 1.0 MobileNet-192 | 69.1 | 418 | 4.2 |
| 1.0 MobileNet-160 | 67.2 | 290 | 4.2 |
| 1.0 MobileNet-128 | 64.4 | 186 | 4.2 |
| GoogleNet | 69.8 | 1,550 | 6.8 |
| VGG 16 | 71.5 | 15,300 | 138 |
Figure 2Some images used in the deep learning training step. (Top) Children who have autism. (Bottom) Children who do not have autism.
Dataset breakdown.
|
|
|
|
|---|---|---|
| Train | 1,327 autistic | 88 |
| Validation | 80 autistic | 5.3 |
| Test | 140 autistic | 9.3 |
| Total | 1,507 autistic | 100 |
Figure 3The algorithm architecture of the proposed model, illustrating the use of MobileNet, followed by two dense layers to perform image recognition. MobileNet uses CNN to predict what is the shape of the object present and what is matched with it from the images.
Body architecture of the developed MobileNet (Howard et al., 2017) for autism image recognition.
|
|
|
|
|
|---|---|---|---|
| 224 * 224 * 3 | 3 * 3 * 3 * 32 | Convolution | S2 |
| 112 * 112 * 32 | Depth-Wise 3 * 3 * 32 | Depth-Wise | S1 |
| 112 * 112 * 32 | 1 * 1 * 32 * 64 | Convolution | S1 |
| 112 * 112 * 64 | Depth-Wise 3 * 3 * 64 | Depth-Wise | S2 |
| 56 * 56 * 64 | 1 * 1 * 64 * 128 | Convolution | S1 |
| 56 * 56 * 128 | Depth-Wise 3 * 3 * 128 | Depth-Wise | S1 |
| 56 * 56 * 128 | 1 * 1 * 128 * 128 | Convolution | S1 |
| 56 * 56 * 128 | Depth-wise 3 * 3 * 128 | Depth-Wise | S2 |
| 28 * 28 * 128 | 1 * 1 * 128 * 256 | Convolution | S1 |
| 28 * 28 * 256 | Depth-Wise 3 * 3 * 256 | Depth-Wise | S1 |
| 28 * 28 * 256 | 1 * 1 * 256 * 256 | Convolution | S1 |
| 28 * 28 * 256 | Depth-Wise 3 * 3 * 256 | Depth-Wise | S2 |
| 14 * 14 * 256 | 1 * 1 * 256 * 512 | Convolution | S1 |
| 14 * 14 * 512 | Depth-Wise 3 * 3 * 512 | Depth-Wise | S1 |
| 14 * 14 * 512 | 1 * 1 * 512 * 512 | Convolution | S1 |
| 14 * 14 * 512 | Depth-Wise 3 * 3 * 512 | Depth-Wise | S1 |
| 14 * 14 * 512 | 1 * 1 * 512 * 512 | Convolution | S1 |
| 14 * 14 * 512 | Depth-Wise 3 * 3 * 512 | Depth-Wise | S1 |
| 14 * 14 * 512 | 1 * 1 * 512 * 512 | Convolution | S1 |
| 14 * 14 * 512 | Depth-Wise 3 * 3 * 512 | Depth-Wise | S1 |
| 14 * 14 * 512 | 1 * 1 * 512 * 512 | Convolution | S1 |
| 14 * 14 * 512 | Depth-Wise 3 * 3 * 512 | Depth-Wise | S1 |
| 14 * 14 * 512 | 1 * 1 * 512 * 512 | Convolution | S1 |
| 14 * 14 * 512 | Depth-Wise 3 * 3 * 512 | Depth-Wise | S2 |
| 7 * 7 * 512 | 1 * 1 * 512 * 1,024 | Convolution | S1 |
| 7 * 7 * 1,024 | Depth-Wise 3 * 3 * 1,024 | Depth-Wise | S2 |
| 7 * 7 * 1,024 | Depth-Wise 1 * 1 * 1,024 | Convolution | S1 |
| 7 * 7 * 1,024 | Pool 7 * 7 | Average pooling | S1 |
| 1 * 1 * 1,024 | 1,024 * 1,000 | Fully connected | S1 |
| 1 * 1 * 1,000 | Classifier | Softmax | S1 |
Figure 4By adding one epoch at a time, this figure shows how the loss changes.
Figure 5With the addition of one epoch, this figure shows how accuracy changed for training, validation, and testing data.