| Literature DB >> 34745371 |
Hadeer A Helaly1,2, Mahmoud Badawy2,3, Amira Y Haikal2.
Abstract
Alzheimer's disease (AD) is a chronic, irreversible brain disorder, no effective cure for it till now. However, available medicines can delay its progress. Therefore, the early detection of AD plays a crucial role in preventing and controlling its progression. The main objective is to design an end-to-end framework for early detection of Alzheimer's disease and medical image classification for various AD stages. A deep learning approach, specifically convolutional neural networks (CNN), is used in this work. Four stages of the AD spectrum are multi-classified. Furthermore, separate binary medical image classifications are implemented between each two-pair class of AD stages. Two methods are used to classify the medical images and detect AD. The first method uses simple CNN architectures that deal with 2D and 3D structural brain scans from the Alzheimer's Disease Neuroimaging Initiative (ADNI) dataset based on 2D and 3D convolution. The second method applies the transfer learning principle to take advantage of the pre-trained models for medical image classifications, such as the VGG19 model. Due to the COVID-19 pandemic, it is difficult for people to go to hospitals periodically to avoid gatherings and infections. As a result, Alzheimer's checking web application is proposed using the final qualified proposed architectures. It helps doctors and patients to check AD remotely. It also determines the AD stage of the patient based on the AD spectrum and advises the patient according to its AD stage. Nine performance metrics are used in the evaluation and the comparison between the two methods. The experimental results prove that the CNN architectures for the first method have the following characteristics: suitable simple structures that reduce computational complexity, memory requirements, overfitting, and provide manageable time. Besides, they achieve very promising accuracies, 93.61% and 95.17% for 2D and 3D multi-class AD stage classifications. The VGG19 pre-trained model is fine-tuned and achieved an accuracy of 97% for multi-class AD stage classifications.Entities:
Keywords: Alzheimer’s disease; Brain MRI; Convolutional neural network (CNN); Deep learning; Medical image classification
Year: 2021 PMID: 34745371 PMCID: PMC8563360 DOI: 10.1007/s12559-021-09946-2
Source DB: PubMed Journal: Cognit Comput ISSN: 1866-9956 Impact factor: 4.890
Fig. 1A proportion of people affected by AD according to ages in the United States [5]
Demographic data for 300 subjects
| Subject number | 75 | 75 | 75 | 75 |
| Male/female | 21/54 | 51/24 | 43/32 | 32/43 |
| Age (mean ± STD) | 75.95 ± 0.91 | 76.08 ± 0.89684 | 77.44 ± 1.33801 | 75.68 ± 0.469617 |
Fig. 2Slices of MR images: Accelerated Sagittal MPRAGE view, Axial Field Mapping view, and 3 Plane. Localizer view from left to right of AD patient
Training, validation, and test set size
| 0 AD | 9600 | 1200 | 1200 | 12,000 |
| 1 EMCI | 9600 | 1200 | 1200 | 12,000 |
| 2 LMCI | 9600 | 1200 | 1200 | 12,000 |
| 3 NC | 9600 | 1200 | 1200 | 12,000 |
| Total | 38,400 | 4800 | 4800 | 48,000 |
Fig. 3The proposed framework E2AD2C architecture
Fig. 4Example of the normalization methods applied on MRI image
Fig. 5Illustration of the convolutional operation
Fig. 6The difference among the sigmoid, Relu, and LRelu activation functions [24]
Fig. 7The 2D-M2IC model architecture
Fig. 8The 3D-M2IC model architecture
The tuning applied in the vgg19 model
| vgg19 (functional) | (None, 3, 3, 512) | 20,024,384 |
| flatten (Flatten) | (None, 4608) | 0 |
| dense (Dense) | (None, 1024) | 4,719,616 |
| dense_1 (Dense) | (None, 512) | 524,800 |
| dense_2 (Dense) | (None, 256) | 131,328 |
| dropout (Dropout) | (None, 256) | 0 |
| dense_3 (Dense) | (None, 128) | 32,896 |
| dropout_1 (Dropout) | (None, 128) | 0 |
| dense_4 (Dense) | (None, 4) | 516 |
| Total params: 25,433,540 | ||
| Trainable params: 25,433,540 | ||
| Non-trainable params: 0 | ||
Summarization of the applied performance metrics
| •It is the number of the correct prediction to the total number of predictions | where TP, TN, FP, and FN represent the True Positive, True Negative, False Positive, and False Negative values | |
| •For binary classification, we use binary cross-entropy loss | where | |
| •For multi-classification, we use categorical cross-entropy loss | where | |
| •It is the harmonic mean of precision and recall. It has a range of [0, 1]. The higher the F1 Score is, the better the model performance is | ||
| •It is the correct positive result amount to all relevant sample amount | ||
| •It is the correct positive result amount to the positive result amount predicted by the classifier | ||
| •It picks a good cut-off Threshold for the model from plotting True Positive Rate (TPR) against False Positive Rate (FPR) for different values of the Threshold in the range of [0, 1] | ||
| •The higher the correlation between True and predicted values is, the better the model prediction is | MCC = | |
•It is the complete description of the model performance •It gives a matrix as an output, and it forms the basis of other types of metrics that depend on TP, TN, FP, and FN metrics | To understand the definition of TP, TN, FP, and FN, assume the proposed binary model classifies between AD and NC then: – TP: The case that – TN: The case that – FP: The case that – FN: The case that |
Comparison of the proposed models with the state-of-the-art models
| Approach | Dataset | Modality | Type of classification | Accuracy | |
|---|---|---|---|---|---|
| 755 in each class (AD, MCI, and HC) | ADNI | MRI | Binary, multi | ||
| 302 subjects (211 AD, 91 NC) | ADNI | MRI, fMRI | Binary | ||
| 210 subjects (70 AD, 70 NC, 70 MCI) | CAD-dementia | MRI | Binary, multi | ||
| 50 AD, 43 LMCI, 77 EMCI, 61 NC | ADNI | MRI | Binary | ||
| 98 AD, 98 NC | Local hospitals, OASIS | MRI | Binary | ||
| 53 AD, 228 MCI, 250 NC | ADNI | sMRI and DTI | |||
| 530 subjects (185 AD, 185 MCI, 160 HC) | ADNI | MRI | Multi | ||
| AD 192, 184 NC | ADNI | MRI | Binary | ||
| 35 AD, 30 aMCI, 40 NC | Beijing Xuanwu Hospital | DTI, fMRI | Multi | ||
| 28 AD, 28 NC | OASIS | MRI | Binary | ||
| 150 subjects (AD 50, NC 50, MCI 50) | ADNI | sMRI | Multi, binary | ||
| AD 12, NC 12, EMCI 12, LMCI 12 | ADNI | DTI | Multi | ||
| 337 subjects (198 AD, 139 NC) | ADNI | MRI | Binary | ||
| 120 subjects, 30 for each class (AD, EMCI, LMCI, NC) | ADNI | 4D FMRI | Multi-classification | ||
| 407 HC, 418 AD, 280 c-MCI, 533 stable MCI [s-MCI] | ADNI | 3D MRI | Binary | ||
| 787 subjects for (AD, MCIc, MCInc, HC) classes | ADNI | 3D MRI | Binary | ||
| 600 brain MRI images | ADNI | 3D MRI | Multi | ||
300 subjects (75 AD, 75 EMCI, 75 LMCI, 75 NC) Total size = 48,000 MRI images | ADNI | 2D MRI | Multi, binary | ||
| 3D MRI | Multi, binary | ||||
| 2D MRI | Multi | ||||
Fig. 9The comparison of the proposed models with other models for multi-class medical image classification
Fig. 10The comparison among the proposed models (2D-M2IC, 3D-M2IC, 2D-BMIC, 3D-BMIC, and fine-tuned VGG19 model) with one another
Comparison of the performance metrics of the two proposed models (2D-M2IC model, 3D-M2IC model)
| 0.96 | 0.93 | 0.95 | 0.98 | 0.94 | 0.96 | 1200 | |
| 0.90 | 0.97 | 0.94 | 0.92 | 0.96 | 0.94 | 1200 | |
| 0.98 | 0.90 | 0.93 | 0.97 | 0.88 | 0.92 | 1200 | |
| 0.98 | 0.95 | 0.96 | 0.97 | 0.98 | 0.98 | 1200 | |
| 0.95 | 0.94 | 0.95 | 0.96 | 0.95 | 0.95 | 4800 | |
| 0.95 | 0.94 | 0.95 | 0.96 | 0.94 | 0.95 | 4800 | |
| 0.95 | 0.94 | 0.95 | 0.96 | 0.95 | 0.95 | 4800 | |
| 0.94 | 0.94 | 0.94 | 0.95 | 0.95 | 0.95 | 4800 | |
Fig. 11Training and validation accuracy and loss for 2D-M2IC
Fig. 12Training and validation accuracy and loss for 3D-M2IC
The confusion metric and normalized confusion metric for the proposed models (2D-M2IC model, 3D-M2IC).
Fig. 13The ROC-AUC of the proposed 2D-M2IC
Fig. 14The ROC-AUC of the proposed 3D-M2IC
Fig. 15The AD stage prediction for MRI medical images