| Literature DB >> 36010183 |
Sarang Sharma1, Sheifali Gupta1, Deepali Gupta1, Ayman Altameem2, Abdul Khader Jilani Saudagar3, Ramesh Chandra Poonia4, Soumya Ranjan Nayak5.
Abstract
Alzheimer's disease (AD) is a degenerative condition of the brain that affects the memory and reasoning abilities of patients. Memory is steadily wiped out by this condition, which gradually affects the brain's ability to think, recall, and form intentions. In order to properly identify this disease, a variety of manual imaging modalities including CT, MRI, PET, etc. are being used. These methods, however, are time-consuming and troublesome in the context of early diagnostics. This is why deep learning models have been devised that are less time-intensive, require less high-tech hardware or human interaction, continue to improve in performance, and are useful for the prediction of AD, which can also be verified by experimental results obtained by doctors in medical institutions or health care facilities. In this paper, we propose a hybrid-based AI-based model that includes the combination of both transfer learning (TL) and permutation-based machine learning (ML) voting classifier in terms of two basic phases. In the first phase of implementation, it comprises two TL-based models: namely, DenseNet-121 and Densenet-201 for features extraction, whereas in the second phase of implementation, it carries out three different ML classifiers like SVM, Naïve base and XGBoost for classification purposes. The final classifier outcomes are evaluated by means of permutations of the voting mechanism. The proposed model achieved accuracy of 91.75%, specificity of 96.5%, and an F1-score of 90.25. The dataset used for training was obtained from Kaggle and contains 6200 photos, including 896 images classified as mildly demented, 64 images classified as moderately demented, 3200 images classified as non-demented, and 1966 images classified as extremely mildly demented. The results show that the suggested model outperforms current state-of-the-art models. These models could be used to generate therapeutically viable methods for detecting AD in MRI images based on these results for clinical prospective.Entities:
Keywords: Alzheimer’s disease; DenseNet121; DenseNet201; SVM; XGBoost; convolutional neural network; deep learning; gaussian NB
Year: 2022 PMID: 36010183 PMCID: PMC9406825 DOI: 10.3390/diagnostics12081833
Source DB: PubMed Journal: Diagnostics (Basel) ISSN: 2075-4418
Comparison of existing state-of-art models.
| Citation | Approach | Objective | Challenges of the Approach |
|---|---|---|---|
| [ | Non Linear SVM with 2D CNN | To develop an automated technique to classify normal, early and late mild AD subjects. | Dataset consisted of 1167 MRI images. It was able to achieve 75% while performing Bin.c. |
| [ | 2D CNN, 3D CNN 3D CNN-SVM | To distinguish AD and MCI individuals from normal individuals and to improve value based care of affected individuals in medical facilities. | Dataset contained 3127, 3T T1-weighted images. It performed tertiary classification and was able to an accuracy of 88.9%. It also aims to focus on reverting MCI individuals to normal individuals, predict AD progression and improve diagnosis of AD in future. |
| [ | GoogleNet, AlexNet, VGGNet16VGGNet19SqueezeNetResNet18 | To detect AD on MRI scans using D.L techniques. | Dataset consisted of 177 images. It performed Bin.c and achieved an accuracy of 84.38%. To include other neuro-imaging modalities such as PET scans or features in the system to take different aspects of AD into consideration. |
| [ | Data Augmentation, CNN. | To classify AD by using Cross-Modal Transfer Learning | Dataset contained 416; sMRI image scans and it implemented Bin.c and achieved an accuracy of 83.57%. To proceed with a longitudinal dataset and develop a method based on spatial optimization of ROI. |
| [ | DTCWT, PCA, FNN | To develop a CAD system to early diagnose AD individuals. | Dataset contained 416; T1- weighted image scans and it performed Bin.c and achieved an accuracy of 90.06%. Various feature reduction methods such as ICA, LDA and PCA were utilized for swarm optimization. |
| [ | SVM, CNN | To classify AD from MCI by using semi-supervised SVM-CNN. | Dataset contained 359; T1- weighted images and it performed Bin.c and achieved an accuracy of 82.91%. To distinguish brain MRI images semi semi-supervised SVM is applied. |
| [ | SVM-REF, CNN | To classify AD by using SVM-REF-CNN. | Dataset contained 1167; T1-weighted image scans and it performed Bin.c and achieved an accuracy of 81%. To distinguish brain images by using SVM-REF. |
| [ | 2D-CNN, VGG16 | To classify AD by using ensemble based CNN. | Dataset contained 798; T1-weighted image scans and it performed Bin.c and achieved an accuracy of 90.36%. To distinguish AD from MCI images by using 2D-CNN. |
| [ | SVM, CNN | To distinguish MCI from AD by using an SVM classifier with a linear kernel. | Dataset contained 1167; T1-weighted image scans and it performed Bin.c and achieved an accuracy of 69.37%. To distinguish AD from MCI images by using SVM-CNN. |
| [ | SVM, k-NN, CNN | To distinguish MCI from AD by using SVM and k-NN. | Dataset contained 1311; T1 & T2 weighted image scans and it performed Bin.c and achieved an accuracy of 75%. To distinguish AD from MCI images by using SVN-CNN, KNN. |
Figure 1Block Diagram of Proposed Research Model.
Kaggle available Alzheimer’s Disease Dataset.
| Dataset Source | Class Name | Training Images | Validating Images | Total Images |
|---|---|---|---|---|
| Kaggle | M.D | 717 | 179 | 896 |
Figure 2Alzheimer’s Disease MRI Dataset: (a) M.D, (b) Mod.D (c) N.D and (d) V.M.D.
Figure 3Flipping applied to dataset (a) original (b) horizontal flipping (c) vertical flipping (d) 90 degree anticlockwise (e) 270 degree anticlockwise and (f) brightness factor 0.7.
Alzheimer’s Dataset with Augmentation.
| S.No. | Name of the Class | Number of Images before Augmentation | Images after Augmentation | |
|---|---|---|---|---|
| Training Images | Validating Images | |||
| 1 | M.D | 896 | 2150 | 538 |
| 2 | Mod.D | 64 | 512 | 128 |
| 3 | N.D | 3200 | 2800 | 700 |
| 4 | V.M.D | 1966 | 3145 | 787 |
Layers Description of Conventional Neural Network DenseNet121.
| Block Name | Layer Name | Input Size | Output Size | Filter Size | Number of Filters | Number of Times Block Run |
|---|---|---|---|---|---|---|
| Conv_1 | Conv_1_1 | 224 × 224 | 112 × 112 | 7 × 7 | 64 | 1 |
| Conv_2 | Conv_2_1:Conv_2_6 | 112 × 112 | 56 × 56 | 1 × 1 | 128 | 6 |
| Conv_3 | Conv_3_1:Conv_3_12 | 56 × 56 | 28 × 28 | 1 × 1 | 256 | 12 |
| Conv_4 | Conv_4_1:Conv_4_48 | 28 × 28 | 14 × 14 | 1 × 1 | 512 | 48 |
| Conv_5 | Conv_5_1:Conv_5_32 | 14 × 14 | 7 × 7 | 1 × 1 | 1024 | 32 |
Layers Description of Conventional Neural Network 201.
| Block Name | Layer Name | Input Size | Output Size | Filter Size | Number of Filters | Number of Times Block Run |
|---|---|---|---|---|---|---|
| Conv_1 | Conv_1_1 | 224 × 224 | 112 × 112 | 7 × 7 | 64 | 1 |
| Conv_2 | Conv_2_1:Conv_2_12 | 112 × 112 | 56 × 56 | 1 × 1 | 128 | 12 |
| Conv_3 | Conv_3_1:Conv_3_24 | 56 × 56 | 28 × 28 | 1 × 1 | 512 | 24 |
| Conv_4 | Conv_4_1:Conv_4_96 | 28 × 28 | 14 × 14 | 1 × 1 | 896 | 96 |
| Conv_5 | Conv_5_1:Conv_5_64 | 14 × 14 | 7 × 7 | 1 × 1 | 1920 | 64 |
Filter visualization and image conception for convolution layers of DenseNet121.
| Name of Corresponding Block | Filter for First Convolution Layer | Image for First Convolution Layer | Filter for Last Convolution Layer | Image for Last Convolution Layer |
|---|---|---|---|---|
| Conv_1 |
|
|
|
|
| Conv_2 |
|
|
|
|
| Conv_3 |
|
|
|
|
| Conv_4 |
|
|
|
|
| Conv_5 |
|
|
|
|
Filter visualization and image visualization for each convolution layer of DenseNet 201.
| Name of Block | Filter for First Convolution Layer | Image for First Convolution Layer | Filter for last Convolution Layer | Image for Last Convolution Layer |
|---|---|---|---|---|
| Conv_1 |
|
|
|
|
| Conv_2 |
|
|
|
|
| Conv_3 |
|
|
|
|
| Conv_4 |
|
|
|
|
| Conv_5 |
|
|
|
|
Layers Description of Hybrid DenseNet121 Model.
| Block Name | Layer Name | Input Size | Output Size | Filter Size | Number of Filters | Number of Times Block Run |
|---|---|---|---|---|---|---|
| Conv_5 | Conv_5_1:Conv_5_32 | 14 × 14 | 7 × 7 | 1 × 1 | 1024 | 32 |
| Conv_5 | Machine Learning Classifiers | 7 × 7 | 4 × 1 | 1 × 1 | 1024 | 32 |
| Dense_4 | Dense | 4 × 1 | 4 × 1 | N.A | N.A | 1 |
Layers Description of Hybrid DenseNet201 Model.
| Block Name | Layer Name | Input Size | Output Size | Filter Size | Number of Filters | Number of Times Block Run |
|---|---|---|---|---|---|---|
| Conv_5 | Conv_5_1:Conv_5_64 | 14 × 14 | 7 × 7 | 1 × 1 | 1920 | 64 |
| Conv_5 | Machine Learning Classifiers | 7 × 7 | 4 × 1 | 1 × 1 | 1920 | 64 |
| Dense_4 | Dense | 4 × 1 | 4 × 1 | N.A | N.A | 1 |
Figure 4Categorical Hinge Loss vs. Epoch Curve for hybrid DenseNet121 model with classifiers (a) SVM, (b) Gaussian NB and (c) XG Boost.
Training and Validation Loss of Hybrid DenseNet121 Model with Varying Epochs and Fixed Batch Size 64.
| SVM | GNB | XG | ||||
|---|---|---|---|---|---|---|
| Epoch | Train Loss | Valid Loss | Train Loss | Valid Loss | Train Loss | Valid Loss |
| 200 | 0.264 | 0.554 | 0.265 | 0.497 | 0.262 | 0.531 |
| 400 | 0.14 | 0.467 | 0.141 | 0.402 | 0.125 | 0.45 |
| 600 | 0.089 | 0.422 | 0.088 | 0.348 | 0.086 | 0.405 |
| 800 | 0.068 | 0.394 | 0.065 | 0.323 | 0.059 | 0.384 |
| 1000 | 0.051 | 0.313 | 0.051 | 0.38 | 0.05 | 0.372 |
Figure 5Confusion Matrix of Hybrid DenseNet121 with Three Machine Learning Classifiers: (a) SVM, (b) Gaussian NB and (c) XG Boost.
Figure 6Confusion Matrix Parameters of Hybrid DenseNet121 with Three Machine Learning Classifiers.
Confusion Matrix Parameters of Hybrid DenseNet121 Model (in %).
| SVM | GNB | XG | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Type | P | S | Sp | F1 | P | S | Sp | F1 | P | S | Sp | F1 |
| Average | 92 | 89 | 96 | 90 | 88 | 89 | 96 | 89 | 90 | 83 | 95 | 85 |
| Accuracy | 89.89 | 89.18 | 88.25 | |||||||||
Figure 7Categorical hinge loss vs. epoch curve for hybrid DenseNet201 model with classifiers (a) SVM, (b) Gaussian NB and (c) XG Boost.
Training and Validation Loss of Hybrid DenseNet201 Model with Varying Epochs and Fixed Batch Size 64.
| SVM | GNB | XG | ||||
|---|---|---|---|---|---|---|
| Epoch | Train Loss | Valid Loss | Train Loss | Valid Loss | Train Loss | Valid Loss |
| 200 | 0.16 | 0.418 | 0.158 | 0.427 | 0.157 | 0.459 |
| 400 | 0.075 | 0.326 | 0.073 | 0.348 | 0.07 | 0.373 |
| 600 | 0.047 | 0.294 | 0.047 | 0.317 | 0.046 | 0.353 |
| 800 | 0.035 | 0.292 | 0.033 | 0.299 | 0.031 | 0.326 |
| 1000 | 0.027 | 0.291 | 0.028 | 0.265 | 0.025 | 0.318 |
Figure 8Confusion Matrix of DenseNet201 with Three Machine Learning Classifiers: (a) SVM, (b) Gaussian NB and (c) XG Boost.
Figure 9Confusion Matrix Parameters of DenseNet201 with Three Machine Learning Classifiers.
Confusion Matrix Parameters of DenseNet201 (in %).
| SVM | GNB | XG | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Type | P | S | Sp | F | P | S | Sp | F | P | S | Sp | F |
| Average | 93 | 92 | 96 | 92 | 93 | 89 | 97 | 90 | 93 | 92 | 98 | 92 |
| Accuracy | 91.03 | 91.75 | 91.13 | |||||||||
Figure 10Confusion Matrix Parameters of both Hybrid DenseNet Models with Machine Learning Classifiers depicting (a) Precision, (b) Sensitivity, (c) Specificity and (d) F1-Score.
Figure 11Average Performance Parameters of both Hybrid DenseNet Models with Machine Learning Classifiers.
Comparison with existing state-of-art models.
| Study | Dataset Source | No. of Images | Technique Used | Accuracy |
|---|---|---|---|---|
| Rallabandi et al. [ | ADNI | 1167 | SVM with D.L | 75% |
| Feng et al. [ | ADNI | 3127 | 2D-CNN with D.L | 82.57% |
| Ebrahimi-Ghahnavieh et al. [ | ADNI | 177 | DenseNet-201 | 84.38% |
| Aderghal, K. et al. [ | OASIS | 416 | Cross-Modal Transfer Learning | 83.57% |
| Jha et al. [ | OASIS | 416 | DTCWT and PCA with FNN | 90.06% |
| Filipovych et al. [ | ADNI | 359 | SVM, CNN | 82.91% |
| Rathore et al. [ | ADNI | 1167 | SVM-REF, CNN | 81% |
| Kang et al. [ | ADNI | 798 | 2D-CNN, VGG16 | 90.36% |
| Li et al. [ | ADNI | 1167 | SVM, CNN | 69.37% |
| Venugopalan et al. [ | ADNI | 1311 | SVM, k-NN, CNN | 75% |
| Proposed Methodology | Kaggle | 6400 | DenseNet201-Gaussian NB | 91.75% |
| DenseNet201-XG Boost | 91.13% | |||
| DenseNet201-SVM | 91.03% |