| Literature DB >> 35720439 |
Ziquan Zhu1, Siyuan Lu1, Shui-Hua Wang1,2, Juan Manuel Gorriz3, Yu-Dong Zhang1,2,4.
Abstract
Aims: Brain diseases refer to intracranial tissue and organ inflammation, vascular diseases, tumors, degeneration, malformations, genetic diseases, immune diseases, nutritional and metabolic diseases, poisoning, trauma, parasitic diseases, etc. Taking Alzheimer's disease (AD) as an example, the number of patients dramatically increases in developed countries. By 2025, the number of elderly patients with AD aged 65 and over will reach 7.1 million, an increase of nearly 29% over the 5.5 million patients of the same age in 2018. Unless medical breakthroughs are made, AD patients may increase from 5.5 million to 13.8 million by 2050, almost three times the original. Researchers have focused on developing complex machine learning (ML) algorithms, i.e., convolutional neural networks (CNNs), containing millions of parameters. However, CNN models need many training samples. A small number of training samples in CNN models may lead to overfitting problems. With the continuous research of CNN, other networks have been proposed, such as randomized neural networks (RNNs). Schmidt neural network (SNN), random vector functional link (RVFL), and extreme learning machine (ELM) are three types of RNNs.Entities:
Keywords: DenseNet; MRI; brain diseases; convolutional neural network; randomized neural network
Year: 2022 PMID: 35720439 PMCID: PMC9204288 DOI: 10.3389/fnsys.2022.838822
Source DB: PubMed Journal: Front Syst Neurosci ISSN: 1662-5137
Contributions of state-of-the-art methods.
| Method | Contribution |
|---|---|
| Noreen et al. ( | A multi-level method using two DensNet201 and Inception-v3 was proposed to diagnose early brain tumors. |
| Amin et al. ( | A model according to the LSTM model method using magnetic resonance images was introduced to classify brain tumors automatically. |
| Amin et al. ( | A deep learning model was used to predict healthy and unhealthy brain tumor slices. |
| Arunkumar et al. ( | A new model was introduced to train MRI brain tumors to identify ROI location. |
| Purushottam Gumaste and Bairagi ( | An algorithm was proposed to extract left and right brain features. This article also introduced different statistical feature extraction methods and used a Support Vector Machine to extract tumor regions from statistical features. |
| Chatterjee and Das ( | A novel method was proposed for the segmentation of brain images. |
| Bhanothu et al. ( | A new method based on R-CNN was presented to detect tumors and mark their location. |
| Natekar et al. ( | Various technologies were compared for brain tumor segmentation models. |
| Aboelenein et al. ( | The HTTU-Net was proposed for brain tumor cutting. |
| Huang et al. ( | The DFNN was proposed. The method introduced DFM blocks and combined SE blocks. |
| Hu and Razmjooy ( | A meta heuristic-based system was presented to detect tumors. |
| Sadad et al. ( | A novel model according to UNET architecture and ResNet50 as the backbone was proposed for the detection of brain tumors. |
| Kalaiselvi et al. ( | The PR2G was proposed to detect and segment tumors. |
| Kaplan et al. ( | Then LBP and αLBP were used to classify the three different types of brain tumors. |
| Khalil et al. ( | The DA clustering was proposed to improve the accuracy of extracting initial contour points to detect three-dimensional magnetic resonance brain tumors better. |
| Khan et al. ( | The PART was introduced to detect brain tumors of grade I to grade IV brain tumors. |
| Ma and Zhang ( | A method was proposed to intelligently detect brain tumors based on a lightweight neural network. |
| Hollon et al. ( | A new method was proposed for the automatic detection of brain tumors by combining SRH 5–7, CNN, and the label-free optical imaging method. |
| Saba et al. ( | A new method was proposed to detect brain tumors. The Grasp cut method was used to segment brain tumor symptoms, and VGG-19 was used to obtain features. |
| Sharif et al. ( | An unsupervised fuzzy set method was introduced for brain tumor segmentation. |
| Xu et al. ( | A new structure was proposed for the early detection of brain tumors. The new structure was mainly composed of five parts: tumor segmentation, morphology, denoising, feature extraction, and classification. |
| Hemanth et al. ( | The HSBPN was proposed to segment MR brain tumor images. |
| Nayef et al. ( | A novel structure was presented for the classification of the MRI dataset. |
| Chen et al. ( | An improved method was introduced for detecting pathological brains. |
| Shoeibi et al. ( | A review was presented on the segmentation of the Covid-19 by DL. |
| Shoeibi et al. ( | A comprehensive survey about the application of DL in the detection of Multiple Sclerosis |
| Sadeghi et al. ( | A survey was presented on the automatic diagnosis of the SZ by AI. |
| Shoeibi et al. ( | A comprehensive review was presented on applying the various AI techniques in the diagnosis of Epileptic seizures. |
| Shoeibi et al. ( | A review of various methods based on DL for automatic diagnosis of SZ by electroencephalogram (EEG) signals was completed. |
| Shoeibi et al. ( | A new model was proposed to detect Epileptic seizures automatically. The proposed model was based on the DL and the fuzzy theory. |
| Odusami et al. ( | A method was proposed for the recognition of AD. They tested two CNN models (DenseNet201 and ResNet18) to perform this task. |
| Razzak et al. ( | A new network (PartialNet) was introduced to detect AD based on MRIs. This network achieved improvements in AD detection. |
| Ashraf et al. ( | Different CNN models were experimented with to detect AD based on transfer learning. Finally, the fine-tuned DenseNet got the highest accuracy (99.05%). |
Figure 1(A)Unhealthy and (B) healthy brain images in the dataset.
Acronym and full explanation.
| Acronym | Full explanation |
|---|---|
| AD | Alzheimer’s disease |
| Acc | Accuracy |
| Avr | Average |
| BN | Batch normalization |
| CNN | Convolution neural network |
| DCNN | Deep convolution neural network |
| DELM | DenseNet-based extreme learning machine |
| DL | Deep learning |
| DRVFL | DenseNet-based random vector functional link |
| DSNN | DenseNet-based Schmidt neural network |
| ELM | Extreme learning machine |
| F1 | F1-score |
| FC | Fully connected |
| ML | Machine learning |
| Pre | Precision |
| RVFL | Random vector functional link |
| RNNs | Randomized neural networks |
| Sen | Sensitivity |
| SNN | Schmidt neural network |
| Spe | Specificity |
| Std | Standard deviation |
The definition of the parameter.
| Parameter | Definition |
|---|---|
|
| The output of the |
|
| The nonlinear transformation |
|
| The given dataset |
|
| The input dimension |
|
| The output dimension |
|
| The weights vector |
|
| The bias of the |
|
| The final output weights |
|
| The output biases of SNN |
|
| The ground-truth label matrix of the dataset |
|
| The input matrix |
| The sigmoid function | |
|
| The number of hidden nodes |
|
| The output matrix of the hidden layer |
Pseudocode of the proposed DSNN.
| Step 1: Load the pre-trained DenseNet. |
| Step 2: Modify the pre-trained DenseNet. |
| Step 2.1 Remove softmax and classification layer from the pre-trainedDenseNet. |
| Step 2.2 Add FC128, ReLU, BN, FC2, softmax, and classification layer. |
| Step 3: Divide the dataset into five groups of the same size and set |
| Step 4: Use the |
| Step 5: Fine-tune the modified DenseNet. |
| Step 5.1: Input is the training set. |
| Step 5.2: Target is the corresponding label. |
| Step 6: Replace the last five layers of the fine-tuned DenseNet with SNN. |
| Step 7: Extract features |
| Step 8: Train the classifier of the DSNN on the extracted features |
| Step 8.1: Input is the extracted features. |
| Step 8.2: The target is the label of the training set. |
| Step 8.3: SNN is the classifier of the DSNN. |
| Step 9: Test the trained DSNN on the test set. |
| Step 10: Report the test classification performance of the trained DSNN. |
| Step 11: Set |
| Step 12: Average test classification performance. |
Figure 2The pipeline of the proposed DSNN.
Figure 3Backbone of the proposed DSNN. (A) The general view of DenseNet.(B) The modifications in the pre-trained DenseNet.
Figure 4Structure of SNN.
Figure 5Thestructures of (A) RVFL and (B) ELM.
The hyper-parameter settings of the proposed DSNN.
| Hyper-parameter | Value |
|---|---|
| Mini-batch size | 10 |
| Max-epoch | 4 |
| Learning rate | 10−4 |
| Number of the hidden nodes V | 400 |
The classification performance based on five-fold cross-validation (unit: %).
| Methods | Fold | Acc | Sen | Spe | Pre | F1 |
|---|---|---|---|---|---|---|
| DSNN(Ours) | F 1 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 |
| F 2 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | |
| F 3 | 94.87 | 100.00 | 50.00 | 94.59 | 97.22 | |
| F 4 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | |
| F 5 | 97.44 | 100.00 | 75.00 | 97.22 | 98.59 | |
| Avr |
|
|
|
|
| |
| Std | ±2.05 | ±0.00 | ±20.00 | ±2.17 | ±1.11 | |
| DRVFL(Ours) | F 1 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 |
| F 2 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | |
| F 3 | 89.74 | 100.00 | 0.00 | 89.74 | 94.59 | |
| F 4 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | |
| F 5 | 97.44 | 100.00 | 75.00 | 97.22 | 98.59 | |
| Avr | 97.44 | 100.00 | 75.00 | 97.39 | 98.64 | |
| Std | ±3.97 | ±0.00 | ±38.73 | ±3.97 | ±2.10 | |
| DELM(Ours) | F 1 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 |
| F 2 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | |
| F 3 | 92.31 | 100.00 | 25.00 | 92.11 | 95.89 | |
| F 4 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | |
| F 5 | 97.44 | 100.00 | 75.00 | 97.22 | 98.59 | |
| Avr | 97.95 | 100.00 | 80.00 | 97.87 | 98.90 | |
| Std | ±2.99 | ±0.00 | ±29.15 | ±3.07 | ±1.60 | |
| Fine-tuned DenseNet | F 1 | 87.50 | 86.11 | 100.00 | 100.00 | 92.54 |
| F 2 | 82.05 | 80.00 | 100.00 | 100.00 | 88.89 | |
| F 3 | 89.74 | 88.57 | 100.00 | 100.00 | 93.94 | |
| F 4 | 85.00 | 83.33 | 100.00 | 100.00 | 90.91 | |
| F 5 | 79.49 | 77.14 | 100.00 | 100.00 | 87.10 | |
| Avr | 84.76 | 83.03 | 100.00 | 100.00 | 90.67 | |
| Std | ±3.67 | ±4.10 | ±0.00 | ±0.00 | ±2.46 | |
| AlexNet-SNN | F 1 | 89.74 | 100.00 | 0.00 | 89.74 | 94.59 |
| F 2 | 89.74 | 100.00 | 0.00 | 89.74 | 94.59 | |
| F 3 | 90.00 | 97.22 | 25.00 | 92.11 | 94.59 | |
| F 4 | 90.00 | 97.22 | 25.00 | 92.11 | 94.59 | |
| F 5 | 89.74 | 97.14 | 25.00 | 91.89 | 94.44 | |
| Avr | 89.84 | 98.32 | 15.00 | 91.12 | 94.56 | |
| Std | ±0.13 | ±1.38 | ±12.25 | ±1.13 | ±0.06 | |
| ResNet-18-SNN | F 1 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 |
| F 2 | 97.50 | 100.00 | 75.00 | 97.30 | 98.63 | |
| F 3 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | |
| F 4 | 94.87 | 100.00 | 50.00 | 94.59 | 97.22 | |
| F 5 | 94.87 | 97.14 | 75.00 | 97.14 | 97.14 | |
| Avr | 97.45 | 99.43 | 80.00 | 97.81 | 98.60 | |
| Std | ±2.29 | ±1.14 | ±18.71 | ±2.03 | ±1.26 | |
| ResNet-50-SNN | F 1 | 95.00 | 94.44 | 100.00 | 100.00 | 97.14 |
| F 2 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | |
| F 3 | 97.44 | 97.14 | 100.00 | 100.00 | 98.55 | |
| F 4 | 95.00 | 100.00 | 50.00 | 94.74 | 97.30 | |
| F 5 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | |
| Avr | 97.49 | 98.32 | 90.00 | 98.95 | 98.60 | |
| Std | ±2.24 | ±2.23 | ±20.00 | ±2.10 | ±1.24 | |
| VGG-SNN | F 1 | 97.50 | 100.00 | 75.00 | 97.30 | 98.63 |
| F 2 | 87.50 | 94.44 | 25.00 | 91.89 | 93.15 | |
| F 3 | 94.87 | 97.14 | 75.00 | 97.14 | 97.14 | |
| F 4 | 89.74 | 100.00 | 0.00 | 89.74 | 94.59 | |
| F 5 | 87.18 | 88.57 | 75.00 | 96.88 | 92.54 | |
| Avr | 91.36 | 96.03 | 50.00 | 94.59 | 95.21 | |
| Std | ±4.12 | ±4.26 | ±31.62 | ±3.16 | ±2.33 | |
| Restricted DenseNet-SNN | F 1 | 94.87 | 94.29 | 100.00 | 100.00 | 97.06 |
| F 2 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | |
| F 3 | 97.37 | 97.06 | 100.00 | 100.00 | 98.51 | |
| F 4 | 94.87 | 100.00 | 50.00 | 94.59 | 97.22 | |
| F 5 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | |
| Avr | 97.42 | 98.27 | 90.00 | 98.92 | 98.56 | |
| Std | ±2.09 | ±1.83 | ±1.97 | ±2.09 | ±1.69 |
The bold values are results of our proposed model.
Figure 6ROC curve of DSNN.
Figure 7Modelcomparison. (A) Comparison of three proposed models (unit:%). (B) Comparison with the fine-tuned DenseNet (unit: %).(C) The classification performance of the proposed DSNN withdifferent backbones (unit: %). (D) Comparison with thespiking neural network (unit: %).
The final result of the spiking neural network (unit: %).
| Model | Acc | Sen | Spe | Pre | F1 |
|---|---|---|---|---|---|
| Spiking neural network | 86.05 | 100.00 | 0.00 | 96.05 | 92.50 |
|
|
|
|
|
|
|
The bold values are results of our proposed model.
Figure 8Explainability of the proposed DSNN.
Comparison with other state-of-the-art methods (unit: %).
| Methods | Sen | Spe | Pre | Acc | F1 |
|---|---|---|---|---|---|
| ANN (Arunkumar et al., | 89.00 | - | - | 92.14 | - |
| PR2G (Kalaiselvi et al., | 98.46 | - | - | 83.90 | - |
| SRH + CNNs (Hollon et al., | - | - | - |
| - |
| BPNN (Hemanth et al., | 57.54 | 54.50 | 91.71 | 57.23 | 70.72 |
| LVQNN (Nayef et al., | 59.94 | 61.00 | 93.08 | 60.05 | 72.92 |
| LRC (Chen et al., | 100.00 | 58.50 | 95.47 | 95.74 | 97.68 |
| DSNN (Ours) |
|
|
|
|
|
Bold means the best results, - means not available.
Figure 9Comparison with other state-of-the-art methods (unit: %).