| Literature DB >> 33888650 |
Gokhan Guney1, Busra Ozgode Yigin1, Necdet Guven1, Yasemin Hosgoren Alici2, Burcin Colak3, Gamze Erzin4, Gorkem Saygili1.
Abstract
Deep learning (DL) algorithms have achieved important successes in data analysis tasks, thanks to their capability of revealing complex patterns in data. With the advance of new sensors, data storage, and processing hardware, DL algorithms start dominating various fields including neuropsychiatry. There are many types of DL algorithms for different data types from survey data to functional magnetic resonance imaging scans. Because of limitations in diagnosing, estimating prognosis and treatment response of neuropsychiatric disorders; DL algorithms are becoming promising approaches. In this review, we aim to summarize the most common DL algorithms and their applications in neuropsychiatry and also provide an overview to guide the researchers in choosing the proper DL architecture for their research.Entities:
Year: 2021 PMID: 33888650 PMCID: PMC8077051 DOI: 10.9758/cpn.2021.19.2.206
Source DB: PubMed Journal: Clin Psychopharmacol Neurosci ISSN: 1738-1088 Impact factor: 2.582
Fig. 1The diagram shows deep learning is a subfield of machine learning, which is a subfield of artificial intelligence.
Fig. 2(A) Perceptron, the smallest part of the artificial neural network (ANN) model, is defined by the linear function y = W.x + b. In biological neural networks, information from the axon is collected by the dendrites and processed by the cell body to generate electrical pulses and chemical signals. Communication between two different neurons is achieved by means of neurotransmitters in the synapses between the axons and dendrites of two adjacent neurons when the neuron meets the threshold level. Similarly, in ANNs, each input, xi, is weighted by wi according to its contribution to obtain the final output, f (y). The output unit is obtained by passing the weighted sum of the inputs through an activation function. (B) ANN architecture with multiple layers. It has 1 input layer (the first layer), 3 hidden layers (in between layers), and finally 1 output layer (the last layer) with 1 output units.
Fig. 3Convolutional neural network (CNN). A CNN contains two basic parts: feature extraction and classification. The feature extraction part consists of successive convolutional and pooling layers. A convolutional layer applies convolutional filters called a kernel to the image for exploring low and high-level structures. These structures are obtained by shifting these kernels, so called convolution, in the image with a set of weights. After multiplying the elements of these kernels with the corresponding receiving field elements, a feature map is obtained. These maps are passed through nonlinear activation function (e.g., a rectified linear unit). The task of pooling layer is to reduce the feature map size and the total number of parameters to be optimized in the network. It works by gathering similar information in the neighborhood of the receptive field and find a representative value (e.g., maximum or average) within this local region. Flatten layer converts matrices from the convolution layers into a one-dimensional array for the next layer. Fully connected layer computes the final outputs using back propagation and gradient descent as for standard artificial neural networks.
Fig. 4Recurrent neural network: Given architecture has an input layer X, hidden layer S and output layer ŷ. In the network, Xt, ŷt, and St define the current input, output and states respectively. U and W are the weights of the relevant layer and V is the output function. St is calculated using the information from previous state as: St = f (UXt + WSt-1) and, ŷt is calculated as: ŷt = V(St).
Fig. 5The metaphor used by Ian Goodfellow to explain the generative adversarial networks (GANs) model. GANs consists of two different network structures; generator and discriminator networks. While the discriminator network creates new data from a sample database, the discriminator network tries to distinguish between real and fake samples by looking at the data produced by the generator with some noise.
Fig. 6Flow diagram for study selection (modified from Preferred Reporting Items for Systematic Reviews and Meta-Analyses: The PRISMA Statement). ANNs, artificial neural networks; CNNs, convolutional neural networks; RNNs, recurrent neural networks; GANs, generative adversarial networks.
Studies using ANN in neuropsychiatry
| Reference | Method | Year | Modality | Application | Result |
|---|---|---|---|---|---|
| Vyškovský | ANN | 2016 | MRI | Schizophrenia classification | Overall accuracy = 68% |
| Jafri and Calhoun 2006 [ | ANN | 2006 | fMRI | Schizophrenia classification | Accuracy = 76% |
| Fonseca | ANN | 2018 | Array collection data | Classification of bipolar and schizophrenia disorders | Accuracy = 90% |
| Lins | ANN | 2017 | Array collection data | Classification of mild cognitive impairment and dementia | Sensitivity = 98% and Specificity = 96% |
| Narzisi | ANN | 2015 | Array collection data | Classification of children with a positive response to TAU | Accuracy = 89.24% |
ANN, artificial neural network; MRI, magnetic resonance imaging, fMRI, functional MRI; TAU, treatment as usual.
Studies using CNN in neuropsychiatry
| Reference | Method | Year | Modality | Application | Result |
|---|---|---|---|---|---|
| Gupta | Simple CNN + Sparse Automatic Encoder | 2013 | sMRI | Classification of MCI and AD | Accuracy = 94.7% for NC vs. AD |
| Payan and Montana 2015 [ | 3D CNN | 2015 | 3D MRI | Classification of MCI and AD | Accuracy = 95.4% for NC vs. AD |
| Hosseini-asl | 3D CNN with pre- trained 3D con-volution auto-matic encoder | 2018 | 3D sMRI | Classification of MCI and AD | Accuracy = 99.3% for NC vs. AD |
| Wang | CNN | 2018 | sMRI | Classification of AD | Accuracy = 97.65% |
| Duc | 3D CNN | 2020 | fMRI | Classification of AD | Accuracy = 85.27% for NC vs. AD |
| Sarraf and Ghassem 2016 [ | LeNet and GoogleNet | 2016 | sMRI and fMRI | Classification of AD | Accuracy = 94.32% for NC vs. AD with the fMRI |
| Spasov | 3D CNN | 2018 | sMRI, genetic mea-sures (APOe4) and clinical assessment | Classification of AD | Accuracy = 99% for NC vs. AD |
| Liu | ResNet and | 2020 | sMRI | Classification of MCI and AD | Accuracy = 88.9% |
| Farooq | GoogleNet and ResNet | 2017 | sMRI | Classification of AD, EMCI, and LMCI | Accuracy = 98.88% for GoogleNet |
| Korolev | Plain 3D CNN (VoxCNN) and ResNet with six VoxRes blocks | 2017 | 3D sMRI | Classification of AD, EMCI, and LMCI | Accuray (VoxCNN) = 79% for NC vs. AD |
| Senanayake | ResNet, DenseNet, and GoogleNet | 2018 | 3D MR volumes and neuropsychological measure based feature vectors | Classification of MCI and AD | Accuracy = 79% for NC vs. AD |
| Zou | 3D CNN | 2017 | Resting-state fMRI signals | Classification of ADHD | Accuracy = 65.67% |
| Zou | Multi-modality 3D CNN | 2017 | fMRI and sMRI | Classification of ADHD | Accuracy = 69.15% |
| Chen | 3D CNN and 2D CNN | 2019 | A new form of representation of multi channel EEG data | Detection of personalized spatial-frequency ab-normality in EEGs from children with ADHD | Accuracy = 90.29% ± 0.58% |
| Campese | SVM, 2D CNN, and three different 3D architectures (VNet, UNet, and LeNet) | 2019 | 2D and 3D sMRI | Classification of SZ and BP | AUC score: 86.30 ± 9.35 using VNet + SVM for Dataset A |
| Choi | 3D CNN | 2017 | FP-CIT SPECT | Classification of PD | Accuracy = 96% for the PPMI dataset |
CNN, convolutional neural network; 3D, three-dimensional; 2D, two-dimensional; ResNet, residual networks; DenseNet, densely connected networks; MRI, magnetic resonance imaging; fMRI, functional MRI; sMRI, structural MRI; EEG, electroencephalography; FP-CIT, dopamine-trans-porterscintigrafie; SPECT, single photon emission computerized tomography; MCI, mild cognitive impairment; AD, Alzheimer’s disease; EMCI/LMCI, early/late mild cognitive impairment; ADHD, attention deficit and hyperactivity disorder; SZ, spectrum disorder; BP, bipolar disorder; PD, Parkinson’s disease; NC, normal cognitive; AUC, area under the curve; SVM, support vector machine; PPMI, Parkinsons progression markers initiative; SNUH, Seoul National University Hospital.
Studies using RNN in neuropsychiatry
| Reference | Method | Year | Modality | Application | Result | |
|---|---|---|---|---|---|---|
| Petrosian | RNN | 2000 | EEG | Prediction of epileptic seizures | Existence of preictal stage in some minutes reported as feasible to predict seizure | |
| Petrosian | RNN | 2001 | EEG | Early prediction of AD | Sensitivity = 80% | |
| Wang | LSTM | 2018 | Array collection data | AD progression prediction | Accuracy = 99% ± 0.0043 | |
| Dakka | LSTM | 2017 | 4D fMRI | Learning invariant markers of schizophrenia disorder | Average accuracy using LSTM = 66.4% | |
| Kumar | RNN | 2019 | CT, MRI, and PET | Classification of dementia, AD, and autism disorders | Dementia | Accuracy = 82.8% |
| AD | Accuracy = 72.2% | |||||
| Autism | Accuracy = 78.2% | |||||
| BRNN | Dementia | Accuracy = 95.3% | ||||
| AD | Accuracy = 89.6% | |||||
| Autism | Accuracy = 91.9% | |||||
| Talathi 2017 [ | GRU | 2017 | EEG | Early epileptic seizure detection | Accuracy = 99.6% | |
| Che | GRU | 2017 | Parkinson’s progression markers initiative (PPMI) challenge dataset | Personalized predictions of Parkinson’s disease | Personalized LR | RMSE = 0.658 |
| Personalized SVM | RMSE = 0.695 | |||||
| Multiclass LR | RMSE = 0.719 | |||||
| Multiclass SVM | RMSE = 0.742 | |||||
| LSTM | RMSE = 0.785 | |||||
| KNN | RMSE = 0.957 | |||||
| Yao | IndRNN | 2019 | EEG | Classification of epileptic seizure | IndRNN | Average accuracy = 87% ± 0.03 |
| LSTM | Average accuracy = 84.4% ± 0.02 | |||||
| CNN | Average accuracy = 82.9% ± 0.02 | |||||
RNN, recurrent neural network; LSTM, long-short term memory; BRNN, bidirectional RNN; GRU, gated recurrent unit; IndRNN, independent RNN; EEG, electroencephalography; 4D, four-dimensional; MRI, magnetic resonance imaging; fMRI, functional MRI; CT, computed tomography; PET, positron emission tomography; AD, Alzheimer’s disease; CNN, convolutional neural network; RCNN, recurrent CNN; SVM, support vector machine; LR, logistic regression; KNN, K nearest neighbors; RMSE, root mean square error.
Studies using RNN in neuropsychiatry
| Reference | Method | Year | Modality | Application | Result |
|---|---|---|---|---|---|
| Truong | DCGAN | 2018 | EEG | Seizure prediction | AUC = 80% |
| Wei | cGAN | 2018 | Multimodal MRI | Predicting myelin content | Dice index between ground truth and prediction = 0.83 |
| Palazzo | LSTM and cGAN | 2017 | EEG | Reading the mind | Maximum test accuracy = 83.9% for the LSTM-based |
RNN, recurrent neural network; DCGAN, deep convolutional generative adversarial network; cGAN, conditional generative adversarial network; EEG, electroencephalography; MRI, magnetic resonance imaging; AUC, area under the curve; LSTM, long-short term memory.