| Literature DB >> 35265830 |
Zaniar Ardalan1, Vignesh Subbian1,2.
Abstract
Deep learning algorithms have been moderately successful in diagnoses of diseases by analyzing medical images especially through neuroimaging that is rich in annotated data. Transfer learning methods have demonstrated strong performance in tackling annotated data. It utilizes and transfers knowledge learned from a source domain to target domain even when the dataset is small. There are multiple approaches to transfer learning that result in a range of performance estimates in diagnosis, detection, and classification of clinical problems. Therefore, in this paper, we reviewed transfer learning approaches, their design attributes, and their applications to neuroimaging problems. We reviewed two main literature databases and included the most relevant studies using predefined inclusion criteria. Among 50 reviewed studies, more than half of them are on transfer learning for Alzheimer's disease. Brain mapping and brain tumor detection were second and third most discussed research problems, respectively. The most common source dataset for transfer learning was ImageNet, which is not a neuroimaging dataset. This suggests that the majority of studies preferred pre-trained models instead of training their own model on a neuroimaging dataset. Although, about one third of studies designed their own architecture, most studies used existing Convolutional Neural Network architectures. Magnetic Resonance Imaging was the most common imaging modality. In almost all studies, transfer learning contributed to better performance in diagnosis, classification, segmentation of different neuroimaging diseases and problems, than methods without transfer learning. Among different transfer learning approaches, fine-tuning all convolutional and fully-connected layers approach and freezing convolutional layers and fine-tuning fully-connected layers approach demonstrated superior performance in terms of accuracy. These recent transfer learning approaches not only show great performance but also require less computational resources and time.Entities:
Keywords: convolutional neural network; domain adaptation; fine tuning; medical imaging; neuroimaging; transfer learning
Year: 2022 PMID: 35265830 PMCID: PMC8899512 DOI: 10.3389/frai.2022.780405
Source DB: PubMed Journal: Front Artif Intell ISSN: 2624-8212
Figure 1The LeNet architecture for letter recognition, one of the first CNN architectures for image processing. FC, Fully-connected layer. The architecture was designed for handwritten digit recognition.
Figure 2Demonstration of transferring weights of a convolution filter from source domain to target domain.
Figure 3Flowchart illustrating literature search process and extraction of studies meeting the scoping review inclusion criteria.
Frequency of datasets in source and target domains.
|
|
|
|
|---|---|---|
| ImageNet | 19 | 0 |
| Alzheimer's Disease neuroimaging initiative (ADNI) | 17 | 26 |
| Human Connectome Project (HCP) | 1 | 3 |
| Brain Tumor Segmentation (BraTS) | 1 | 2 |
| The Open Access Series of Imaging Studies (OASIS) | 1 | 3 |
| Nathan Kline Institute-Rockland Sample (NKI-RS) Nooner et al., | 0 | 1 |
| The National Alliance for Medical Imaging Computing (NAMIC) | 1 | 1 |
| Kirby | 1 | 1 |
| Rotterdam Scan Study (RSS) | 1 | 1 |
| MRBrains | 1 | 1 |
| Internet Brain Segmentation Repository (IBSR) | 1 | 1 |
| MS Lesions | 1 | 1 |
| Brain-computer interface (BCI) Blankertz et al., | 0 | 1 |
| High gamma dataset (HGD) | 0 | 1 |
| Private datasets | 9 | 12 |
| WHO grade status | 0 | 1 |
| MS dataset | 0 | 1 |
| The Cancer Genome Atlas (TCGA) | 0 | 1 |
| Harvard The Whole Brain Atlas (AANLIB) | 0 | 1 |
| Autism Brain Imaging Data Exchange (ABIDE) Li L et al., | 1 | 2 |
| UK Bio-Bank (UKBB) | 0 | 1 |
| An Asian Cohort Saba et al., | 0 | 1 |
| National Alzheimer's Coordinating Center (NACC) | 0 | 1 |
| Hammers Adult Atlases (HAA) | 1 | 0 |
| Multi-Atlas Labeling Challenge (MALC) | 1 | 0 |
| Functional connectivity dataset | 1 | 0 |
MIDAS—Collection NAMIC: Public Data.
Databases | Kennedy Krieger Institute.
MRBrainS13 | Evaluation framework for MR Brain Image Segmentation.
NITRC: Longitudinal Multiple Sclerosis Lesion Imaging Archive: Tool/Resource Info.
The Cancer Genome Atlas Program—National Cancer Institute.
NACC Researcher home page, NACC, Alzheimer's disease research, FTLD, NIA/NIH, database, neuropathology.
2012 MICCAI Multi-Atlas Labeling Challenge Data.
Data types for source and target domains.
|
|
| ||
|---|---|---|---|
|
|
|
|
|
| MRI | 31 | MRI | 36 |
| Natural Images | 18 | PET | 4 |
| fMRI | 5 | fMRI | 8 |
| EEG | 4 | EEG | 2 |
| PET | 3 | CSF | 1 |
| Cerebrospinal fluid (CSF) | 1 | ||
Figure 4Residual block of the ResNet algorithm (left) and the ResNet 12 architecture (right).
Figure 5The GoogLeNet inception modules. Left: Naïve version of inception module. Right: Inception module with dimensionality reduction.
Figure 6The AlexNet architecture.
Different algorithms used for neuroimaging problems and their usage frequency.
|
|
|
|
|
|---|---|---|---|
| Classification algorithms | 66 (91%) | VGG | 11 (15%) |
| AlexNet | 8 (11%) | ||
| ResNet | 9 (13%) | ||
| Inception/GoogLeNet | 9 (13%) | ||
| SVM | 7 (10%) | ||
| Custom CNN | 14 (19%) | ||
| SqueezeNet | 1 (1%) | ||
| ConnectomeCNN | 1 (1%) | ||
| DenseNet | 1 (1%) | ||
| Logistic Regression | 1 (1%) | ||
| TrAdaboost | 1 (1%) | ||
| Lasso | 1 (1%) | ||
| LSTM | 2 (3%) | ||
| Segmentation | 6 (9%) | U-Net | 4 (6%) |
| Custom CNN | 2 (3%) | ||
|
|
|
Machine learning vs. deep learning algorithms used for neuroimaging problems and their usage frequency.
|
|
|
|
|---|---|---|
| Deep learning 62 (86%) | VGG | 11 (15%) |
| AlexNet | 8 (11%) | |
| ResNet | 9 (13%) | |
| Inception/GoogLeNet | 9 (13%) | |
| U-Net | 4 (6%) | |
| Custom CNN | 16 (22%) | |
| SqueezeNet | 1 (1%) | |
| ConnectomeCNN | 1 (1%) | |
| DenseNet | 1 (1%) | |
| LSTM | 2 (3%) | |
| Machine leaning 10 (14%) | Lasso | 1 (1%) |
| SVM | 7 (10%) | |
| Logistic Regression | 1 (1%) | |
| TrAdaboost | 1 (1%) | |
|
|
|
Different transfer learning approaches.
|
|
|
|
|---|---|---|
| Kernel learning (KL) | 8 | 13% |
| Freeze convolution and FC layers (FF) | 12 | 19% |
| Freeze convolution layers and fine-tune FC layers (FT) | 9 | 14% |
| Freeze convolution layers and randomly initialize FC layers (FI) | 8 | 13% |
| Fine-tune convolution and FC layers (TT) | 23 | 37% |
| Fine-tune convolution and initialize FC layers randomly (TI) | 3 | 5% |
| All | 63 | 100% |
Different research problem discussed by the literature.
|
|
|
|
|---|---|---|
| AD | 29 | 58% |
| Brain mapping | 8 | 16% |
| Age classification | 2 | 4% |
| Brain tumor | 3 | 6% |
| MS | 1 | 2% |
| Obsessive-compulsive disorder | 1 | 2% |
| Autism | 1 | 2% |
| Arterial spin labeling | 1 | 2% |
| Brain diseases | 2 | 4% |
| Alcoholics detection | 1 | 2% |
| Parkinson's disease | 1 | 2% |
|
|
|
|
Strength and limitations of transfer learning approaches.
|
|
|
|
|---|---|---|
| KL | Needs little computational resources and time. | Cannot be implemented on CNN algorithms. |
| FF | Does not need training or fine-tuning. Also, does not need large computational resources. | Cannot generalize well on very different source and target datasets. Fewer applications. |
| FT | Convolution layers act as feature extractor and do not need to be trained again. Diverse applications. Faster than most methods other than FF. Second in performance after TT. | Needs more hardware resources than FF because of fine-tuning of FC layers. May not be as successful as TT if source and target datasets are very different. |
| FI | Convolution layers act as feature extractor and do not need to be trained again. Diverse applications. Faster than most methods other than FF. | Needs more resources for FC layers to be trained from scratch. Slow. |
| TT | Best performance among all methods. Very flexible. | Needs more resources for fine-tuning convolution and FC layers. Slow. |
| TI | Strong performance, almost as good as FT. Much more flexible than other methods, including TT. | Slowest. Needs much more resources than other methods. |