| Literature DB >> 35190567 |
Reece Walsh1, Mohamed H Abdelpakey2, Mohamed S Shehata2, Mostafa M Mohamed3.
Abstract
Classifying and analyzing human cells is a lengthy procedure, often involving a trained professional. In an attempt to expedite this process, an active area of research involves automating cell classification through use of deep learning-based techniques. In practice, a large amount of data is required to accurately train these deep learning models. However, due to the sparse human cell datasets currently available, the performance of these models is typically low. This study investigates the feasibility of using few-shot learning-based techniques to mitigate the data requirements for accurate training. The study is comprised of three parts: First, current state-of-the-art few-shot learning techniques are evaluated on human cell classification. The selected techniques are trained on a non-medical dataset and then tested on two out-of-domain, human cell datasets. The results indicate that, overall, the test accuracy of state-of-the-art techniques decreased by at least 30% when transitioning from a non-medical dataset to a medical dataset. Reptile and EPNet were the top performing techniques tested on the BCCD dataset and HEp-2 dataset respectively. Second, this study evaluates the potential benefits, if any, to varying the backbone architecture and training schemes in current state-of-the-art few-shot learning techniques when used in human cell classification. To this end, the best technique identified in the first part of this study, EPNet, is used for experimentation. In particular, the study used 6 different network backbones, 5 data augmentation methodologies, and 2 model training schemes. Even with these additions, the overall test accuracy of EPNet decreased from 88.66% on non-medical datasets to 44.13% at best on the medical datasets. Third, this study presents future directions for using few-shot learning in human cell classification. In general, few-shot learning in its current state performs poorly on human cell classification. The study proves that attempts to modify existing network architectures are not effective and concludes that future research effort should be focused on improving robustness towards out-of-domain testing using optimization-based or self-supervised few-shot learning techniques.Entities:
Mesh:
Year: 2022 PMID: 35190567 PMCID: PMC8861170 DOI: 10.1038/s41598-022-06718-2
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1The process proposed for training and testing the nine selected few-shot learning techniques on out-of-domain data.
Figure 2A temporal overview of notable few-shot learning techniques proposed within the past 5 years. "Optimization-based" few-shot learning techniques refer to those proposing changes to optimization processes employed by a network. "Metric-based" few-shot learning techniques refer to those proposing a metric from which a similarity score between a set of samples can be obtained from. "Augmented Metric-based" few-shot learning techniques refer to those proposing an augmentation (such as application of a self-supervised or transductive process) to a metric-based few-shot learning technique.
Figure 3The optimization-based process for optimizing towards three tasks illustrated.
Figure 4An illustrated example of transductive few-shot learning. (A) Grey circles represent unlabelled points (the query set) and coloured circles represent labelled points (the support set). (B) All unlabelled points are labelled based on their position within the labelled data.
An overview of the differing details between the models trained and tested.
| Model name | Technique | Backbone | Preprocessing | Extra training data |
|---|---|---|---|---|
| AmdimNet[ | Self-supervised Metric | AmdimNet | No | Yes |
| EPNet[ | Transductive Metric | WRN28-10 | No | Yes |
| SimpleCNAPS[ | Metric | ResNet18 | No | Yes |
| PT+MAP[ | Metric | WRN28-10 | Yes | No |
| LaplacianShot[ | Metric | WRN28-10 | No | No |
| S2M2R[ | Self-supervised Metric | WRN28-10 | Yes | No |
| Reptile[ | Optimization | CONV4 | No | No |
| MAML[ | Optimization | CONV4 | No | No |
| ProtoNet[ | Metric | CONV4 | No | No |
Parameter details specific to each technique.
| Model | Optimizer | Momentum | Weight decay | Batch size |
|---|---|---|---|---|
| AmdimNet[ | Adam | – | – | 100 |
| EPNet[ | SGD | 0.9 | 0.0005 | 128 |
| SimpleCNAPS[ | Adam | – | – | 256 |
| PT+MAP[ | Adam | – | – | 16 |
| LaplacianShot[ | SGD | 0.9 | 0.0001 | 128 |
| S2M2R[ | Adam | – | – | 16 |
| Reptile[ | Adam | – | – | 5 |
| MAML[ | Adam | – | – | 32 |
| ProtoNet[ | Adam | – | – | 5 |
Test accuracy results from baseline experiments run against the mini-ImageNet test set, BCCD, and HEp-2.
| Model | Mini-ImageNet | BCCD | HEp-2 |
|---|---|---|---|
| AmdimNet[ | 89.75 ± 0.12 | 48.35 ± 0.18 | 54.32 ± 0.21 |
| EPNet[ | 88.66 ± 0.24 | 47.39 ± 0.22 | |
| SimpleCNAPS[ | 47.06 ± 0.72 | 53.15 ± 0.84 | |
| PT+MAP[ | 88.02 ± 0.13 | 42.94 ± 0.17 | 54.73 ± 0.22 |
| LaplacianShot[ | 82.27 ± 0.15 | 34.75 ± 0.13 | 44.69 ± 0.17 |
| S2M2R[ | 82.81 ± 0.31 | 44.15 ± 0.23 | 54.41 ± 0.27 |
| Reptile[ | 65.62 ± 0.28 | 51.76 ± 0.13 | |
| MAML[ | 64.62 ± 0.19 | 42.81 ± 0.21 | 45.21 ± 0.24 |
| ProtoNet[ | 67.88 ± 0.12 | 46.89 ± 0.13 | 50.70 ± 0.17 |
Testing using the BCCD dataset was performed using additional global pooling layers.
The highest accuracy relative to each dataset is in bold.
Test accuracy results from in-domain training on HEp-2 and testing on BCCD.
| Model | BCCD |
|---|---|
| EPNet[ | 45.31 ± 0.21 |
| Reptile[ | 40.24 ± 0.23 |
Test accuracy results from using different backbone variations in EPNet on and testing on mini-ImageNet, BCCD, and HEp-2.
| Backbone (%) | Mini-ImageNet (%) | BCCD | HEp-2 (%) |
|---|---|---|---|
| WideResNet28-10 (Original Backbone) | 88.7 | 47.4 | 55.1 |
| EfficientNetV2 (Default Width) | 59.8 | 18.3 | 26.3 |
| EfficientNetV2 (0.5 Width) | 67.3 | 25.7 | 33.3 |
| EfficientNetV2 (0.75 Width) | 69.2 | 28.0 | 35.7 |
| EfficientNetV2 (2.75 Width) | 70.8 | 29.5 | 37.1 |
| ResNet-18 | 68.2 | 26.8 | 34.7 |
| DenseNet | 78.8 | 37.4 | 45.0 |
Each backbone was trained on mini-ImageNet’s training set before testing.
Test accuracy results using different model additions within EPNet.
| Model addition | Mini-ImageNet | BCCD (%) | HEp-2 |
|---|---|---|---|
| No additions | 88.7% | 47.4 | 55.12% |
| RandAugment (Magnitude = 5–15) | 75.8% | 34.6 | 42.1% |
| RandAugment (Magnitude = 5–10) | 69.3% | 27.8 | 35.6% |
| RandAugment (Magnitude = 5) | 70.1% | 28.9 | 36.1% |
| Squeeze and excitation (Reduction = 0.10) | 68.7% | 27.6 | 35.2 |
| Squeeze and excitation (Reduction = 0.25) | 68.2% | 26.2 | 34.9% |
| Squeeze and excitation (Reduction = 2) | 63.9% | 22.4 | 30.5% |
| Mixup (Alpha = 0.10) | 76.2% | 34.7 | 42.8% |
| Label smoothing (A = 0.10) | 65.3 | 23.2 | 31.4% |
| Exponential moving average | 78.6% | 38.2 | 44.1% |
Each model addition was independently trained on the mini-ImageNet training set and tested on the mini-ImageNet test set, BCCD, and HEp-2.