| Literature DB >> 35565352 |
Se-Woon Choe1,2, Ha-Yeong Yoon3, Jae-Yeop Jeong3, Jinhyung Park1, Jin-Woo Jeong3.
Abstract
Microscopic image-based analysis has been intensively performed for pathological studies and diagnosis of diseases. However, mis-authentication of cell lines due to misjudgments by pathologists has been recognized as a serious problem. To address this problem, we propose a deep-learning-based approach for the automatic taxonomy of cancer cell types. A total of 889 bright-field microscopic images of four cancer cell lines were acquired using a benchtop microscope. Individual cells were further segmented and augmented to increase the image dataset. Afterward, deep transfer learning was adopted to accelerate the classification of cancer types. Experiments revealed that the deep-learning-based methods outperformed traditional machine-learning-based methods. Moreover, the Wilcoxon signed-rank test showed that deep ensemble approaches outperformed individual deep-learning-based models (p < 0.001) and were in effect to achieve the classification accuracy up to 97.735%. Additional investigation with the Wilcoxon signed-rank test was conducted to consider various network design choices, such as the type of optimizer, type of learning rate scheduler, degree of fine-tuning, and use of data augmentation. Finally, it was found that the using data augmentation and updating all the weights of a network during fine-tuning improve the overall performance of individual convolutional neural network models.Entities:
Keywords: cancer cell taxonomy; convolutional neural network; deep learning; ensemble approach
Year: 2022 PMID: 35565352 PMCID: PMC9100154 DOI: 10.3390/cancers14092224
Source DB: PubMed Journal: Cancers (Basel) ISSN: 2072-6694 Impact factor: 6.575
Figure 1Workflow of the proposed approach.
Figure 2Image preprocessing step. (a) Captured microscope image, (b) grayscale image, (c) noise removed image, (d) identified cell contour, (e) segmented image patches.
Figure 3Example of data augmentation: (A) original, (B) rotation, (C) translation, (D) vertical flip, (E) all.
Figure 4Degree of fine-tuning.
Figure 5Overview of ensemble approaches.
Summary of hyper-parameters.
| Parameter | Option | Note |
|---|---|---|
| Data augmentation | O | Rotation, translation, and vertical flip |
| X | Without any augmentation | |
| Fine-tuning | Without freeze | All weights are updated |
| 25% freeze | Only 75% of weights are updated | |
| Optimizer | SGD | Stochastic gradient descent |
| AdaGrad | Adaptive gradient-based optimization | |
| Learning rate scheduler | O | Exponential decay |
| X | Learning rate is fixed to 0.001 |
Comparison of the model performance (Best scores in each algorithm are marked in bold).
| Algorithm | Model | Accuracy | Precision | Recall | F1-Score |
|---|---|---|---|---|---|
| Machine Learning |
| ||||
| RF | 49.55 ± 0.32 | 49.01 ± 0.33 | 49.55 ± 0.32 | 49.3 ± 0.32 | |
| LDA | 46.26 ± 0.98 | 44.81 ± 0.98 | 45.26 ± 0.98 | 45.03 ± 0.98 | |
| KNN | 44.05 ± 0.92 | 45.86 ± 1.1 | 44.05 ± 0.92 | 44.93 ± 0.94 | |
| Average | 49.39 ± 5.94 | 49.51 ± 5.52 | 49.39 ± 5.94 | 49.44 ± 5.72 | |
| Deep Learning |
| ||||
| EfficientNetB2 | 96.195 ± 0.23 | 96.23 ± 0.272 | 96.176 ± 0.194 | 96.203 ± 0.232 | |
| ResNet50 | 96.265 ± 0.138 | 96.274 ± 0.13 | 96.265 ± 0.138 | 96.269 ± 0.134 | |
| InceptionV3 | 95.57 ± 0.322 | 95.604 ± 0.376 | 95.556 ± 0.298 | 95.58 ± 0.336 | |
| MobileNetV2 | 95.412 ± 0.223 | 95.446 ± 0.229 | 95.412 ± 0.224 | 95.429 ± 0.226 | |
| Average | 96.071 ± 0.584 | 96.1 ± 0.58 | 96.06 ± 0.581 | 96.08 ± 0.58 | |
| Ensemble (Single-architecture) |
| ||||
| EfficientNetB2 | 96.757 ± 0.202 | 96.763 ± 0.294 | 96.757 ± 0.294 | 96.76 ± 0.294 | |
| ResNet50 | 97.066 ± 0.148 | 97.073 ± 0.145 | 97.066 ± 0.148 | 96.07 ± 0.147 | |
| InceptionV3 | 96.342 ± 0.196 | 96.345 ± 0.202 | 96.342 ± 0.196 | 96.343 ± 0.199 | |
| MobileNetV2 | 96.533 ± 0.209 | 96.55 ± 0.226 | 96.533 ± 0.209 | 96.541 ± 0.217 | |
| Average | 96.868 ± 0.5 | 96.875 ± 0.5 | 96.868 ± 0.5 | 96.871 ± 0.5 | |
| Ensemble (Multi-architecture) | Top-1 | 97.563 ± 0.145 | 97.568 ± 0.145 | 97.563 ± 0.145 | 97.565 ± 0.145 |
| Top-2 | 97.673 ± 0.122 | 97.677 ± 0.124 | 97.673 ± 0.122 | 97.675 ± 0.123 | |
|
|
|
|
|
| |
| Average | 97.657 ± 0.144 | 97.661 ± 0.149 | 97.657 ± 0.144 | 97.659 ± 0.144 |
Configuration of the best-performing individual deep learning models.
| Algorithm | Data Augmentation | Degree of Fine-Tuning | Optimizer | Learning Rate Scheduler |
|---|---|---|---|---|
| DenseNet121 | O | All weights | SGD | X |
| EfficientNetB2 | O | All weights | AdaGrad | X |
| ResNet50 | O | All weights | AdaGrad | X |
| InceptionV3 | O | All weights | SGD | O |
| MobileNetV2 | O | Freeze the early 25% layers | SGD | O |
Figure 6Classification accuracy per epoch: (A) training accuracy, (B) validation accuracy.
Figure 7Loss per epoch: (A) training loss, (B) validation loss.
Number of trainable parameters.
| Degree of Fine-Tuning | All Weights | Freeze the First 25% Layers | |
|---|---|---|---|
| Model | |||
| DenseNet121 | 6,957,956 | 6,716,740 | |
| MobileNetV2 | 2,228,996 | 2,197,060 | |
| EfficientNetB2 | 7,706,630 | 7,700,858 | |
| InceptionV3 | 21,776,548 | 21,348,836 | |
| ResNet50 | 23,542,788 | 23,315,972 | |
Figure 8Network design choices (the statistical significance is represented using * (p < 0.05), ** (p < 0.01), and *** (p < 0.001)): (A) optimizer, (B) data augmentation, (C) learning rate scheduler, (D) degree of fine-tuning.
Figure 9Performance change according to the ensemble configuration: (A) single-arch ensemble, (B) multi-arch ensemble.
Comparison with previous studies (“CNN” denotes “Convolutional Neural Network”, “GAN” denotes “Generative Adversarial Network”, “ML” denotes “Machine Learning”, “ANN” denotes “Artificial Neural Network”, “GA” denotes “Generic Algorithm”).
| Ref. | Task | Image Acquisition | Method | Num. of Classes | Metric | Performance | Feature |
|---|---|---|---|---|---|---|---|
| Rubin et al. [ | Cancer cell classification | Low-coherence off-axis holography without statining | GAN-based approach | 4 classes (healthy skin, melanoma cells, colorectal adenocarcinoma colon cells, metastatic colorectal adenocarcinoma cells) | Accuracy | 90–99% | CNN feature |
| Oei et al. [ | Breast cancer cell detection | Confocal immunofluorescence microscopy images with staining | CNN | 2 classes (breast normal cells and cancer cells) | Accuracy | 97.2% | CNN feature |
| Kumar et al. [ | Cervical cancer cell detection | Microscopic biopsy images with staining | RF, SVM, KNN, fuzzy KNN | 2 classese (noncancerous, cancerous) | Accuracy | 92.19% | Texture features, morphology and shape features, HOG, wavelet features, etc. |
| Shi et al. [ | Cervical cancer cell classification | Microscopic images of Pap smear slides with staining | Graph neural network | 5 types of cervical cancer cells (superficial–intermediate, parabasal, koilocytotic, dyskeratotic, and metaplastic cells) | Accuracy | 94.93% | CNN feature |
| Sophea et al. [ | HOG + SVM | 2 classes (normal and abnormal) | Accuracy | 94.7% | HOG | ||
| Chankong et al. [ | Bayes, LDA, KNN, ANN, SVM | 7 classes (superficial squamous, intermediate squamous, columnar, mild dysplasia, moderate dysplasia, severe dysplasia, and carcinoma in situ) | Accuracy | 93.78% | Hand-crafted features (area of cucleus, nucleus-to-cytoplasm ratio, etc.) | ||
| Sharma et al. [ | KNN | Accuracy | 82.9% | ||||
| Gençtav et al. [ | Bayesian, decision tree, SVM | Precision | 91.7% | ||||
| Marinakis et al. [ | GA | Accuracy | 96.73% | ||||
| Our proposed method | Cancer cell classification | Microscopic images of cell culture flask without staining | CNN ensemble | 4 classes of cell culture flask (HeLa, MCF-7, Huh7, and NCI-H1299) | Accuracy | 97.735% | CNN feature |