| Literature DB >> 35741303 |
Youngjin Park1, Mujin Kim1, Murtaza Ashraf1, Young Sin Ko2, Mun Yong Yi1.
Abstract
CNN-based image processing has been actively applied to histopathological analysis to detect and classify cancerous tumors automatically. However, CNN-based classifiers generally predict a label with overconfidence, which becomes a serious problem in the medical domain. The objective of this study is to propose a new training method, called MixPatch, designed to improve a CNN-based classifier by specifically addressing the prediction uncertainty problem and examine its effectiveness in improving diagnosis performance in the context of histopathological image analysis. MixPatch generates and uses a new sub-training dataset, which consists of mixed-patches and their predefined ground-truth labels, for every single mini-batch. Mixed-patches are generated using a small size of clean patches confirmed by pathologists while their ground-truth labels are defined using a proportion-based soft labeling method. Our results obtained using a large histopathological image dataset shows that the proposed method performs better and alleviates overconfidence more effectively than any other method examined in the study. More specifically, our model showed 97.06% accuracy, an increase of 1.6% to 12.18%, while achieving 0.76% of expected calibration error, a decrease of 0.6% to 6.3%, over the other models. By specifically considering the mixed-region variation characteristics of histopathology images, MixPatch augments the extant mixed image methods for medical image analysis in which prediction uncertainty is a crucial issue. The proposed method provides a new way to systematically alleviate the overconfidence problem of CNN-based classifiers and improve their prediction accuracy, contributing toward more calibrated and reliable histopathology image analysis.Entities:
Keywords: confidence calibration; deep learning; histopathology image analysis; prediction uncertainty
Year: 2022 PMID: 35741303 PMCID: PMC9221905 DOI: 10.3390/diagnostics12061493
Source DB: PubMed Journal: Diagnostics (Basel) ISSN: 2075-4418
Figure 1Baseline vs. MixPatch. A single WSI generates multiple patches. The process of tiling creates certain case patches and uncertain case patches. Most parts of a certain patch are covered by a single label, but those of an uncertain patch are mixed. The baseline methods are overconfident, even for uncertain patches and incorrect outputs. The proposed method, MixPatch, overcomes these problems by explicitly incorporating the mixed-region variations in histopathological images into the training process.
Figure 2The overall process of the proposed method. In the existing methods, the patch-level classifier is trained using a CNN model and a cleaned patch dataset, , which pathologists previously confirmed. The proposed method, MixPatch, additionally uses a new subtraining dataset, which consists of image and label . is built by combining randomly selected images from the minipatch dataset. is defined according to the ratio of abnormal mini-patches. In the figure, a minibatch is a randomly built mix of samples from and samples from .
Compositions of datasets.
| Original Training Dataset | Minipatch Dataset | Test Dataset | ||||
|---|---|---|---|---|---|---|
| Class | Normal | Abnormal | Normal | Abnormal | Normal | Abnormal |
| WSIs | 204 | 282 | 204 | 282 | 48 | 50 |
| Patches | 32,063 | 38,492 | 3500 | 3500 | 3733 | 3780 |
Summary of the compared methods.
| Baseline | LS | Cutout | CutMix | MixPatch | |
|---|---|---|---|---|---|
| Data augmentation | X | X | O | O | O |
| Soft labeling | X | O | X | O | O |
| Ratio reflection | X | X | X | O | O |
| All correct labeling | O | O | X | X | O |
| Image |
|
|
|
|
|
| Label | Normal 1.0 | Normal 0.9 | Abnormal 1.0 | Normal 0.8 | Normal 0.4 |
| Actual label | Normal | Normal | Abnormal | Abnormal | Abnormal |
Labeling strategy for a mixed patch.
| Abnormal Patch Ratio | New Ground-Truth Label |
|---|---|
| 0/4 | [0.9, 0.1] |
| 1/4 | [0.4, 0.6] |
| 2/4 | [0.3, 0.7] |
| 3/4 | [0.2, 0.8] |
| 4/4 | [0.1, 0.9] |
The confusion matrix for outcome of predictions.
| Actual | |||
|---|---|---|---|
| Abnormal (Positive) | Normal (Negative) | ||
| Prediction | Abnormal (Positive) | True positive ( | False positive ( |
| Normal (Negative) | False negative ( | True negative ( | |
Performance comparison of the alternative methods.
| Training Methods | Accuracy ↑ | Sensitivity ↑ | Specificity ↑ | AUROC ↑ | ECE ↓ |
|---|---|---|---|---|---|
| Baseline |
|
|
|
|
|
| LS |
|
|
|
|
|
| Cutout |
|
|
|
|
|
| CutMix |
|
|
|
|
|
| MixPatch |
|
|
|
|
|
Figure 3ROC curve for the different methods.
Figure 4Integrated reliability diagram for patch-level classifiers trained using each method.
Confidence distributions of each method.
| Methods | Confidence Distributions | |
|---|---|---|
| False Predictions | True Predictions | |
| Baseline |
|
|
| Label smoothing |
|
|
| CutMix |
|
|
| Cutout |
|
|
| MixPatch |
|
|
Confusion matrix for each method with a threshold approach (X = prediction, Y = true).
|
| |||||
|---|---|---|---|---|---|
| Model | 0.5 (Baseline) | 0.4 | 0.3 | 0.2 | 0.1 |
| Baseline |
|
|
|
|
|
| LS |
|
|
|
|
|
| CutMix |
|
|
|
|
|
| Cutout |
|
|
|
|
|
| MixPatch |
|
|
|
|
|
Figure 5The Grad-Cam [62] visualization examples for uncertain patch images.
Performance in WSI classification.
| WSI Classifiers | WSI-Level Accuracy ↑ (In Percent) |
|---|---|
| Baseline |
|
| LS |
|
| Cutout |
|
| CutMix |
|
| MixPatch |
|