| Literature DB >> 31019547 |
Lian Zou1,2,3, Shaode Yu1,4, Tiebao Meng5, Zhicheng Zhang1,2, Xiaokun Liang1,2,6, Yaoqin Xie1.
Abstract
This study reviews the technique of convolutional neural network (CNN) applied in a specific field of mammographic breast cancer diagnosis (MBCD). It aims to provide several clues on how to use CNN for related tasks. MBCD is a long-standing problem, and massive computer-aided diagnosis models have been proposed. The models of CNN-based MBCD can be broadly categorized into three groups. One is to design shallow or to modify existing models to decrease the time cost as well as the number of instances for training; another is to make the best use of a pretrained CNN by transfer learning and fine-tuning; the third is to take advantage of CNN models for feature extraction, and the differentiation of malignant lesions from benign ones is fulfilled by using machine learning classifiers. This study enrolls peer-reviewed journal publications and presents technical details and pros and cons of each model. Furthermore, the findings, challenges and limitations are summarized and some clues on the future work are also given. Conclusively, CNN-based MBCD is at its early stage, and there is still a long way ahead in achieving the ultimate goal of using deep learning tools to facilitate clinical practice. This review benefits scientific researchers, industrial engineers, and those who are devoted to intelligent cancer diagnosis.Entities:
Mesh:
Year: 2019 PMID: 31019547 PMCID: PMC6452645 DOI: 10.1155/2019/6509357
Source DB: PubMed Journal: Comput Math Methods Med ISSN: 1748-670X Impact factor: 2.238
Figure 1The diagram of the main flow chart of ML-based CAD (a) and major architectures of CNN-based CAD (b). The black dashed line indicates the blocks are modifiable. The green dashed line denotes each step in the ML-based model is interpretable, and the red solid line indicates the CNN-based model is data-driven when the architecture is fixed.
Figure 2The architecture of VGG16. It consists of 13 convolutional layers, 3 full-connection layers, 5 pooling layers, and 1 softmax layer in addition to the input and output layers.
Figure 3The diagram of knowledge transferred from the source domain to a different but related target domain. In the source domain, a model is trained with sufficient high-quality instances (data and labels) and transfer learning enables the model used in a related target domain. It relieves the requirement of huge amount of instances for the training of deep models in the target domain which is critically helpful in medical imaging field.
A summary of CNN based MBCD methods.
| Year | Database | No. of images | AUC | ACC | SPE | SEN | |
|---|---|---|---|---|---|---|---|
| [ | 2016 | DDSM | 600 | 0.967 | |||
| [ | 2016 | BCDR-F03 | 736 | 0.82 ± 0.03 | |||
| [ | 2016 | In-house | 607 | 0.86 | |||
|
| |||||||
| [ | 2017 | In-house | 3158 | 0.88 | 0.82 | 0.72 | 0.81 |
| [ | 2017 | In-house | 2454 | 0.82 ± 0.02 | |||
| [ | 2017 | In-house | 245 | 0.86 ± 0.01 | |||
| [ | 2017 | INbreast | 115 | 0.91 ± 0.12 | 0.95 ± 0.05 | ||
| [ | 2017 | In-house | 560 | 0.79 ± 0.02 | |||
| [ | 2017 | IRMA | 2796 | 0.839 | 0.837 | 0.854 | 0.797 |
|
| |||||||
| [ | 2018 | In-house | 78 | 0.81 ± 0.05 | |||
| [ | 2018 | In-house | 3290 | 0.7274 | |||
| [ | 2018 | DDSM | 600 | 0.974 | |||
| MIAS | 120 | 0.967 | |||||
| [ | 2018 | BCDR-F03 | 736 | 0.813 | |||
| [ | 2018 | DDSM | 5316 | 0.98 | 0.9735 | ||
| BCDR-F03 | 600 | 0.96 | 0.9667 | ||||
| INbreast | 200 | 0.97 | 0.9550 | ||||
| [ | 2018 | DDSM | 600 | 0.97 | |||
| [ | 2018 | DREAM | 82,000 | 0.85 | |||
| INbreast | 115 | 0.95 | |||||
| [ | 2018 | BCDR-F03 | 736 | 0.891 | 0.852 | ||
| [ | 2018 | BCDR-F03 | 736 | 0.88 | 0.81 | ||
Confusion matrix.
| Predicted positive | Predicted negative | |
|---|---|---|
| Histologically verified positive | True positive (TP) | False negative (FN) |
| Histologically verified negative | False positive (FP) | True negative (TN) |
Summary of CNN-based MBCD models from the model building to its pros and cons analysis.
| Publication (year) | Approach | Pros (+)/cons (−) |
|---|---|---|
| [ | (1) An 8-layered CNN | +parameter initialization |
| (2) SVM-based decision mechanism | +decision mechanism | |
| (3) Compared to ML- and CNN-based models | −256 mid- and 2048 high-level features | |
|
| ||
| [ | (1) A 3-layered CNN | +medical instances for training |
| (2) SVM-based classification | −17 low- and 400 high-level features | |
| (3) Compared to ML- and CNN-based models | −a shallow CNN model | |
|
| ||
| [ | (1) Transferred AlexNet | +parameter initialization |
| (2) SVM-based classification | +soft-voting-based decision mechanism | |
| (3) Classifier-based soft voting | −29 low- and 3795 high-level features | |
|
| ||
| [ | (1) Modified LeNet | +semisupervised learning |
| (2) Graph based semisupervised learning | +a few labeled data used for training | |
| (3) Feature dimension reduction | +less sensitive to initial labeled data | |
| (4) Using unlabeled data | ||
|
| ||
| [ | (1) Modified AlexNet | +parameter initialization |
| (2) Multitask transfer learning | +improved generalizability | |
|
| ||
| [ | (1) Transferred the VGG | +parameter initialization |
| (2) SVM-based classification | +decision mechanism | |
| (3) Compared to ML- and CNN-based models | −38 low- and 1472 high-level features | |
|
| ||
| [ | (1) R-CNN for detection and diagnosis | +minimal user intervention in image analysis |
| (2) Feature regression | −781 low-level features for CNN feature regression | |
| (3) RF-based classification | ||
|
| ||
| [ | (1) A 4-layered CNN | +medical instances for training |
| −a shallow CNN model | ||
|
| ||
| [ | (1) A 3-layered CNN | +medical instances for training |
| (2) SVM-based classification | +image analysis in transformed domain | |
| (3) Data augmentation | −a shallow CNN model | |
|
| ||
| [ | (1) VGG for feature extraction | +2 features selected for diagnosis |
| (2) Stepwise feature selection | ||
| (3) SVM-based classification | ||
|
| ||
| [ | (1) Transferred AlexNet | +parameter initialization |
| (2) Data augmentation | ||
| (3) Compared to CNN models | ||
|
| ||
| [ | (1) A 7-layered CNN | +parameter initialization |
| (2) Parasitic metric learning | +parasitic metric learning | |
|
| ||
| [ | (1) Transferred VGG | +parameter initialization |
| (2) Compared to CNN-based models | ||
|
| ||
| [ | (1) Transferred VGG/ResNet/Inception | +parameter initialization |
| (2) Comparison on 3 databases | +systematic comparison | |
| −time consuming | ||
|
| ||
| [ | (1) YOLO and tensor structure | +medical instances for training |
| (2) Data augmentation | +simultaneous detection and classification | |
|
| ||
| [ | (1) Faster R-CNN and VGG | +medical instances for training |
| (2) Pretrained with the DDSM | +both detection and diagnosis | |
| +evaluated on a large-scale screening dataset | ||
|
| ||
| [ | (1) GoogLeNet for feature extraction | +medical instances for training |
| (2) Attention mechanism for feature selection | +multiview and clinical information fusion | |
| (3) LSTM for feature fusion | ||
|
| ||
| [ | (1) Transferred AlexNet/GoogLeNet | +parameter initialization |
| (2) Data augmentation | ||
| (3) Compared to ML- and CNN-based models | ||
Figure 4The flow chart and an example of dedicated MBCD models. The flow chart highlights the CNN is a newly designed or modified network, and the example describes the architecture of a CNN model in [58]. It should be noted that parameters of dedicated models are with random initialization followed by iterative optimization with medical instances.
Figure 5The flow chart and an example of transferred MBCD models. The flow chart emphasizes transfer learning (dashed arrows) and fine-tuning, and the example comes from [64] which makes use of pretrained VGG16 for malignancy prediction. It should be noted that parameters of pretrained models are well-determined in the source domain, while fine-tuning attempts to use medical instances for further optimization of these parameters toward the target task.
Figure 6The flow chart and an example of CNN performing as feature extractors. The flow chart highlights the information fusion which can be further divided into two approaches, feature fusion followed by a classifier or decision fusion of lesion malignancy predicted by using one or more classifiers. The example comes from [51] which develops a new CNN model and the model is pretrained on ImageNet. At last, the model fuses the prediction results from SVM classifiers which separately use 384 midlevel features and 2014 high-level features as its input.
Technical highlights.
| New architecture | Transfer learning | Fine-tuning | Information fusion | |
|---|---|---|---|---|
| Dedicated CNN models | ✔ | ✖ | ✖ | ✖ |
| Transferred CNN models | — | ✔ | ✔ | ✖ |
| CNN models as feature extractors | — | — | — | ✔ |