| Literature DB >> 33062608 |
Alessandro Bruno1, Edoardo Ardizzone2, Salvatore Vitabile3, Massimo Midiri3.
Abstract
BACKGROUND: Deep learning methods have become popular for their high-performance rate in the classification and detection of events in computer vision tasks. Transfer learning paradigm is widely adopted to apply pretrained convolutional neural network (CNN) on medical domains overcoming the problem of the scarcity of public datasets. Some investigations to assess transfer learning knowledge inference abilities in the context of mammogram screening and possible combinations with unsupervised techniques are in progress.Entities:
Keywords: Classification; computer-assisted image processing; computing methodologies; deep learning; digital mammography
Year: 2020 PMID: 33062608 PMCID: PMC7528986 DOI: 10.4103/jmss.JMSS_31_19
Source DB: PubMed Journal: J Med Signals Sens ISSN: 2228-7477
A list of some state-of-the-art methods
| Author | Method | Dataset | Reported results | Pros and cons |
|---|---|---|---|---|
| Cao | Spatial clustering[ | Mini-MIAS | Sensitivity 88.7% | The method is based on a RIC algorithm. |
| Ribli | Faster R-CNN[ | INbreast | AUC 0.85, sensitivity 90%, and 0.14 false-positive marks per image | It sets the state-of-the-art classification performance on INbreast. The size of the publicly available dataset is small |
| Hu | Adaptive thresholding[ | Mini-MIAS | Sensitivity 91.8% | The global and local thresholds are chosen adaptively without artificial intelligence. Tests over mammograms with different spatial resolutions are missing |
| Huynh | Transfer learning from deep convolutional neural networks[ | Dataset from the University of Chicago Medical Center. 219 digital mammograms and 607 ROIs | AUC 0.86 and true positive and false-positive fractions | The article shows performances of several architectures |
| Xi | Deep convolutional neural networks[ | DDSM dataset | Accuracy 95% | The authors investigated the powerfulness of some state-of-the-art CNNs |
| Pereira | Multilevel thresholding[ | Mini-MIAS | Sensitivity 90% | The method runs both detection and segmentation tasks. The detection accuracy rate is high, while segmentation performs a bit lower |
| Dunghel | Combination of cascade of deep learning and random forest[ | DDSM and INbreast | True positive rate of 0.96±0.03 at 1.2 false positive per image on INbreast. True positive rate of 0.75 at 4.8 false positive per image on DDSM | The method achieves very good performances on both datasets. The computational burden of the method seems to be quite expensive (the execution time is almost 20 s) |
| Tavakoli | CNNs and a decision scheme[ | Mini-MIAS | Sensitivity 93.33% | ROIs in the proposed method are not rescaled to preserve the quality of the image |
| Burçin | Havrda and Charvat entropy and Otsu N thresholding[ | Mini-MIAS | Sensitivity 90.2% | The method detects abnormalities from mammograms using an unsupervised approach. A check of the robustness of the features extracted over another dataset is missing |
| Akila Agnes | Multiscale all convolutional neural network[ | Mini-MIAS | Sensitivity 96% and 0.99 AUC | The method exhaustively exploits the powerfulness of CNN over Mini-MIAS reaching out impressive performances |
| Sampaio | Cellular neural network[ | Mini-MIAS | Sensitivity 90.9% | The method allows for detecting and segmenting suspicious regions even though the latter task has some drawbacks (10% of masses were lost) |
| Levy and Jain | Deep convolutional neural networks[ | DDSM | Accuracy 92.9%, precision 92.4%, recall 93.4% | Preprocessing, data augmentation, and transfer learning steps are run to obtain state-of-the-art performances |
| Vikhe and Thool | Wavelet processing and adaptive thresholding[ | Mini-MIAS and DDSM | Sensitivity 91% | The method runs suspicious region detection on two subsets of the existing databases reaching out more or less the same accuracy levels |
| Anitha | WPAT, Dual-Stage adaptive thresholding[ | Mini-MIAS | Sensitivty 93% | The method relies upon dual-stage adaptive thresholding which is, at the same time, dependent on pectoral muscle removal step |
| Teare | Genetic search of image enhancement methods and a dual deep convolutional neural networks[ | DDSM and ZMDS | Specificity 91% Specificity 80% | False-color enhancement technique to mammography images and utilizing a dual deep CNN engine. Some details on the reliability of the whole system are missing |
| Jaffar | DuSAT, deep convolutional neural network with support vector machine[ | Mini-MIAS, DDSM | Sensitivity 93.25% | Performances over two different datasets are very similar. Comparisons over high-resolution images are missing |
CNNs–Convolutional neural networks, WPAT–Wavelet processing and adaptive thresholding, DuSAT–Dual-stage adaptive thresholding, DDSM–Digital Database for Screening Mammography, MIAS–Medical image analysis, RIC–Robust information clustering, AUC–Area under the curve, ROIs–Regions of Interest, ZMDS–Zebra Mammography Dataset
Figure 1A patch sample from Suspicious Region Detection on Mammogram from PP (a) and a sample of patches generated with data augmentation (b)
Figure 2The histogram of a sample from the Suspicious Region Detection on Mammogram from PP dataset is given (two main modalities can be observed)
Figure 3The logistic function (a) and the nonparametric kernel-smoothing distribution (b) are used as fitting functions of the histogram for the Breast profile regions. These functions are, then, used to generate two new versions of the given mammogram
Figure 4The overall working scheme of the Scale Invariant Feature Transform based module is represented with respect to all the steps which it is made of: (a) the input image is specified into two new mammograms (b) using the logistic function and the nonparametric kernel-smoothing distribution [Figure 2]; Scale invariant feature transform keypoints are extracted on both the mammogram versions considering the radius parameter and discarding those keypoints with negative Laplacian (c), then the intersection between all the keypoints extracted on both mammogram versions is performed as a sort of result integration (d)
Figure 5The overall scheme of the deep learning technique we adopt for our purpose: employment of pretrained convolutional neural networks, transfer learning, data augmentation, regularization, and fine-tuning on biomedical data
Figure 6The training accuracy rates of AlexNET on Mini-MIAS are shown with respect to different number of epochs and Mini-Batch
Figure 9The training accuracy rates of PyramidNet on Suspicious Region Detection on Mammogram from PP are shown with respect to different number of epochs and histogram specifications
Figure 7The training accuracy rates of PyramidNet on Mini-MIAS are shown with respect to different number of epochs and Mini-Batch
Figure 8The training accuracy rates of AlexNET on Suspicious Region Detection on Mammogram from PP are shown with respect to different number of epochs and histogram specifications
5-fold cross validation performance of PyramidNet over 4916 patches from SuReMaPP
| Fold test | Suspicious (%) | Nonsuspicious (%) | ||
|---|---|---|---|---|
| True | False | True | False | |
| 1st fold | 461 (93.8) | 30 (6.2) | 458 (93) | 34 (7) |
| 2nd fold | 463 (94.2) | 28 (5.8) | 464 (94.3) | 28 (5.7) |
| 3rd fold | 465 (94.7) | 26 (5.3) | 467 (94.9) | 25 (5.1) |
| 4th fold | 466 (94.9) | 25 (5.1) | 469 (95.3) | 23 (4.7) |
| 5th fold | 466 (94.9) | 25 (5.1) | 471 (95.7) | 21 (4.3) |
| Average (%) | 94.5±0.48 | 5.5±0.48 | 94.64±1.05 | 5.36±1.05 |
We have a total number of 4916 patches from SuReMaPP. To apply 5-fold cross-validation, we split up the whole dataset into five subsets counting 983 patches. The remaining amount of patches are used as the training set over the category suspicious and non-suspicious regions. The process aims to detect the capabilities of the model to infer knowledge over the classification task. We repeat the steps on each of the five subsets. The results for each of the 5 groups are described in each table row using the true positives and false positives. The average percentage in the bottom row gives us a measure of the knowledge inference of the Deep Learning Model
Figure 10The overall scheme of our novel technique which consists of the integration of a scale invariant feature transform-based method and a deep learning module with transfer learning: input (mammogram image); histogram specifications (logistic function and nonparametric kernel-smoothing distribution); scale invariant feature transform-Based method which extract keypoints on candidate suspicious regions; PyramidNet fine-tuned over mammogram images (specified with the same histogram specification as in scale invariant feature transform-based method); Mechanism of voting using Softmax function
Figure 11Blue grids above are with the same size (224 × 224) of the input layer of PyramidNet. The red circle spots the suspicious region detected by the radiologists (a). A patch (yellow square) centered on a keypoint (green dot) that does not intercept any suspicious region (b) is a false positive (FP). A patch (yellow square) centered on a keypoint that intercepts the suspicious region (c) is a true positive (TP). In the last case (d), we count the blue dotted patch containing the suspicious region as a false negative (the system was able to detect only a false positive, see the yellow square)
Figure 12A comparison between different fine-tuning data combination is given with respect to sensitivity and specificity metrics
Figure 13The proposed method is compared with spatial clustering[24] (SC), adaptive thresholding[25] (AT), multilevel thresholding[26] (MT), Havrda and Charvat entropy and Otsu N thresholding[73] (HC), cellular neural network[27] (CelNN), wavelet processing and adaptive thresholding[28] (WPAT), dual-stage adaptive thresholding[29] (DuSAT), deep convolutional neural network with support vector machine[40] (DCNN_SVM), convolutional neural networks and a decision scheme[45] (convolutional neural networks + DS), multiscale all convolutional neural Network[50] (M All convolutional neural network) and our proposed method
Figure 14The most important steps of our solution are resumed with green keypoints (b) (scale invariant feature transform-based module) and red keypoints (c) (validated by transfer learning module) overlaid on a mammogram. Only the smaller red circles in the integrated results (c) turn out to be true positives