| Literature DB >> 31350843 |
Syed Jamal Safdar Gardezi1, Ahmed Elazab1, Baiying Lei1, Tianfu Wang1.
Abstract
BACKGROUND: Machine learning (ML) has become a vital part of medical imaging research. ML methods have evolved over the years from manual seeded inputs to automatic initializations. The advancements in the field of ML have led to more intelligent and self-reliant computer-aided diagnosis (CAD) systems, as the learning ability of ML methods has been constantly improving. More and more automated methods are emerging with deep feature learning and representations. Recent advancements of ML with deeper and extensive representation approaches, commonly known as deep learning (DL) approaches, have made a very significant impact on improving the diagnostics capabilities of the CAD systems.Entities:
Keywords: breast cancer; convolutional neural networks; deep learning; lesion classification; machine learning; malignant tumor
Year: 2019 PMID: 31350843 PMCID: PMC6688437 DOI: 10.2196/14464
Source DB: PubMed Journal: J Med Internet Res ISSN: 1438-8871 Impact factor: 5.428
Figure 1Multiview breast mammogram of a patient. The first column presents two views of the right breast: right craniocaudal (RCC) view and right mediolateral oblique (RMLO) view. The second column presents two views of the left breast: left craniocaudal (LCC) view and left mediolateral oblique (LMLO) view.
Figure 2Difference between 2 pipelines: conventional machine learning pipeline (left) and deep learning pipeline (right).
Figure 3(a) Original mammogram image 1024×1024. (b) Preprocessing to remove annotations. (c) pectoral muscle (PM) removal by region growing. (d) PM removal by adaptive segmentation.
Figure 4Pixel-level illustration of true positive, false positive, and false negative compared with ground truth.
Summary of advantages and disadvantages of segmentation methods.
| Methods | Advantages | Disadvantages |
| GTa | Widely used as preprocessing step in image processing as these methods are easy to implement | Not suitable for segmentation of ROIsb, as GT methods produce high false positive detections |
| Local thresholding | Works well compared with GT, sometimes used to improve the GT results | Widely used in literature as initialization step of other algorithms, but local thresholding fails to separate the pixels accurately into suitable regions |
| Region growing | Uses pixel connectivity properties to grow iteratively and sum up the region having similar pixel properties | Need initialization point, that is, a seed point to begin with and highly dependent on initial guess |
| Region clustering | No seed point required to initialize; it can directly search the cluster regions. | Total number of clusters need to be predefined at initial stage |
| Edge detection | Highly suitable for detecting the object boundaries and contours of the suspected ROIs | Requires some information about object properties |
| Template matching | Needs ground truth and are easily implemented. Easy implementation; if the prototypes are suitably selected, it can produce good results. | Need prior information about the region properties of the objects such as size, shape, and area. |
| Multiscale technique | Do not require any prior knowledge about object properties | Requires empirical evaluation to select the appropriate wavelet transform |
| Easily discriminate among the coefficients at different level and scale of decompositions | Need to select scale of decompositions |
aGT: Global thresholding.
bROI: region of interest.
Figure 5An overview of mammogram processing using computer-aided diagnosis based on machine learning algorithms.
Summary of convolutional neural network–based methods for breast density estimation.
| Author | Method | Dataset/number | Task | Performance metric/s (value/s) | Code availability |
| Mohamed et al [ | CNNa (AlexNet; transfer learning) | Private, University of Pittsburgh/200,00 DMb (multiview) | Breast density estimation | AUCc (0.9882) | —d |
| Ahn et al [ | CNN (transfer learning) | Private, Seoul University Hospital/397 DM (multiview) | Breast density estimation | Correlation coefficient (0.96) | — |
| Xu et al [ | CNN (scratch based) | Public, INbreast dataset/410 DM (multiview) | Breast density estimation | Accuracy (92.63%) | — |
| Wu et al [ | CNN (transfer learning) | Private, New York University School of Medicine/201,179 cases (multiview) | Breast density estimation | Mean AUC (0.934) | [ |
| Kallenberg et al [ | Conventional sparse autoencoder, ie, CNN+stacked autoencoder | Private, Dutch Breast Cancer Screening Program and Mayo Mammography, Minnesota/493+668 images (multiview) | Breast density estimation and risk scoring | Mammographic texture (0.91) and AUC (0.61) | — |
| Ionescu et al [ | CNN | Private dataset/67,520 DM (multiview) | Breast density estimation and risk scoring | Average match concordance index of (0.6) | — |
| Geras et al [ | Multiview deep neural network | Private, New York University/886,000 image (multiview) | Breast density estimation and risk score | Mean AUC (0.735) | — |
aCNN: convolutional neural network.
bDM: digital mammogram.
cAUC: area under the curve.
dNot available.
Figure 6Sample results from the study by Ribli et al for mass detection and classification.
Figure 7An overview of conditional generative adversarial network adapted from the study by Singh et al for mass segmentation and shape classification. CNN: convolutional neural network.
Summary of convolutional neural network–based methods for breast mass detection.
| Author | Method | Dataset/number | Task | Performance metric/s (value/s) | Code availability |
| Dhungel et al [ | Hybrid CNNa+level set | Public, INbreast dataset/410 images (multiview) | Mass detection, classification of benign, and malignant | Accuracy (0.9) and sensitivity (0.98) | —b |
| Dhungel et al [ | CRFc+CNN | Public, INbreast and DDSMd/116 and 158 images (multiview) | Lesion detection and segmentation | Dice score (0.89) | — |
| Zhu et al [ | Fully convolutional network+ CRF | Public, INbreast and DDSM/116 and 158 images (multiview) | Lesion segmentation | Dice score (0.97) | [ |
| Wang et al [ | Stacked autoencoder (transfer learning) | Private, Sun Yat-Sen University/1000 Digital mammogram | Detection and classification of calcifications and masses | Accuracy (0.87) | — |
| Riddli et al [ | Faster R-CNN (transfer learning) | Public, DDSM (2620), INbreast (115), and private dataset by Semmelweis University Budapest/847 images | Detection and classification | AUCe (0.95) | Semmelweis dataset: [ |
| Singh et al [ | Conditional generative adversarial network and CNN | Public and private, DDSM and Reus Hospital Spain dataset/567+194 images | Lesion segmentation and shape classification | Dice score (0.94) and Jaccard Index (0.89) | — |
| Agarwal and Carson [ | CNN (scratch based) | Public, DDSM/8750 images (multiview) | Classification of mass and calcifications | Accuracy (0.90) | — |
| Gao et al [ | Shallow-deep convolutional neural network, ie, 4 layers CNN+ResNet | Private, Mayo Clinic Arizona (49 subjects) and public, INbreast dataset (89 subjects) (multiview) | Lesion detection and classification | Accuracy (0.9) and AUC (0.92) | — |
| Hagos et al [ | Multi-input CNN | Private (General Electric, Hologic, Siemens) dataset/28,294 images/(multiview) | Lesion detection and classification | AUC (0.93) and CPM (0.733) | — |
| Tuwen et al [ | Fast R-CNN and Mask R-CNN with ResNet variants as backbone | Private (General Electric, Hologic, Siemens) dataset/23,405 images (multiview) | Lesion detection and classification | Sensitivity (0.97) with 3.56 FPf per image | — |
| Jung et al [ | RetinaNet model | Public and private, INbreast and GURO dataset by Korea University Guro Hospital/410+222 images (multiview) | Mass detection and classification | Accuracy (0.98) with 1.3 FP per image | [ |
| Shen et al [ | CNN end-to-end (transfer learning through visual geometry group 16 and ResNet) | Public, DDSM and INbreast/2584 +410 (multiview) | Classification of masses | AUC (0.96) | [ |
aCNN: convolutional neural network.
bNot available.
cCRF: conditional random field.
dDDSM: Digital Database for Screening Mammography.
eAUC: area under the curve.
fFP: false positive.
Figure 8Sample results from the study by Wu et al for synthetic generation of data using conditional generative adversarial network. GAN: generative adversarial network.
Summary of convolutional neural network–based methods for breast mass classification.
| Author | Method | Dataset/number | Task | Performance metric/s (value/s) | Code availability |
| Levy and Jain [ | AlexNet and GoogleNet (transfer learning) | Public, DDSMa dataset/1820 images (multiview) | Breast mass classification | Accuracy (0.924), precision (0.924), and recall (0.934) | —b |
| Samala et al [ | Multistage fine-tuned CNNc (transfer learning) | Private+public, University of Michigan and DDSM/4039 ROIsd (multiview) | Classification performance on varying sample sizes | AUCe (0.91) | [ |
| Jadoon et al [ | CNN- Discrete wavelet and CNN-curvelet transform | Public, image retrieval in medical applications dataset/2796 ROI patches | Classification | Accuracy (81.83 and 83.74) and receiver operating characteristic curve (0.831 and 0.836) for both methods | — |
| Huynh et al [ | CNN (transfer learning) | Private, University of Chicago/219 images (multiview) | Classification of benign and malignant tumor | AUC (0.86) | — |
| Domingues and Cardoso [ | Autoencoder | Public, INbreast/116 ROIs | Classification of mass vs normal | Accuracy (0.99) | [ |
| Wu et al [ | GANf and ResNet50 | Public, DDSM dataset/10,480 images (multiview) | Detection and classification of benign and malignant calcifications and masses | AUC (0.896) | [ |
| Sarah et al [ | CNN (transfer learning) | Public, Full-field digital mammography and DDSM/14,860 images (multiview) | Classification | AUC (0.91) | — |
| Wang et al [ | CNN and long short-term memory | Public, Breast Cancer Digital Repository (BCDR-F03)/763 images (multiview) | Classification of breast masses using contextual information | AUC (0.89) | — |
| Shams et al [ | CNN and GAN | Public, INbreast and DDSM (multiview) | Classification | AUC (0.925) | — |
| Gastounioti et al [ | Texture feature+CNN | Private/106 cases (mediolateral oblique view only) | Classification | AUC (0.9) | — |
| Dhungel et al [ | Multi-ResNet | Public, INbreast (multiview) | Classification | AUC (0.8) | — |
aDDSM: Digital Database for Screening Mammography.
bNot available.
cCNN: convolutional neural network.
dROIs: region of interest.
eAUC: area under the curve.
fGAN: generative adversarial network.