| Literature DB >> 31167642 |
Dina Abdelhafiz1,2, Clifford Yang3, Reda Ammar4, Sheida Nabavi4.
Abstract
BACKGROUND: The limitations of traditional computer-aided detection (CAD) systems for mammography, the extreme importance of early detection of breast cancer and the high impact of the false diagnosis of patients drive researchers to investigate deep learning (DL) methods for mammograms (MGs). Recent breakthroughs in DL, in particular, convolutional neural networks (CNNs) have achieved remarkable advances in the medical fields. Specifically, CNNs are used in mammography for lesion localization and detection, risk assessment, image retrieval, and classification tasks. CNNs also help radiologists providing more accurate diagnosis by delivering precise quantitative analysis of suspicious lesions.Entities:
Keywords: Breast cancer; Classification; Computer-aided detection (CAD); Convolutional neural networks (CNNs); Deep learning (DL); Feature detection; Machine learning (ML); Mammograms (MGs); Transfer learning (TL)
Mesh:
Year: 2019 PMID: 31167642 PMCID: PMC6551243 DOI: 10.1186/s12859-019-2823-4
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
Inclusion/exclusion for the systematic review
| Category | Criteria |
|---|---|
| Time period | Published from 1995 to the present (Nov 2017). |
| Databases | Private and public databases. |
| Publication | English articles in print. |
| Excluded articles accepted for publication before appearance in journals or conferences as of Aug 2017. | |
| Research focus | All Implementation of CNNs for breast cancer in Mammography. |
| Keywords | Deep learning, convolutional neural networks, breast cancer, mammography and transfer learning. |
| Abnormalities | Mass, calcification, architectural distortion and asymmetries. |
Fig. 1A breakdown of the studies included in this survey in the year of publication grouped by their neural network task. Since 2016 the number of studies on CNN for MGs has increased significantly
Comparison between widely used databases in literature respect to size of images, views (CC, MLO), digital or film databases, the format of images, bits/pixel (bpp) and the distribution of normal, benign and malignant images
| Database | Image-size | Views | Type | Format | bpp | #Normal | #Benign | #Malignant |
|---|---|---|---|---|---|---|---|---|
| DDSM | 3118 ×5001 | Both | FSM | LJPEG | 12 | 914 | 870 | 695 |
| IRMA | Several | Both | Both | PNG | 12 | 1108 | 1284 | 1284 |
| INbreast | Several | Both | FFDM | DICOM | 16 | 67 | 220 | 49 |
| MIAS | 1024 ×1024 | MLO | FSM | PGM | 8 | 207 | 69 | 56 |
| BCDR-F01 | 720 ×1168 | Both | FSM | TIF | 8 | 0 | 187 | 175 |
| BCDR-F02 | 720 ×1168 | Both | FSM | TIF | 8 | 0 | 426 | 90 |
| BCDR-F03 | 720 ×1168 | Both | FSM | TIF | 8 | 0 | 426 | 310 |
| BCDR-D01 | Several | Both | FFDM | DICOM | 14 | 0 | 85 | 58 |
| BCDR-D02 | Several | Both | FFDM | DICOM | 14 | 0 | 405 | 51 |
| BCDR-DN01 | Several | Both | FFDM | DICOM | 14 | 200 | 0 | 0 |
A summary for the strengths and limitations of the DDSM, IRMA, INbreast, MIAS and BCDR databases
| Database | Strength | Limitation |
|---|---|---|
| DDSM | Big widely used database. | Non-standard format. |
| Shape variations of different lesions. | Not precise position of lesions. | |
| IRMA | Accurate position of lesions. | Non-standard format. |
| High resolution. | ||
| INbreast | Accurate position of lesions. | Limited size. |
| Limited mass shape variations. | ||
| Standard file format. | ||
| Old database. | ||
| No more supported. | ||
| MIAS | Still widely used. | Limited size. |
| Images are of low resolution. | ||
| Has MLO view only. | ||
| Different resolutions. | ||
| BCDR | Accurate position of lesions. | Limited size. |
| Standard file format. | ||
| Still in their development phase. |
Fig. 2The CNN architecture is a stack of Convolutional layer (Conv), Nonlinear layer (e.g. ReLU), Pooling layer (Pool), and a Loss function (e.g. SVM/Softmax) on the last (Fully connected) layer. The output can be a single class (e.g. Normal, Benign, Malignant)
The configurations of AlexNET, ZF-NET, GoogLeNET, VGG-NET and ResNET models
| AlexNet [ | ZF-Net [ | GoogLeNet [ | VGG-Net [ | ResNet [ | |
|---|---|---|---|---|---|
| Year | 2012 | 2013 | 2014 | 2014 | 2015 |
| Image Resolution | 227 ×227 | 227 ×227 | 224 ×224 | 224 ×224 | 2244 ×224 |
| Number of layers | 8 | 8 | 22 | 19 | 152 |
| Number of Conv-Pool layers | 5 | 5 | 21 | 16 | 151 |
| Number of FC layers | 3 | 3 | 1 | 3 | 1 |
| Full connected layer size | 4096,4096,1000 | 4096,4096,1000 | 1000 | 4096,4096,1000 | 1000 |
| Filter Sizes | 3, 5, 11 | 3, 5, 11 | 1,3,5,7 | 3 | 1,3,7 |
| Number of Filters | 96 - 384 | 96 - 384 | 64 - 384 | 64 - 512 | 64 - 2048 |
| Strides | 1, 4 | 1, 4 | 1, 2 | 1 | 1, 2 |
| Data Augmentation | + | + | + | + | + |
| Dropout | + | + | + | + | + |
| Batch Normalization | - | - | - | - | + |
| Number of GPU | 2 GTX | 1 GTX | A few high-end | 4 Nvidia | |
| 580 GPUs | 580 GPUs | GPUs | Titan Black GPUs | Titan Black GPUs | 8 GPUs |
| Training Time | 5:6 days | 12 days | 1 week | 2:3 weeks | 2:3 weeks |
| Top-5 error | 16.40% | 11.2% | 6.70% | 7.30% | 3.57% |
A comparison between most famous toolkits and libraries for training mammography
| Interface | Languages | Open source | CUDA support | Pre-trained models | Forks (Github) | Contributions (Github) | |
|---|---|---|---|---|---|---|---|
| TensorFlow | Python | C++, Python | Yes | Yes | Yes | 63,603 | 1,481 |
| Keras | Python, R | Python | Yes | Yes | Yes | 11,203 | 681 |
| Cafee | Python, Matlab, C++ | C++, Python | Yes | Yes | Yes | 14,868 | 267 |
| PyTorch | Python | C, Python, CUDA | Yes | Yes | Yes | 3,592 | 644 |
| MatConvNet | Matlab | CUDA | Yes | Yes | Yes | 651 | 24 |
Fig. 3Statistics for the included studies