| Literature DB >> 28607456 |
Monjoy Saha1, Chandan Chakraborty2, Indu Arun3, Rosina Ahmed3, Sanjoy Chatterjee3.
Abstract
Being a non-histone protein, Ki-67 is one of the essential biomarkers for the immunohistochemical assessment of proliferation rate in breast cancer screening and grading. The Ki-67 signature is always sensitive to radiotherapy and chemotherapy. Due to random morphological, color and intensity variations of cell nuclei (immunopositive and immunonegative), manual/subjective assessment of Ki-67 scoring is error-prone and time-consuming. Hence, several machine learning approaches have been reported; nevertheless, none of them had worked on deep learning based hotspots detection and proliferation scoring. In this article, we suggest an advanced deep learning model for computerized recognition of candidate hotspots and subsequent proliferation rate scoring by quantifying Ki-67 appearance in breast cancer immunohistochemical images. Unlike existing Ki-67 scoring techniques, our methodology uses Gamma mixture model (GMM) with Expectation-Maximization for seed point detection and patch selection and deep learning, comprises with decision layer, for hotspots detection and proliferation scoring. Experimental results provide 93% precision, 0.88% recall and 0.91% F-score value. The model performance has also been compared with the pathologists' manual annotations and recently published articles. In future, the proposed deep learning framework will be highly reliable and beneficial to the junior and senior pathologists for fast and efficient Ki-67 scoring.Entities:
Mesh:
Substances:
Year: 2017 PMID: 28607456 PMCID: PMC5468356 DOI: 10.1038/s41598-017-03405-5
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1Ki-67 proliferation scoring by the pathologists with respect to differential color distribution: Three input Ki-67 stained images of breast cancer at 40× with the scores (a) = 30%; (b) = 60%; (c) = 90%; and color spectrum visualization of the inputs images (d–f).
Characterization of Ki-67 scoring approaches.
| Categories | Year | Cancer type | Methodology used | Results |
|---|---|---|---|---|
|
| 2016 | Nasopharyngeal cancer | K-means clustering | 91.8% Segmentation accuracy[ |
| 2015 | Meningiomas and Oligodendrogliomas tumor | Morphology operation, thresholding, feature extraction and classification | The results shows the effectiveness of the proposed algorithm[ | |
| 2014 | Neuroendocrine tumor | Learning based approach | 89% precision, 91% recall, 90% F-score[ | |
| 2014 | Breast Cancer | Otsu thresholding | High correlation observed between manual and automated procedure[ | |
| 2014 | Breast cancer | Aperio Genie and Nuclear v9 software | Misclassification rate 5–7%[ | |
| 2013 | Pancreatic neuroendocrine tumor | Voting-Based Seed Detection, Repulsive Deformable Model, Two step classification | 87.68% classification accuracy, 88.01% sensitivity and 87.12% specificity[ | |
| 2013 | Rabbit Liver | Inform 1.4 image analysis software | Useful in clinical practice[ | |
| 2012 | Breast cancer | K-means clustering | T-test shows reliable proliferation rate[ | |
| 2012 | Not mentioned | Watershed segmentation, Laplacian-of-Gaussian filtering, SVM classifier | 90% sensitivity at confidence level I, 99% sensitivity at confidence level VIII[ | |
| 2012 | Breast Cancer | Slidepath Tissue IA system software | Excellent agreement between manual and automated technique[ | |
| 2010 | Breast cancer | ImmunoRatio software | 20% labeling index as a cutoff, 2.2 hazard ratio[ | |
| 2009 | Meningiomas tumor | Thresholding, watershed and morphological operations, SVM classifier | The proposed method helpful for further research[ | |
|
|
| |||
Figure 2Shows pictorial representation of decision layer connections in CAFFE.
Figure 3Illustrates the flow diagram of the patch detection from original images.
Proposed deep learning approach.
| Layer | Type | Maps | Neurons | Filter size |
|---|---|---|---|---|
| 0 | Input Image | 3 | 71 × 71 | — |
| 1 | Conv-1 | 90 | 70 × 70 | 2 × 2 |
| 2 | MP-1 | 90 | 35 × 35 | 2 × 2 |
| 3 | Conv-2 | 180 | 32 × 32 | 4 × 4 |
| 4 | MP-2 | 180 | 16 × 16 | 2 × 2 |
| 5 | Conv-3 | 360 | 14 × 14 | 3 × 3 |
| 6 | MP-3 | 360 | 7 × 7 | 2 × 2 |
| 7 | Conv-4 | 720 | 6 × 6 | 2 × 2 |
| 8 | MP-4 | 720 | 3 × 3 | 2 × 2 |
| 9 | Conv-5 | 1440 | 2 × 2 | 2 × 2 |
| 10 |
| — | 720 | 1 × 1 |
| 11 | FC-1 | — | 100 | 1 × 1 |
| 12 | FC-2 | — | 2 | 1 × 1 |
Figure 4Shows the flow diagram of the proposed deep learning model.
5-fold cross-validation.
| Cross-Validation | Pr | Re | F-score |
|---|---|---|---|
| 1st | 0.930 | 0.881 | 0.910 |
| 2nd | 0.927 | 0.875 | 0.910 |
| 3rd | 0.926 | 0.879 | 0.920 |
| 4th | 0.930 | 0.881 | 0.900 |
| 5th | 0.931 | 0.880 | 0.910 |
| Average |
|
|
|
Performance based on various combinations of training and testing dataset.
| Training images (%) | Testing images (%) | Pr | Re | F-score |
|---|---|---|---|---|
|
|
| 0.909 | 0.777 | 0.838 |
|
|
| 0.925 | 0.877 | 0.901 |
|
|
| 0.929 | 0.880 | 0.904 |
|
|
| 0.950 | 0.882 | 0.914 |
|
|
| 0.971 | 0.893 | 0.930 |
Quantitative performance measures for Ki-67 scoring.
| Confusion Matrix | Pr | Re | F-score | |
|---|---|---|---|---|
| 17028 | 2277 | 0.93 | 0.88 | 0.91 |
| 1287 | 15840 | |||
Figure 5Regression curve between automated and manual hotspots count.
Figure 6Shows precision and recall curve.
Figure 7Visualization of feature maps of various convolution layers.
Figure 8Ki-67 detection results by using the proposed algorithm.
Figure 9Overall detection of hotspots in breast cancer IHC images at different proliferation levels.
Overall proliferation score.
|
|
|
|
|
|---|---|---|---|
| Less proliferate (<15%) | Expert-1 | 12.87 | 13.00 |
| Expert-2 | 13.01 | 13.00 | |
|
|
|
| |
| Average proliferate (16–30%) | Expert-1 | 27.29 | 27.99 |
| Expert-2 | 28.00 | 27.99 | |
|
|
|
| |
| Highly proliferate (>31%) | Expert-1 | 90.00 | 90.00 |
| Expert-2 | 90.00 | 90.00 | |
|
|
|
|
‘MPS’: Manual Proliferation Score.
Figure 10Training performance graph.
Figure 11ROC graph for showing the overall performance of the proposed methodology.
Comparison with the existing methods.
| Comparison parameters | P. Shi | N. Khan | Proposed Methodology |
|---|---|---|---|
| Image type | Human nasopharyngeal carcinoma Xenografts | Neuroendocrine tumor | Breast cancer |
| Sample size | 100 images | 57 images | 450 images |
| Image size | 2040 × 1536 | 10 × 5 K | 2048 × 1536 |
| Image Magnification | 40x | 40x | 40x |
| Methodology used | Conventional techniques (smoothing, color channel decomposition, local feature extraction, K-means, watershed segmentation) | Conventional technique (Perceptual clustering) |
|
| Accuracy (%) | 91.8 | 94.60 |
|
| Computation time (sec) | 1.7 | 7 |
|
| CPU or GPU used | CPU | CPU |
|
| Error rate | 0.82 | Not mentioned |
|
Comparison of performance measures.
| Conditions | Pr (%) | Re (%) | F-score (%) |
|---|---|---|---|
|
| 89 | 80 | 82 |
|
|
| 65 | 76 |
|
| 93 | 80 | 86 |
|
| 87 |
| 87 |
|
|
|
|
|