| Literature DB >> 31882817 |
Pejman Rasti1, Christian Wolf2, Hugo Dorez3, Raphael Sablong3, Driffa Moussata3, Salma Samiei1, David Rousseau4.
Abstract
In this article, we address the problem of the classification of the health state of the colon's wall of mice, possibly injured by cancer with machine learning approaches. This problem is essential for translational research on cancer and is a priori challenging since the amount of data is usually limited in all preclinical studies for practical and ethical reasons. Three states considered including cancer, health, and inflammatory on tissues. Fully automated machine learning-based methods are proposed, including deep learning, transfer learning, and shallow learning with SVM. These methods addressed different training strategies corresponding to clinical questions such as the automatic clinical state prediction on unseen data using a pre-trained model, or in an alternative setting, real-time estimation of the clinical state of individual tissue samples during the examination. Experimental results show the best performance of 99.93% correct recognition rate obtained for the second strategy as well as the performance of 98.49% which were achieved for the more difficult first case.Entities:
Year: 2019 PMID: 31882817 PMCID: PMC6934609 DOI: 10.1038/s41598-019-56583-9
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1Top: Human samples of colon’s wall images: healthy (left) and unhealthy (right) tissues observed from fluorescent confocal endomicroscopy. Bottom: Mouse samples of colon’s wall images: healthy (left) and unhealthy (right) tissues observed from fluorescent confocal endomicroscopy.
Number of mice in each dataset.
| Healthy mice | Mice with cancer | Mice with inflammation | |
|---|---|---|---|
| Training | 5 | 7 | 7 |
| Validation | 1 | 2 | 2 |
| Testing | 3 | 4 | 7 |
Left: Results of cross-subject training with full data, where all images of 6 healthy mice, 9 mice with cancer, and 9 mice with inflammation used for training the system. Right: Confusion matrix of cross-subject performance where our proposed CNN architecture is used.
| Left | Right | |||||
|---|---|---|---|---|---|---|
| Classifiers | Transfer learning | Accuracy | True Cancer | True Inflammation | True Healthy | |
| Proposed CNN architecture | — | Predicted Cancer | 13107 | 0 | 0 | |
| DenseNet | X | 94.54% ± 2.9 | Predicted Inflammation | 0 | 5012 | 46 |
| VGG16 + linear SVM | X | 90.60% ± 0.4 | Predicted Healthy | 0 | 75 | 2011 |
| VGG16 | X | 89.62% ± 3.3 | ||||
| ResNet50 | X | 75.93% ± 4.1 | ||||
| VGG16 | — | 74.82% ± 3.2 | ||||
| LBP features + linear SVM | — | 83.01% ± 0.4 | ||||
| Proposed method at[ | — | 77.41% ± 1.3 | ||||
Results of cross-subject training with different numbers of frozen layers when transferring the VGG16 network from ImageNet to the target dataset.
| No. Freezing | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Accuracy | 40.8% ± 17.4 | 65.6 ± 29.9% | 89.6 ± 3.3% | 89.2% ± 3.9 | 42.8% ± 21.9 | 43.4% ± 23.25 | 70% ± 24.1 | 52.8% ± 22.2 | 75.4% ± 23.9 | 82.2% ± 9.4 | 65.8% ± 29.9 | 41.2% ± 18.3 | 33% ± 0 |
Figure 2Dependency on the number of training subjects for cross-subject training (LBP features + SVM classifier).
Figure 3Example of correctly and miss classified images of the proposed CNN architecture for the cross-subject training strategy. Each cell consists from left to right of a grayscale image, a coarse localization map of the important regions in the image for the network[40], and a high-resolution class-discriminative visualization[40]. Cells with dashed lines mean that there is no miss classified images for that class.
Left: Results of cross-sample training with full data. Right: Confusion Matrix of cross-sample performance where our proposed CNN architecture is used.
| Left | Right | |||||
|---|---|---|---|---|---|---|
| Classifiers | Transfer learning | Accuracy | True Cancer | True Inflammation | True Healthy | |
| Proposed CNN architecture | — | Predicted Cancer | 13994 | 0 | 0 | |
| LBP features + linear SVM | — | 97.7% ± 0.39 | Predicted Inflammation | 0 | 4032 | 0 |
| VGG16 + linear SVM | X | 85.9% ± 0.4 | Predicted Healthy | 0 | 5 | 1849 |
| VGG16 | X | 82.12% ± 4.1 | ||||
| ResNet50 | X | 79.94% ± 4.6 | ||||
| DenseNet | X | 79.51% ± 3.8 | ||||
| VGG16 | — | 78.49% ± 1.27 | ||||
Figure 4Average of recognition rate of cross-subject (left) and cross-sample (right) training respectively with sample selection in solid red line versus a random selection of data in dashed blue line as a function of the number of images in the training dataset. Yellow and purple lines show the average recognition rate plus and minus standard deviation respectively.
Figure 5The proposed architecture of the deep network optimized for the task on the cross-validation set.