| Literature DB >> 29065595 |
David Vázquez1,2, Jorge Bernal1, F Javier Sánchez1, Gloria Fernández-Esparrach3, Antonio M López1,2, Adriana Romero2, Michal Drozdzal4,5, Aaron Courville2.
Abstract
Colorectal cancer (CRC) is the third cause of cancer death worldwide. Currently, the standard approach to reduce CRC-related mortality is to perform regular screening in search for polyps and colonoscopy is the screening tool of choice. The main limitations of this screening procedure are polyp miss rate and the inability to perform visual assessment of polyp malignancy. These drawbacks can be reduced by designing decision support systems (DSS) aiming to help clinicians in the different stages of the procedure by providing endoluminal scene segmentation. Thus, in this paper, we introduce an extended benchmark of colonoscopy image segmentation, with the hope of establishing a new strong benchmark for colonoscopy image analysis research. The proposed dataset consists of 4 relevant classes to inspect the endoluminal scene, targeting different clinical needs. Together with the dataset and taking advantage of advances in semantic segmentation literature, we provide new baselines by training standard fully convolutional networks (FCNs). We perform a comparative study to show that FCNs significantly outperform, without any further postprocessing, prior results in endoluminal scene segmentation, especially with respect to polyp segmentation and localization.Entities:
Mesh:
Year: 2017 PMID: 29065595 PMCID: PMC5549472 DOI: 10.1155/2017/4037190
Source DB: PubMed Journal: J Healthc Eng ISSN: 2040-2295 Impact factor: 2.682
Figure 1(a) Colonoscopy image and corresponding labeling: blue for lumen, red for background (mucosa wall), and green for polyp. (b) Proposed pipeline of a decision support system for colonoscopy.
Summary of prior database content. All frames show at least one polyp.
| Database | Number of patients | Number of seq. | Number of frames | Resolution | Annotations |
|---|---|---|---|---|---|
| CVC-ColonDB | 13 | 13 | 300 | 500 × 574 | Polyp, lumen |
| CVC-ClinicDB | 23 | 31 | 612 | 384 × 288 | Polyp |
| CVC-EndoSceneStill | 36 | 44 | 912 | 500 × 574 & 384 × 288 | Polyp, lumen, background, specularity, border (void) |
Figure 2Example of a colonoscopy image and its corresponding ground truth: (a) original image, (b) polyp mask, (c) specular highlights mask, and (d) lumen mask.
FCN8 endoluminal scene semantic segmentation results for different data augmentation techniques. The results are reported on validation set.
| Data augmentation | IoU background | IoU polyp | IoU lumen | IoU spec. | IoU mean | Acc mean |
|---|---|---|---|---|---|---|
| None | 88.93 | 44.45 | 54.02 | 25.54 | 57.88 | 92.48 |
| Zoom | 89.89 | 52.73 | 51.15 | 37.10 | 57.72 | 90.72 |
| Warp | 90.00 | 54.00 | 49.69 |
| 58.97 | 90.93 |
| Shear | 89.60 | 46.61 | 54.27 | 36.86 | 56.83 | 90.49 |
| Rotation | 90.52 | 52.83 |
| 35.81 | 58.89 | 91.38 |
| Combination |
|
| 55.08 | 35.75 |
|
|
FCN8 endoluminal scene semantic segmentation results for different numbers of classes. The results are reported on validation set. In all cases, we selected the model that provided best validation results (with or without class balancing).
| Number of classes | IoU background | IoU polyp | IoU lumen | IoU spec. | IoU mean | Acc mean |
|---|---|---|---|---|---|---|
| 4 | 92.07 | 39.37 |
|
| 57.88 | 92.48 |
| 3 | 92.19 | 50.70 | 56.48 | — | 66.46 | 92.82 |
| 2 |
|
| — | — |
|
|
Results on the test set: FCN8 with respect to previously published methods.
| Data augmentation | IoU background | IoU polyp | IoU lumen | IoU spec. | IoU mean | Acc mean | |
|---|---|---|---|---|---|---|---|
|
| |||||||
| 4 classes | None | 86.36 | 38.51 |
| 32.98 | 50.46 | 87.40 |
| 3 classes | None | 84.66 | 47.55 | 36.93 | — | 56.38 | 86.08 |
| 2 classes | None |
| 50.85 | — | — |
|
|
| 4 classes | Combination | 88.81 |
| 41.21 | 38.87 | 55.13 | 89.69 |
|
| |||||||
| [ | — | 73.93 | 22.13 | 23.82 |
| 41.19 | 75.58 |
Figure 3Examples of predictions for 4-class FCN8 model. Each subfigure represents a single frame, a ground truth annotation, and a prediction image. We use the following color-coding in the annotations: red for background (mucosa), blue for lumen, yellow for polyp, and green for specularity. (a), (b), (c), (d) show correct polyp segmentation, whereas (e), (d) show incorrect polyp segmentation.
Figure 4Localization rate of polyps as a function of IoU. The x-axis represents the degree of overlap between ground truth and model prediction. The y-axis represents the percentage of correctly localized polyps. Different color plots represent different models: FCN8 with 4 classes, FCN8 with 3 classes, and FCN8 with 2 classes and previously published method [13] (referred to as state-of-the-art in the plot).
Summary of processing times achieved by the different methods studied in the paper. FCN results are the same for all four classes considered as segmentation of the four classes is done at the same time∗.
| Method | Polyp | Lumen | Specular highlights | Background |
|---|---|---|---|---|
| FCN | 88 ms∗ | 88 ms∗ | 88 ms∗ | 88 ms∗ |
| State-of-the-art | 10000 ms | 8000 ms | 5000 ms | 23000 ms |