| Literature DB >> 34285783 |
Xihao Chen1, Jingya Yu1, Shenghua Cheng1, Xiebo Geng1, Sibo Liu1, Wei Han1, Junbo Hu2, Li Chen3, Xiuli Liu1, Shaoqun Zeng1.
Abstract
Diverse styles of cytopathology images have a negative effect on the generalization ability of automated image analysis algorithms. This article proposes an unsupervised method to normalize cytopathology image styles. We design a two-stage style normalization framework with a style removal module to convert the colorful cytopathology image into a gray-scale image with a color-encoding mask and a domain adversarial style reconstruction module to map them back to a colorful image with user-selected style. Our method enforces both hue and structure consistency before and after normalization by using the color-encoding mask and per-pixel regression. Intra-domain and inter-domain adversarial learning are applied to ensure the style of normalized images consistent with the user-selected for input images of different domains. Our method shows superior results against current unsupervised color normalization methods on six cervical cell datasets from different hospitals and scanners. We further demonstrate that our normalization method greatly improves the recognition accuracy of lesion cells on unseen cytopathology images, which is meaningful for model generalization.Entities:
Keywords: Cytopathology images; Domain adversarial networks; Generative adversarial learning; Unsupervised image style normalization
Year: 2021 PMID: 34285783 PMCID: PMC8273362 DOI: 10.1016/j.csbj.2021.06.025
Source DB: PubMed Journal: Comput Struct Biotechnol J ISSN: 2001-0370 Impact factor: 7.271
Fig. 1Cytopathology image style diversity. Image style and visual appearance can differ greatly depending on slide staining procedures and scanning devices. Different rows of patches stand for different hospitals while different columns represent different scanning devices.
Fig. 2Cell cross-color phenomenon of CycleGAN-based normalization methods. Results of CycleGAN for style normalization: the target image to be normalized (left), the normalized target image with user-selected style (middle) and the image reconstructed back to the input style (right).
Fig. 3Image style normalization process with two-stage domain adversarial style normalization framework. The proposed framework consists of a style reconstruction network G(.), an intra-domain discriminator D1(.) and an inter-domain discriminator D2(.). It normalizes the target image with style removal and style reconstruction. In the process of style reconstruction, we take multiple losses (including GAN1 Loss (), GAN2 Loss () and L1 Loss ()) to ensure the reconstructed style consistent with the source style and the reconstructed structure identical with its origin.al input images.
Cervical image style normalization datasets.
| Name | Hospital | Scanner | Training slides | Training patches | Test slides | Test patches |
|---|---|---|---|---|---|---|
| S | Hospital1 | Device1 | 170 | 100,000 | 40 | 6,000 |
| T1 | Hospital1 | Device2 | 343 | 50,000 | 119 | 5,000 |
| T2 | Hospital1 | Device3 | 169 | 50,000 | 26 | 5,000 |
| T3 | Hospital1 | Device4 | 60 | 50,000 | 20 | 5,000 |
| T4 | Hospital2 | Device2 | 48 | 50,000 | 16 | 5,000 |
| T5 | Hospital2 | Device4 | 39 | 50,000 | 10 | 5,000 |
Cervical lesion cell recognition datasets.
| Name | Positive Slides | Annotations | Positive Patches | Negative Slides | Negative Patches | |
|---|---|---|---|---|---|---|
| S-train | 84 | 29,360 | 200,000 | 86 | 200,000 | |
| S-test | 20 | 2,734 | 4,000 | 20 | 4,000 | |
| T1-test | 19 | 1,292 | 2,500 | 100 | 2,500 | |
| T2-test | 15 | 728 | 1,300 | 11 | 1,300 | |
| T3-test | 11 | 602 | 1,100 | 9 | 1,100 |
Fig. 4Examples of style normalized images of different normalization methods in the target domain T1. The first row refers to the typical examples of source domain. In general, normalized images generated by methods based on deep learning match the image style of source domain better than traditional methods. Phenomenon of hue inconsistency may be found in CycleGAN-based methods.
Color distribution similarity and structure consistency metrics of style normalization methods at different source-target domain settings. “Our*” refers to , “Our” refers to .
| Method | T1 | T2 | T3 | T4 | T5 |
|---|---|---|---|---|---|
| Bhattacharyya distance | |||||
| SPCN | 0.270/0.803 | 0.329/0.702 | 0.351/0.674 | 0.311/0.748 | 0.381/0.638 |
| Macenko’s | 0.127/0.867 | 0.312/0.657 | 0.234/0.764 | 0.191/0.833 | 0.368/0.615 |
| Reinhard’s | 0.320/0.674 | 0.336/0.654 | 0.392/0.572 | 0.358/0.643 | 0.381/0.603 |
| Khan’s | 0.374/0.654 | 0.447/0.576 | 0.403/0.616 | 0.384/0.639 | 0.404/0.612 |
| Gupta’s | 0.197/0.825 | 0.258/0.756 | 0.200/0.790 | 0.217/0.789 | 0.241/0.743 |
| Zheng’s | 0.155/0.864 | 0.298/0.675 | 0.311/0.706 | 0.206/0.799 | 0.356/0.660 |
| CycleGAN | 0.088/0.929 | 0.081/0.927 | 0.110/0.909 | 0.087/0.920 | 0.104/0.897 |
| StainGAN | 0.123/0.900 | 0.141/0.885 | 0.110/0.908 | 0.142/0.847 | 0.123/0.881 |
| Tellez’s | 0.121/0.896 | 0.367/0.625 | 0.234/0.748 | 0.173/0.847 | 0.225/0.766 |
| Our* | 0.092/0.905 | 0.107/0.891 | 0.161/0.844 | 0.112/0.882 | 0.120/0.881 |
| Our | |||||
| SSIM | |||||
| SPCN | 0.890/21.06 | 0.874/16.18 | 0.895/21.91 | 0.878/20.96 | 0.870/18.11 |
| Macenko’s | 0.916/21.82 | 0.544/16.68 | 0.916/22.17 | 0.895/21.27 | 0.887/17.88 |
| Reinhard’s | 0.703/14.64 | 0.772/13.20 | 0.681/14.82 | 0.672/13.97 | 0.724/16.28 |
| Khan’s | 0.579/12.08 | 0.643/12.95 | 0.577/8.81 | 0.528/12.02 | 0.550/8.98 |
| Gupta’s | 0.947/23.92 | 0.966/22.47 | 0.970/23.90 | 0.931/22.11 | 0.954/19.84 |
| Zheng’s | 0.963/19.14 | 0.972/20.85 | |||
| CycleGAN | 0.831/21.14 | 0.943/22.97 | 0.884/17.09 | 0.858/19.92 | 0.734/14.45 |
| StainGAN | 0.956/27.60 | 0.958/24.44 | 0.891/16.95 | 0.854/20.05 | 0.777/15.23 |
| Tellez’s | 0.988/ | 0.964/19.32 | |||
| Our* | 0.970/26.75 | 0.963/25.56 | 0.947/21.16 | 0.965/25.71 | 0.937/20.44 |
| Our | 0.962/26.87 | 0.929/22.93 | 0.880/17.03 | 0.892/22.42 | 0.842/16.29 |
Fig. 5Distribution matching property in “L*a*b*” color space of different normalization methods at T1S. Each curve in the plots represents the source domain S or the target domain T1 or its normalized target domain by different normalization methods. The closer two curves are, the better their color distribution similarity is.
Ablation experiment results of the color-encoding mask in the target domain T1.
| Metrics | With mask | With mask | Without mask |
|---|---|---|---|
| using “ | using “ | ||
| Bhattacharyya distance | 0.059 | 0.092 | 0.078 |
| Histogram intersection | 0.940 | 0.905 | 0.932 |
| SSIM | 0.962 | 0.920 | 0.948 |
| PSNR | 26.87 | 23.64 | 25.25 |
Fig. 6Ablation experiment results of the color-encoding mask in the target domain T1. The normalized images without using the color mask or with using unsuitable color mask have the cell hue inconsistency problem.
Ablation experiment results of training individual and united target domains T1-T5.
| Individual target domains | |||||
|---|---|---|---|---|---|
| Metrics | T1 | T2 | T3 | T4 | T5 |
| Bhattacharyya distance | 0.059 | 0.060 | 0.087 | 0.076 | 0.070 |
| Histogram intersection | 0.940 | 0.943 | 0.927 | 0.928 | 0.933 |
| SSIM | 0.962 | 0.929 | 0.880 | 0.892 | 0.842 |
| PSNR | 26.87 | 22.93 | 17.03 | 22.42 | 16.29 |
| United target domains | |||||
| Metrics | T1 | T2 | T3 | T4 | T5 |
| Bhattacharyya distance | 0.081 | 0.179 | 0.148 | 0.093 | 0.153 |
| Histogram intersection | 0.932 | 0.823 | 0.868 | 0.918 | 0.851 |
| SSIM | 0.915 | 0.774 | 0.895 | 0.912 | 0.889 |
| PSNR | 22.95 | 17.90 | 19.43 | 22.78 | 18.74 |
Ablation experiment results of different loss weights (, , ) in the the target domain T1.
| Metrics | 100:0.5:1 | 100:1:1 | 100:1:0.5 | 100:1:0 |
|---|---|---|---|---|
| Bhattacharyya distance | 0.073 | 0.059 | 0.075 | 0.092 |
| Histogram intersection | 0.936 | 0.940 | 0.930 | 0.905 |
| SSIM | 0.936 | 0.962 | 0.946 | 0.970 |
| PSNR | 24.46 | 26.87 | 25.26 | 26.75 |
Improved accuracies (%) of unseen target domain images.
| Method | T1 | T2 | T3 | Avg |
|---|---|---|---|---|
| Wild | 64.6 | 61.4 | 71.9 | 62.1 |
| SPCN | 66.7 | 49.8 | 71.6 | 62.7 |
| Macenko’s | 71.1 | 45.7 | 60.2 | 59.0 |
| Reinhard’s | 62.3 | 63.4 | 61.9 | 62.5 |
| Khan’s | 50.0 | 51.5 | 50.5 | 50.7 |
| Gupta’s | 77.5 | 62.1 | 57.9 | 65.8 |
| Zheng’s | 74.5 | 50.0 | 70.3 | 64.9 |
| Cycle-GAN | 75.3 | 67.6 | 60.4 | 67.8 |
| StainGAN | 66.4 | 66.2 | 72.9 | 68.5 |
| Tellez’s | 51.8 | 50.0 | 53.3 | 51.7 |
| Our |
Fig. 7T-SNE of the task network ’s feature representations from wild images (a) and corresponding normalized images by our method (b). Green and blue dots refer to positive and negative patches of S; purple and red dots refer to positive and negative patches of T2.