| Literature DB >> 33324026 |
Zhuomin Zhang1, Dolzodmaa Davaasuren1, Chenyan Wu1, Jeffery A Goldstein2, Alison D Gernand1, James Z Wang1.
Abstract
We propose a multi-region saliency-aware learning (MSL) method for cross-domain placenta image segmentation. Unlike most existing image-level transfer learning methods that fail to preserve the semantics of paired regions, our MSL incorporates the attention mechanism and a saliency constraint into the adversarial translation process, which can realize multi-region mappings in the semantic level. Specifically, the built-in attention module serves to detect the most discriminative semantic regions that the generator should focus on. Then we use the attention consistency as another guidance for retaining semantics after translation. Furthermore, we exploit the specially designed saliency-consistent constraint to enforce the semantic consistency by requiring the saliency regions unchanged. We conduct experiments using two real-world placenta datasets we have collected. We examine the efficacy of this approach in (1) segmentation and (2) prediction of the placental diagnoses of fetal and maternal inflammatory response (FIR, MIR). Experimental results show the superiority of the proposed approach over the state of the art.Entities:
Keywords: Pathology; Photo image analysis; Placenta; Transfer learning
Year: 2020 PMID: 33324026 PMCID: PMC7727399 DOI: 10.1016/j.patrec.2020.10.004
Source DB: PubMed Journal: Pattern Recognit Lett ISSN: 0167-8655 Impact factor: 3.756
Fig. 1Challenges for cross-domain placenta image segmentation. (a) Cross-domain images: Placenta images from two different hospitals. (b) Their corresponding segmentation results using the trained model on the first dataset.
Fig. 2The pipeline of the proposed approach. As shown in the left part, the introduced attention network A can divide the placenta image x into attended regions such as ruler and background, and unattended regions that include disc and cord. The translated image is a combination of translated attended parts and original unattended parts. The attention-consistent loss and a saliency-consistent loss are added to preserve the semantics in together with the image-level adaptation as composed of the pixel GAN loss Lgan and the cycle loss Lcyc.
Segmentation evaluation accuracy.
| Method | Pixel Accu. | Mean Accu. | Mean IoU |
|---|---|---|---|
| No adaptation | 0.5256 | 0.4164 | 0.2049 |
| CycleGAN | 0.7022 | 0.5850 | 0.3508 |
| AGGAN | 0.7807 | 0.6497 | 0.4672 |
| MSL(w/o | 0.8291 | 0.6502 | 0.4812 |
| MSL(w/o | 0.7852 | 0.6591 | 0.4722 |
| MSL(w/o | 0.7593 | 0.6203 | 0.3874 |
| MSL |
Fig. 3Segmentation result comparisons. (a) Original images. (b) Ground truth. (c) Segmentation results without adaptation. (d)(f)(h) Translation results using CycleGAN, AGGAN, and our model, respectively. (e)(g)(i) Segmentation results using the translated images to the left of it.
Fig. 4Confusion matrices of baseline, CycleGAN, AGGAN and MSL.
Fig. 5Ablation Study. Left: original images. Middle: results of MSL. Right: results without the corresponding loss.
Fig. 6Prediction results comparison using original or segmented images for FIR (top) and MIR (bottom).