Nassim Bouteldja1,2, Barbara M Klinkhammer2, Tarek Schlaich1, Peter Boor2,3, Dorit Merhof1,4. 1. Institute of Imaging and Computer Vision, RWTH Aachen University, Aachen, Germany. 2. Institute of Pathology, RWTH Aachen University Hospital, Aachen, Germany. 3. Department of Nephrology and Immunology, RWTH Aachen University Hospital, Aachen, Germany. 4. Fraunhofer Institute for Digital Medicine MEVIS, Bremen, Germany.
Abstract
Background: In digital pathology, many image analysis tasks are challenged by the need for large and time-consuming manual data annotations to cope with various sources of variability in the image domain. Unsupervised domain adaptation based on image-to-image translation is gaining importance in this field by addressing variabilities without the manual overhead. Here, we tackle the variation of different histological stains by unsupervised stain-to-stain translation to enable a stain-independent applicability of a deep learning segmentation model. Methods: We use CycleGANs for stain-to-stain translation in kidney histopathology, and propose two novel approaches to improve translational effectivity. First, we integrate a prior segmentation network into the CycleGAN for a self-supervised, application-oriented optimization of translation through semantic guidance, and second, we incorporate extra channels to the translation output to implicitly separate artificial meta-information otherwise encoded for tackling underdetermined reconstructions. Results: The latter showed partially superior performances to the unmodified CycleGAN, but the former performed best in all stains providing instance-level Dice scores ranging between 78% and 92% for most kidney structures, such as glomeruli, tubules, and veins. However, CycleGANs showed only limited performance in the translation of other structures, e.g. arteries. Our study also found somewhat lower performance for all structures in all stains when compared to segmentation in the original stain. Conclusions: Our study suggests that with current unsupervised technologies, it seems unlikely to produce "generally" applicable simulated stains.
Background: In digital pathology, many image analysis tasks are challenged by the need for large and time-consuming manual data annotations to cope with various sources of variability in the image domain. Unsupervised domain adaptation based on image-to-image translation is gaining importance in this field by addressing variabilities without the manual overhead. Here, we tackle the variation of different histological stains by unsupervised stain-to-stain translation to enable a stain-independent applicability of a deep learning segmentation model. Methods: We use CycleGANs for stain-to-stain translation in kidney histopathology, and propose two novel approaches to improve translational effectivity. First, we integrate a prior segmentation network into the CycleGAN for a self-supervised, application-oriented optimization of translation through semantic guidance, and second, we incorporate extra channels to the translation output to implicitly separate artificial meta-information otherwise encoded for tackling underdetermined reconstructions. Results: The latter showed partially superior performances to the unmodified CycleGAN, but the former performed best in all stains providing instance-level Dice scores ranging between 78% and 92% for most kidney structures, such as glomeruli, tubules, and veins. However, CycleGANs showed only limited performance in the translation of other structures, e.g. arteries. Our study also found somewhat lower performance for all structures in all stains when compared to segmentation in the original stain. Conclusions: Our study suggests that with current unsupervised technologies, it seems unlikely to produce "generally" applicable simulated stains.
Authors: Stanley J Robboy; Sally Weintraub; Andrew E Horvath; Bradden W Jensen; C Bruce Alexander; Edward P Fody; James M Crawford; Jimmy R Clark; Julie Cantor-Weinberg; Megha G Joshi; Michael B Cohen; Michael B Prystowsky; Sarah M Bean; Saurabh Gupta; Suzanne Z Powell; V O Speights; David J Gross; W Stephen Black-Schaffer Journal: Arch Pathol Lab Med Date: 2013-06-05 Impact factor: 5.534
Authors: Faisal Mahmood; Daniel Borders; Richard J Chen; Gregory N Mckay; Kevan J Salimian; Alexander Baras; Nicholas J Durr Journal: IEEE Trans Med Imaging Date: 2020-10-28 Impact factor: 10.048
Authors: Michael Gadermayr; Laxmi Gupta; Vitus Appel; Peter Boor; Barbara M Klinkhammer; Dorit Merhof Journal: IEEE Trans Med Imaging Date: 2019-02-14 Impact factor: 10.048
Authors: Yair Rivenson; Hongda Wang; Zhensong Wei; Kevin de Haan; Yibo Zhang; Yichen Wu; Harun Günaydın; Jonathan E Zuckerman; Thomas Chong; Anthony E Sisk; Lindsey M Westbrook; W Dean Wallace; Aydogan Ozcan Journal: Nat Biomed Eng Date: 2019-03-04 Impact factor: 25.671
Authors: Meyke Hermsen; Thomas de Bel; Marjolijn den Boer; Eric J Steenbergen; Jesper Kers; Sandrine Florquin; Joris J T H Roelofs; Mark D Stegall; Mariam P Alexander; Byron H Smith; Bart Smeets; Luuk B Hilbrands; Jeroen A W M van der Laak Journal: J Am Soc Nephrol Date: 2019-09-05 Impact factor: 10.121
Authors: Peter Bankhead; Maurice B Loughrey; José A Fernández; Yvonne Dombrowski; Darragh G McArt; Philip D Dunne; Stephen McQuaid; Ronan T Gray; Liam J Murray; Helen G Coleman; Jacqueline A James; Manuel Salto-Tellez; Peter W Hamilton Journal: Sci Rep Date: 2017-12-04 Impact factor: 4.379
Authors: Nassim Bouteldja; Barbara M Klinkhammer; Roman D Bülow; Patrick Droste; Simon W Otten; Saskia Freifrau von Stillfried; Julia Moellmann; Susan M Sheehan; Ron Korstanje; Sylvia Menzel; Peter Bankhead; Matthias Mietsch; Charis Drummer; Michael Lehrke; Rafael Kramann; Jürgen Floege; Peter Boor; Dorit Merhof Journal: J Am Soc Nephrol Date: 2020-11-05 Impact factor: 10.121
Authors: Catherine P Jayapandian; Yijiang Chen; Andrew R Janowczyk; Matthew B Palmer; Clarissa A Cassol; Miroslav Sekulic; Jeffrey B Hodgin; Jarcy Zee; Stephen M Hewitt; John O'Toole; Paula Toro; John R Sedor; Laura Barisoni; Anant Madabhushi Journal: Kidney Int Date: 2020-08-22 Impact factor: 10.612