| Literature DB >> 35647616 |
Zhanghexuan Ji1, Mohammad Abuzar Shaikh1, Dana Moukheiber1, Sargur N Srihari1, Yifan Peng2, Mingchen Gao1.
Abstract
Self-supervised learning provides an opportunity to explore unlabeled chest X-rays and their associated free-text reports accumulated in clinical routine without manual supervision. This paper proposes a Joint Image Text Representation Learning Network (JoImTeRNet) for pre-training on chest X-ray images and their radiology reports. The model was pre-trained on both the global image-sentence level and the local image region-word level for visual-textual matching. Both are bidirectionally constrained on Cross-Entropy based and ranking-based Triplet Matching Losses. The region-word matching is calculated using the attention mechanism without direct supervision about their mapping. The pre-trained multi-modal representation learning paves the way for downstream tasks concerning image and/or text encoding. We demonstrate the representation learning quality by cross-modality retrievals and multi-label classifications on two datasets: OpenI-IU and MIMIC-CXR. Our code is available at https://github.com/mshaikh2/JoImTeR_MLMI_2021.Entities:
Keywords: Attention; Multi-modality; Self-supervised learning
Year: 2021 PMID: 35647616 PMCID: PMC9134785 DOI: 10.1007/978-3-030-87589-3_12
Source DB: PubMed Journal: Mach Learn Med Imaging