| Literature DB >> 35197110 |
Amara Tariq1, Omar Kallas2, Patricia Balthazar2, Scott Jeffery Lee2, Terry Desser3, Daniel Rubin3,4, Judy Wawira Gichoya5,2, Imon Banerjee5,2.
Abstract
BACKGROUND: Transfer learning is a common practice in image classification with deep learning where the available data is often limited for training a complex model with millions of parameters. However, transferring language models requires special attention since cross-domain vocabularies (e.g. between two different modalities MR and US) do not always overlap as the pixel intensity range overlaps mostly for images.Entities:
Keywords: BERT; Language model; Radiology report; Transfer learning; Word2vec
Mesh:
Year: 2022 PMID: 35197110 PMCID: PMC8867666 DOI: 10.1186/s13326-022-00262-8
Source DB: PubMed Journal: J Biomed Semantics
Fig. 1Cross domain finetuning language modeling schemes for MR and US domains
Fig. 2Sample MR reports: (a) Sample with LI-RADS structured template, (b) Sample free-text report
Statistics of the cohorts before processing - Stanford US dataset and EUH MRI dataset
| Stanford US dataset | EUH MRI dataset | |
|---|---|---|
| Number of unique words | 17194 | 19828 |
| Common words in two domains | 1790 | |
| Average number of words (+/- std) | 167 (+/- 39) | 197 (+/-47) |
| Average number of sentences (+/- std) | 27 (+/-7) | 32 ((+/-8) |
| Number of unique words in templated reports | 2774 | |
| Number of unique words in reports without template | 7930 | |
| Average number of words (+/- std) describing | ||
| liver related finding in templated reports | 36 (+/- 25) | 109 (+/- 63) |
| Average number of words (+/- std) describing | ||
| liver related finding in reports without template | 47 (+/- 27) | 104 (+/- 52) |
Fig. 3Sample reports: (a) Embedding of structured and unstructured US reports structured template, (b) Embedding of structured and unstructured MR reports structured template
Performance of language model and classifiers on structured MR reports (reports with template) and unstructured MR reports (reports without template). Models trained over MR domain as well as cross-domain models (MR-finetuned) have been tested
| Report document with template | Report document without template | |||||
|---|---|---|---|---|---|---|
| Precision | Recall | f1-score | Precision | Recall | f1-score | |
| 0.94 | 0.90 | 0.92 | 0.73 | 0.76 | 0.74 | |
| 0.93 | 0.95 | 0.94 | 0.94 | 0.93 | 0.94 | |
| 0.95 | 0.93 | 0.94 | 0.70 | 0.76 | 0.73 | |
| 0.95 | 0.96 | 0.95 | 0.94 | 0.92 | 0.93 | |
| 0.95 | 0.98 | 0.96 | 0.68 | 0.62 | 0.65 | |
| 0.98 | 0.96 | 0.97 | 0.91 | 0.93 | 0.92 | |
| 0.97 | 0.96 | 0.97 | 0.40 | 0.19 | 0.26 | |
| 0.97 | 0.98 | 0.98 | 0.83 | 0.93 | 0.88 | |
| 0.94 | 0.90 | 0.92 | 0.71 | 0.57 | 0.63 | |
| 0.93 | 0.95 | 0.94 | 0.91 | 0.95 | 0.92 | |
| 0.94 | 0.91 | 0.92 | 0.77 | 0.48 | 0.59 | |
| 0.94 | 0.95 | 0.94 | 0.89 | 0.97 | 0.93 | |
Performance of language model and classifiers on structured US reports (reports with template) and unstructured US reports (reports without template). Models trained over US domain as well as cross-domain models (US-finetuned) have been tested
| Report document with template | Report document without template | |||||
|---|---|---|---|---|---|---|
| Precision | Recall | f1-score | Precision | Recall | f1-score | |
| 0.68 | 0.52 | 0.59 | 0.75 | 0.27 | 0.40 | |
| 0.95 | 0.97 | 0.96 | 0.94 | 0.99 | 0.97 | |
| 0.71 | 0.59 | 0.64 | 0.67 | 0.18 | 0.29 | |
| 0.96 | 0.97 | 0.96 | 0.94 | 0.99 | 0.96 | |
| 0.68 | 0.66 | 0.67 | 0.70 | 0.64 | 0.67 | |
| 0.96 | 0.97 | 0.96 | 0.97 | 0.98 | 0.97 | |
| 0.68 | 0.72 | 0.70 | 0.64 | 0.64 | 0.64 | |
| 0.97 | 0.96 | 0.97 | 0.97 | 0.97 | 0.97 | |
| 0.91 | 0.34 | 0.50 | 0.33 | 0.09 | 0.14 | |
| 0.93 | 1.00 | 0.96 | 0.93 | 0.98 | 0.96 | |
| 0.67 | 0.34 | 0.45 | 0.33 | 0.09 | 0.14 | |
| 0.93 | 0.98 | 0.96 | 0.93 | 0.98 | 0.96 | |
Fig. 4Heatmap of liver-related text of sample MR and US reports with ‘Malignant’ label: (a) MR Word2Vec Embedding+1DCNN - predicted label: ‘Benign’, (b) MR-finetuned Word2Vec Embedding+1DCNN - predicted label: ‘Malignant’,(c) US Word2Vec Embedding+1DCNN - predicted label: ‘Benign’, (d) US-finetuned Word2Vec Embedding+1DCNN - predicted label: ‘Malignant’
Fig. 5Word2Vec Language Spaces; (a): US Language Model, (b): US-finetuned Language Model, c): New words in US-finetuned Language Model, (d): MR Language Model, (e): MR-finetuned Language Model, (f): New words in MR-finetuned Language Model
Fig. 6Mean vectors of reports in BERT and Word2Vec language spaces; left column: BERT space, right column: Word2Vec space; row 1: MR reports with MR language models, row 2: MR reports with MR-finetuned language models, row 3: US reports with US language models, row 4: US reports with US-finetuned language models