| Literature DB >> 32332804 |
Young Sang Cho1, Kyeongwon Cho2,3, Chae Jung Park2,4, Myung Jin Chung2,5, Jong Hyuk Kim4, Kyunga Kim4,6, Yi-Kyung Kim5, Hyung-Jin Kim5, Jae-Wook Ko7, Baek Hwan Cho8,9, Won-Ho Chung10.
Abstract
Ménière's Disease (MD) is difficult to diagnose and evaluate objectively over the course of treatment. Recently, several studies have reported MD diagnoses by MRI-based endolymphatic hydrops (EH) analysis. However, this method is time-consuming and complicated. Therefore, a fast, objective, and accurate evaluation tool is necessary. The purpose of this study was to develop an algorithm that can accurately analyze EH on intravenous (IV) gadolinium (Gd)-enhanced inner-ear MRI using artificial intelligence (AI) with deep learning. In this study, we developed a convolutional neural network (CNN)-based deep-learning model named INHEARIT (INner ear Hydrops Estimation via ARtificial InTelligence) for the automatic segmentation of the cochlea and vestibule, and calculation of the EH ratio in the segmented region. Measurement of the EH ratio was performed manually by a neuro-otologist and neuro-radiologist and by estimation with the INHEARIT model and were highly consistent (intraclass correlation coefficient = 0.971). This is the first study to demonstrate that automated EH ratio measurements are possible, which is important in the current clinical context where the usefulness of IV-Gd inner-ear MRI for MD diagnosis is increasing.Entities:
Mesh:
Substances:
Year: 2020 PMID: 32332804 PMCID: PMC7181627 DOI: 10.1038/s41598-020-63887-8
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1The proposed INHEARIT framework. MRC images (384 ×324 pixels) are cropped into patches (100 × 100 pixels) during the data preparation stage, and the patches are fed into the deep-learning network. The segmentation results are applied to HYDROPS-Mi2 patches as masks, and the endolymphatic hydrops (EH) ratio is calculated from the segmented areas.
Performance of INHEARIT (INner ear Hydrops Estimation via ARtificial InTelligence) trained with the dataset according to the annotation: fully annotated dataset (FA), selectively annotated dataset (SA), and both FA and SA (FASA).
| Model | Experiment | Number of Original Patches | Dataset | Augmentation | IoU |
|---|---|---|---|---|---|
| concat 3into1VGG | 262 | SA | Low | 0.497 ± 0.022 | |
| Moderate | |||||
| High | 0.528 ± 0.017 | ||||
| 3into3VGG | 1 | 110 | FA | Low | 0.580 ± 0.637 |
| Moderate | 0.646 ± 0.033 | ||||
| High | |||||
| 2 | 262 | SA | Low | 0.620 ± 0.026 | |
| Moderate | |||||
| High | 0.711 ± 0.037 | ||||
| 3 | 372 | FASA | Low | 0.635 ± 0.013 | |
| Moderate | 0.705 ± 0.002 | ||||
| High |
IoU = intersection-over-union; Avg = average; SD = standard deviation; FA = fully annotated dataset; SA = selectively annotated dataset; FASA = both FA and SA.
The two models of concat3into1VGG (three slices were concatenated and entered into a VGG-based network) and 3into3VGG (three slices were independently fed into each of the VGG-based networks) were compared. Numbers in bold indicate the highest performance for each item.
Figure 2AI-based segmentation results from (A) the selectively annotated (SA) dataset and (B) the fully annotated (FA) dataset. Those examples show that AI-based prediction performs well compared to physicians’ annotations (ground truth).
Performance of INHEARIT fine-tuned with item 2 (SA, moderate augmentation) with various dataset combinations.
| Experiment | Number of Original Patches | Dataset | Augmentation | IoU | IoU(2)a |
|---|---|---|---|---|---|
| 4 | 110 | SA → FA | Low | 0.686 ± 0.026 | + |
| Moderate | 0.724 ± 0.030 | + | |||
| High | + | ||||
| 5 | 372 | SA → FASA | Low | 0.702 ± 0.032 | + |
| Moderate | 0.760 ± 0.014 | + | |||
| High | + | ||||
| 6 | 60 | SA → FA’ | Low | 0.610 ± 0.040 | − |
| Moderate | + | ||||
| High | 0.716 ± 0.027 | + | |||
| 7 | 322 | SA → FA’SA | Low | 0.642 ± 0.025 | + |
| Moderate | − | ||||
| High | 0.674 ± 0.016 | − |
IoU = intersection-over-union; Avg = average; SD = standard deviation; SA = selectively annotated dataset; FA = fully annotated dataset; FA’ = main organs only in FA, FASA = both SA and FA, FA’SA = main organs only in FASA; IoU = intersection-over-union.
aLoss (−) or gain (+) in IoU compared with Experiment 2 in Table 1 at the same augmentation scale.
Numbers in bold indicate the highest performance for each item.
Figure 3Agreement analysis of the endolymphatic hydrops ratio via intraclass correlation coefficient (ICC), showing (A) ICC mean values (maximum and minimum) and p-values for the overall, cochlea-only, and vestibule-only and (B) scatter plots for all ICCs between the ground truth and prediction by INHEARIT network values.
Figure 4Bland-Altman plot for the total (cochlea and vestibule) dataset. The green line indicates the upper limit of agreement (ULoA), while the red line indicates the lower limit of agreement (LLoA).
Figure 5Concept of the 3into3VGG of the INHEARIT network. The network received three independent MRC images into each convolutional network (convnet), and features from the three layers are summated before the deconvolutional layers. Two skip connections from the main convolutional network are connected to the deconvolutional network (deconvenet).