| Literature DB >> 35282363 |
Yutian Chen1,2,3, Wen Xie1,2, Jiawei Zhang1,2,4,5, Hailong Qiu1,2, Dewen Zeng6, Yiyu Shi6, Haiyun Yuan1,2, Jian Zhuang1,2, Qianjun Jia1,7, Yanchun Zhang5, Yuhao Dong1,2,7, Meiping Huang1,2,7, Xiaowei Xu1,2.
Abstract
Coronary artery disease (CAD) is the most common cause of death globally, and its diagnosis is usually based on manual myocardial (MYO) segmentation of MRI sequences. As manual segmentation is tedious, time-consuming, and with low replicability, automatic MYO segmentation using machine learning techniques has been widely explored recently. However, almost all the existing methods treat the input MRI sequences independently, which fails to capture the temporal information between sequences, e.g., the shape and location information of the myocardium in sequences along time. In this article, we propose a MYO segmentation framework for sequence of cardiac MRI (CMR) scanning images of the left ventricular (LV) cavity, right ventricular (RV) cavity, and myocardium. Specifically, we propose to combine conventional neural networks and recurrent neural networks to incorporate temporal information between sequences to ensure temporal consistency. We evaluated our framework on the automated cardiac diagnosis challenge (ACDC) dataset. The experiment results demonstrate that our framework can improve the segmentation accuracy by up to 2% in the Dice coefficient.Entities:
Keywords: MRI; cardiac sequences; coronary artery disease; diagnosis; myocardial segmentation; temporal consistency
Year: 2022 PMID: 35282363 PMCID: PMC8914019 DOI: 10.3389/fcvm.2022.804442
Source DB: PubMed Journal: Front Cardiovasc Med ISSN: 2297-055X
Figure 1Structure illustration of a typical CMR image. The images are the slices on the z-axis, y-axis, and x-axis, respectively, from Patient 001 in the automated cardiac diagnosis challenge (ACDC) dataset at end-diastolic (ED) frame with mask, and the second row of the figure is the raw CMR slices of Patient 001.
Figure 2Examples of hard and easy cases of CMR image (slices taken on short-axis). The first and second columns refer to hard cases and the third and fourth columns refer to easy cases.
Figure 3The proposed myocardial (MYO) segmentation architecture, which contains an initial segmentation network (ISN) and a temporal consistency based network (TCN). TSN is based on residual U-net (Res U-Net) (50), while TCN is based on ConvLSTM (51).
Figure 4The network structure of our proposed Res U-net based ISN.
Figure 5The network structure of TCN. The input of TCN is the features extracted from ISN. TCN consists of hierarchical ConvLSTMs and is able to incorporate temporal information between CMR frames.
Figure 6Frames of CMR sequence from three patients. Note the brightness heterogeneity in left ventricular (LV) and right ventricular (RV) on the first few frames. Using LSTM in TCN, the model can get more temporal information from the previous and future frames and result in more accurate segmentation.
Figure 7Illustration of our proposed bi-direction training approach.
Comparison of our proposed framework against residual U-net (Res U-net).
|
|
| |||||||||
|---|---|---|---|---|---|---|---|---|---|---|
|
|
|
|
| |||||||
|
|
|
|
|
|
|
|
|
|
| |
| Res U-net | 0.8856 | 0.8073 | 0.7178 | 0.5583 ± 0.0045 | <0.0001 | 0.8050 | 0.6841 | 0.7554 | 0.4053 ± 0.0379 | <0.0001 |
| Res U-net + | 0.8857 | 0.8082 | 0.7097 | 0.5586 ± 0.0044 | <0.0001 | 0.8056 | 0.6896 | 0.7588 | 0.4186 ± 0.0038 | <0.0001 |
| Our framework |
|
|
|
| – |
|
|
|
| – |
Res U-net+f-ConvLSTM refers to forward ConvLSTM that is trained forwardly from frame 1 to frame T. Res U-net+bi-ConvLSTM (our framework) refers to training both forward ConvLSTM and backward ConvLSTM, where backward ConvLSTM processes input CMR images from frame T to frame 1. Two-sided t-test with 95% CI is used for statistic analysis. The bold values correspond to the optimal performance.
Figure 8Visualization of CMR image segmentation results of three different patients in both end-diastolic (ED) phase and end-systolic (ES) phase. Yellow, orange, and purple areas refer to the LV, myocardial (MYO), and RV, respectively. Each row refers to the segmentation result of Res U-net, our framework, Res U-net + f-ConvLSTM, and ground truth from left to right, respectively. The white arrows in the image specifically point out the segmentation result that is inconsistent. We can see the f-ConvLSTM and bi-ConvLSTM model, which incorporates the temporal information between frames can greatly decrease the existence of such inconsistent segmentation results. Also, most errors in Res U-net + fConvLSTM and our framework are in hard cases like Patient 39, where the input CMR image has low contrast and vague contour between labeled tissue and background tissue.
Quantitative segmentation results of different models for frames in Figure 8.
|
|
|
|
| ||||
|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
| ||
| Patient 16 | ED | 0.5417 | 0.8245 | 0.5417 | 0.8433 | 0.5417 | 0.8499 |
| ES | 0.5729 | 0.8168 | 0.5729 | 0.8196 | 0.5729 | 0.8234 | |
| Patient 39 | ED | 0.6771 | 0.8132 | 0.6771 | 0.8364 | 0.6771 | 0.8348 |
| ES | 0.7396 | 0.7556 | 0.7396 | 0.7688 | 0.7396 | 0.7587 | |
| Patient 64 | ED | 0.6666 | 0.8537 | 0.6667 | 0.8653 | 0.6667 | 0.8616 |
| ES | 0.6875 | 0.8334 | 0.6875 | 0.8368 | 0.6875 | 0.8347 | |
| Patient 90 | ED | 0.6563 | 0.8099 | 0.6563 | 0.8027 | 0.6563 | 0.8043 |
| ES | 0.6563 | 0.7476 | 0.6563 | 0.7482 | 0.6563 | 0.7623 | |
Figure 9Typical segmentation error of our framework. The image on Columns 1 and 4 are the segmentation results and images on Columns 2 and 5 are the corresponding ground truth. The segmentation error is usually caused by brightness heterogeneity, lack of contrast, or the improper input image due to faulty setup of magnetic resonance system or the misoperations of operators. The white arrows in the image specifically point out the segmentation result that is inconsistent.