| Literature DB >> 35966529 |
Zhennong Chen1, Francisco Contijoch1,2, Gabrielle M Colvert1, Ashish Manohar3, Andrew M Kahn4, Hari K Narayan5, Elliot McVeigh1,2,4.
Abstract
Background: The presence of left ventricular (LV) wall motion abnormalities (WMA) is an independent indicator of adverse cardiovascular events in patients with cardiovascular diseases. We develop and evaluate the ability to detect cardiac wall motion abnormalities (WMA) from dynamic volume renderings (VR) of clinical 4D computed tomography (CT) angiograms using a deep learning (DL) framework.Entities:
Keywords: computed tomography; deep learning; left ventricle (LV); volume rendering (VR); wall motion abnormality detection
Year: 2022 PMID: 35966529 PMCID: PMC9366190 DOI: 10.3389/fcvm.2022.919751
Source DB: PubMed Journal: Front Cardiovasc Med ISSN: 2297-055X
Figure 1Automatic generation and quantitative labeling of volume rendering video. This figure contains two parts: Rendering Generation: automatic generation of VR video (left column, white background, step 1-4 in red) and Data Labeling: quantitative labeling of the video (right column, light gray background, step a-d in blue). Rendering Generation: Step 1 and 2: Prepare the greyscale image of LV blood-pool with all other structures removed. Step 3: For each study, 6 volume renderings with 6 view angles rotated every 60 degrees around the long axis were generated. The mid-cavity AHA segment in the foreground was noted under each view. Step 4: For each view angle, a volume rendering video was created to show the wall motion across one heartbeat. Five systolic frames in VR video were presented. ED, end-diastole; ES, end-systole. Data Labeling: Step a: LV segmentation. LV, green. Step b: Quantitative RSCT was calculated for each voxel. Step c: The voxel-wise RSCT map was binarized and projected onto the pixels in the VR video. See Supplementary Material 2 for more details. In rendered RSCT map, the pixels with RSCT ≥−0.20 (abnormal wall motion) were labeled as red and those with RSCT < −0.20 (normal) were labeled as black. Step d: a video was labeled as abnormal if >35% endocardial surface has RSCT ≥−0.20 (red pixels).
Figure 2Deep learning framework. Four frames were input into a pre-trained inception-v3 individually to obtain a 2048-length feature vector for each frame. Four vectors were concatenated into a feature matrix which was then input to the next components in the framework. A Long Short-term Memory followed by fully connected layers was trained to predict a binary classification of the presence of WMA in the video. CNN, convolutional neural network; RNN, recurrent neural network.
DL classification performance in cross-validation and testing.
|
|
| ||||||||
|---|---|---|---|---|---|---|---|---|---|
|
|
| ||||||||
|
|
|
|
| ||||||
|
|
|
|
|
|
|
|
| ||
| DL | Abnormal | 443 | 36 | 87 | 7 | 276 | 45 | 57 | 4 |
| Normal | 49 | 702 | 6 | 105 | 30 | 477 | 5 | 72 | |
| Sens | 0.900 | Sens | 0.935 | Sens | 0.902 | Sens | 0.919 | ||
| Spec | 0.951 | Spec | 0.938 | Spec | 0.914 | Spec | 0.947 | ||
| Acc | 0.931 | Acc | 0.937 | Acc | 0.909 | Acc | 0.935 | ||
| κ | 0.855 | κ | 0.872 | κ | 0.808 | κ | 0.868 | ||
Two hundred five CT studies and 1230 Volume Rendered (VR) videos were used for 5-fold cross-validation. One hundred thirty-eight CT studies and 828 VR videos were in the testing. The four confusion matrices correspond to per-video classification (light gray) and per-study classification (dark gray) for cross-validation (left) and testing (right). N.
Figure 3DL classification accuracy vs. LVEF. The per-video (black) and per-study (gray) accuracy are shown in studies with (LVEF < 40%), (40 < LVEF < 60%) and (LVEF > 60%). *Indicates the significant difference.
DL classification performance in CT studies with 40 < LVEF < 60%.
|
|
| ||||||||
|---|---|---|---|---|---|---|---|---|---|
|
|
| ||||||||
|
|
|
|
| ||||||
|
|
|
|
|
|
|
|
| ||
| DL | Abnormal | 131 | 23 | 33 | 5 | 126 | 23 | 32 | 3 |
| Normal | 37 | 91 | 4 | 5 | 26 | 71 | 1 | 5 | |
| Sens | 0.780 | Sens | 0.892 | Sens | 0.829 | Sens | 0.970 | ||
| Spec | 0.798 | Spec | 0.500 | Spec | 0.755 | Spec | 0.625 | ||
| Acc | 0.787 | Acc | 0.809 | Acc | 0.801 | Acc | 0.902 | ||
| κ | 0.567 | κ | 0.407 | κ | 0.581 | κ | 0.657 | ||
Forty-seven CT studies with 40% < LVEF < 60% were in the cross-validation and 41 CT studies were in the testing. The light gray indicates per-video evaluation, dark gray indicates per-study evaluation.
Results re-binned into six regional LV views.
|
| |||||||||
|---|---|---|---|---|---|---|---|---|---|
|
|
| ||||||||
|
|
|
|
|
| κ |
|
|
| κ |
| 0 | Anterolateral | 0.845 | 0.964 | 0.922 | 0.824 | 0.886 | 0.936 | 0.920 | 0.818 |
| 60 | Inferolateral | 0.938 | 0.952 | 0.946 | 0.888 | 0.909 | 0.915 | 0.913 | 0.805 |
| 120 | Inferior | 0.879 | 0.974 | 0.932 | 0.860 | 0.917 | 0.910 | 0.913 | 0.824 |
| 180 | Inferoseptal | 0.882 | 0.946 | 0.917 | 0.832 | 0.847 | 0.861 | 0.855 | 0.705 |
| 240 | Anteroseptal | 0.963 | 0.944 | 0.951 | 0.899 | 0.927 | 0.952 | 0.942 | 0.879 |
| 300 | Anterior | 0.893 | 0.931 | 0.917 | 0.822 | 0.932 | 0.904 | 0.913 | 0.807 |
This table shows the per-video classification of our DL model when detecting WMA from each regional view of LV. See the definition of regional LV views in Section Production of volume rendering video of LV blood-pool. Sens, sensitivity; Spec, specificity; Acc, accuracy.
Comparison between DL and expert visual assessment.
|
| |||||
|---|---|---|---|---|---|
|
|
| ||||
|
|
|
|
| ||
| DL | Abnormal | 37 | 5 | 33 | 9 |
| Normal | 4 | 54 | 4 | 54 | |
| κ | 0.815 | κ | 0.729 | ||
Per-study comparison were run on 100 CT studies randomly selected from the testing cohort. The light gray indicates per-video evaluation, dark gray indicates per-study evaluation.