| Literature DB >> 33285483 |
Donglin Di1, Feng Shi2, Fuhua Yan3, Liming Xia4, Zhanhao Mo5, Zhongxiang Ding6, Fei Shan7, Bin Song8, Shengrui Li1, Ying Wei2, Ying Shao2, Miaofei Han2, Yaozong Gao2, He Sui5, Yue Gao9, Dinggang Shen10.
Abstract
The coronavirus disease, named COVID-19, has become the largest global public health crisis since it started in early 2020. CT imaging has been used as a complementary tool to assist early screening, especially for the rapid identification of COVID-19 cases from community acquired pneumonia (CAP) cases. The main challenge in early screening is how to model the confusing cases in the COVID-19 and CAP groups, with very similar clinical manifestations and imaging features. To tackle this challenge, we propose an Uncertainty Vertex-weighted Hypergraph Learning (UVHL) method to identify COVID-19 from CAP using CT images. In particular, multiple types of features (including regional features and radiomics features) are first extracted from CT image for each case. Then, the relationship among different cases is formulated by a hypergraph structure, with each case represented as a vertex in the hypergraph. The uncertainty of each vertex is further computed with an uncertainty score measurement and used as a weight in the hypergraph. Finally, a learning process of the vertex-weighted hypergraph is used to predict whether a new testing case belongs to COVID-19 or not. Experiments on a large multi-center pneumonia dataset, consisting of 2148 COVID-19 cases and 1182 CAP cases from five hospitals, are conducted to evaluate the prediction accuracy of the proposed method. Results demonstrate the effectiveness and robustness of our proposed method on the identification of COVID-19 in comparison to state-of-the-art methods.Entities:
Keywords: COVID-19 pneumonia; Hypergraph learning; Uncertainty calculation; Vertex-weighted
Year: 2020 PMID: 33285483 PMCID: PMC7690277 DOI: 10.1016/j.media.2020.101910
Source DB: PubMed Journal: Med Image Anal ISSN: 1361-8415 Impact factor: 8.545
Fig. 1Illustration of lung CT image, infection, lung lobes, and pulmonary segments on a CAP case (left) and a COVID-19 case (right).
Fig. 2Illustration of our proposed Uncertainty Vertex-weighted Hypergraph Learning (UVHL) method for COVID-19 identification. Given a bunch of patients, “Data Uncertainty Measurement” stage calculates and generates the uncertainty score for each case of CAP and COVID-19, denoted as green Gvely. The “Uncertainty-vertex Hypergraph Modelling” then constructs the hypergraph structure for both labeled cases and unknown cases, former of which are embedded and denoted with the color bars. The stage “Uncertainty-vertex Hypergraph Learning” can learn and classify all of the cases into the two diseases, consequently.
Fig. 3Besides the hyperedge weights, the uncertainty-vertex hypergraph contains the uncertainty score of each vertex.
The definition of the confusion matrix for COVID-19 identification.
| Classify as COVID-19 | Classify as CAP | |
|---|---|---|
| True Positive ( | False Negative ( | |
| False Positive ( | True Negative ( |
Fig. 4The prediction accuracy of UVHL and compared methods. The results show that UVHL outperforms other methods for all metrics.
Prediction accuracy comparison of different methods on the pneumonia dataset. For each 10-fold, we compute the accuracy of the proposed method on testing data, and compare them with those of UVHL via paired t-test to generate the p-values for each metric. (“” denotes significance level is reached as -value .)
| Methods | ACC | SEN | SPEC | BAC | PPV | NPV | |||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| ( | 0.84084 | 0.85714 | 0.81034 | 0.83374 | 0.89423 | 0.75200 | |||||||
| ( | 0.84685 | 0.86175 | 0.81897 | 0.84036 | 0.89904 | 0.76000 | |||||||
| ( | 0.85135 | 0.86327 | 0.83052 | 0.84790 | 0.90256 | 0.76866 | |||||||
| ( | 0.86486 | 0.89191 | 0.81743 | 0.85467 | 0.89898 | 0.80547 | |||||||
| ( | |||||||||||||
Experimental comparison on the data uncertainty measurement. For the “proposed uncertainty”, we compute its accuracy on testing data, and compare them with other settings via paired t-test to generate the -values. (“” denotes significance level is reached as -value .)
| Weighting strategy | ACC | SEN | SPEC | BAC | PPV | NPV | |
|---|---|---|---|---|---|---|---|
| 0.85586 | 0.88426 | 0.80342 | 0.84384 | 0.89252 | 0.789912 | ||
| 0.86066 | 0.87021 | 0.84442 | 0.85731 | 0.90983 | 0.78137 | ||
| 0.87387 | 0.918919 | 0.78378 | 0.85135 | 0.89474 | 0.82857 | ||
| 0.88589 | 0.90741 | 0.87678 | 0.83193 | ||||
| 0.84000 | 0.90654 |
Experimental comparison on different feature types and their combination. For the “both” feature type, we compute its accuracy on testing data, and compare them with each other via paired t-test to generate the p-values. (“” denotes significance level is reached as -value .)
| Feature types | ACC | SEN | SPEC | BAC | PPV | NPV |
|---|---|---|---|---|---|---|
| 0.85886 | 0.90323 | 0.77586 | 0.83954 | 0.88288 | 0.81081 | |
| 0.85946 | 0.86982 | 0.85582 | 0.78012 | |||
| 0.84000 | 0.90654 |
Experimental comparison on different feature types and their combination. For each factor, we compute the accuracy of the proposed method on testing data, and compare them with each other via paired t-test to generate the -values. (“” denotes significance level is reached as -value .)
| Factors | % / Subjs | ACC | SEN | SPEC | BAC | PPV | NPV |
|---|---|---|---|---|---|---|---|
| 50.48% | 0.88248 | ||||||
| 49.52% | 0.88963 | 0.92768 | 0.83302 | 0.88035 | 0.89209 | ||
| 54.20% | 0.89236 | 0.87611 | |||||
| 45.80% | 0.93078 | 0.82680 | 0.87879 | 0.81877 | |||
| 59.91% | 0.93084 | 0.76660 | |||||
| 40.09% | 0.87491 | 0.82962 | 0.88247 | 0.80451 |
Fig. 5Prediction accuracy comparison with respect to different scales of training data.