Literature DB >> 33821209

Rapid quantification of COVID-19 pneumonia burden from computed tomography with convolutional LSTM networks.

Kajetan Grodecki, Aditya Killekar, Andrew Lin, Sebastien Cadet, Priscilla McElhinney, Aryabod Razipour, Cato Chan, Barry D Pressman, Peter Julien, Judit Simon, Pal Maurovich-Horvat, Nicola Gaibazzi, Udit Thakur, Elisabetta Mancini, Cecilia Agalbato, Jiro Munechika, Hidenari Matsumoto, Roberto Menè, Gianfranco Parati, Franco Cernigliaro, Nitesh Nerlekar, Camilla Torlasco, Gianluca Pontone, Damini Dey, Piotr J Slomka.   

Abstract

Quantitative lung measures derived from computed tomography (CT) have been demonstrated to improve prognostication in coronavirus disease (COVID-19) patients, but are not part of the clinical routine since required manual segmentation of lung lesions is prohibitively time-consuming. We propose a new fully automated deep learning framework for rapid quantification and differentiation between lung lesions in COVID-19 pneumonia from both contrast and non-contrast CT images using convolutional Long Short-Term Memory (ConvLSTM) networks. Utilizing the expert annotations, model training was performed 5 times with separate hold-out sets using 5-fold cross-validation to segment ground-glass opacity and high opacity (including consolidation and pleural effusion). The performance of the method was evaluated on CT data sets from 197 patients with positive reverse transcription polymerase chain reaction test result for SARS-CoV-2. Strong agreement between expert manual and automatic segmentation was obtained for lung lesions with a Dice score coefficient of 0.876 $\pm$ 0.005; excellent correlations of 0.978 and 0.981 for ground-glass opacity and high opacity volumes. In the external validation set of 67 patients, there was dice score coefficient of 0.767 $\pm$ 0.009 as well as excellent correlations of 0.989 and 0.996 for ground-glass opacity and high opacity volumes. Computations for a CT scan comprising 120 slices were performed under 2 seconds on a personal computer equipped with NVIDIA Titan RTX graphics processing unit. Therefore, our deep learning-based method allows rapid fully-automated quantitative measurement of pneumonia burden from CT and may generate results with an accuracy similar to the expert readers.

Entities:  

Year:  2021        PMID: 33821209      PMCID: PMC8020980     

Source DB:  PubMed          Journal:  ArXiv        ISSN: 2331-8422


Introduction

Coronavirus disease 2019 (COVID-19) is a global pandemic and public health crisis of catastrophic proportions, with over 123 million confirmed cases world-wide as of March 22 2021 [1]. While vaccines are now available–they are not 100% effective, new strains are emerging and not all the population will be vaccinated. It is likely that annual vaccinations will be necessary and continuous monitoring for the disease will be needed. While the diagnosis of COVID-19 relies on a reverse transcription polymerase chain reaction (RT-PCR) test in respiratory tract specimens, computed tomography (CT) remains the central modality in disease staging [5,9,15]. Specific CT lung features include peripheral and bilateral ground-glass opacities (GGO), with round and other specific morphology as well as peripheral consolidations and increasing extension of such opacities has been associated with the risk of critical illness[26,21,4]. Whereas conventional visual scoring of the COVID-19 pneumonia extent correlates with clinical disease severity, it requires proficiency in cardiothoracic imaging and ignores lesion features such as volumes, density, or inhomogeneity[28,17]. On the other hand, CT-derived quantitative lung measures are not part of the clinical routine, despite being demonstrated to improve prognostication in COVID-19 patients, due to prohibitively time-consuming manual segmentation of the lung lesions required for computation [11,10,6]. Deep learning, a class of artificial intelligence (AI), has shown to be very effective for automated object detection and image classification from a wide range of data[8]. A variety of AI systems has been introduced to aid radiologists in the detection of lung involvement in COVID-19 with several presenting the potential to improve the performance of junior radiologists to the senior level [10,29]. However, most medical segmentation networks consume a lot of memory in storing the intermediate features for skip connections[7,19]. Considering the spatiotemporal nature of CT images, utilization of adjacent slices of input may improve performance in semantic segmentation tasks. Convolutional long short-term memory (ConvLSTM) networks have the capability of preserving relevant features with simultaneous dismission of irrelevant ones in the form of the feedback loop, which translates into a memory-sparing strategy for the holistic analysis of the images. In this paper, we employ the ConvLSTM to facilitate rapid segmentation and accurate 3D quantification of the disease involvement of lung lesions in COVID-19 pneumonia from both contrast and non-contrast CT images.

Methods

Patient population

The cohort used in this study comprised 264 patients, who underwent chest CT and had a positive RT-PCR test result for SARS-CoV-2. A total of 197 patients were included into training cohort and 67 were used for external validation. Datasets for 187 out of 197 patients from the training cohort were collected from the prospective, international, multicenter registry involving centers from North America (Cedars-Sinai Medical Center, Los Angeles, USA [n=75]), Europe (Centro Cardiologico Monzino [n = 64], and Istituto Auxologico Italiano [n = 17]; both Milan, Italy), Australia (Monash Medical Centre, Victoria, Australia [n=6]) and Asia (Showa Medical University, Tokyo, Japan [n=25]), where either non-contrast [n=157] or contrast-enhanced [n=30] chest CT was performed to aid in the triage of patients with a high clinical suspicion for COVID-19, in the setting of a pending or comorbidities associated with severe illness from COVID-19. The population is described in Table 1. Datasets for the remaining 10 COVID-19 patients were derived from an open-access repository of non-contrast CT images, therefore no clinical data were provided for this cohort. Out of 31,560 transverse slices available 15,588 had lesions Table 2. External validation cohort comprised 67 non-contrast CT scans of COVID-19 patients: 50 from an open-access repository [20] and 17 additional from Italy (Centro Cardiologico Monzino). There were 12,102 transverse slices available in this cohort and 6,503 had lesions Table 2. All data were deidentified prior to being enrolled in this study. The CT images from each patient and the clinical database were fully anonymized and transferred to one coordinating center for core lab analysis. The study was conducted with the approval of local institutional review boards (Cedars-Sinai Medical Center IRB# study 617) and written informed consent was waived for fully anonymized data analysis.
Table 1

Patient baseline characteristics and imaging data in a training cohort.

Baseline characteristicsN=187
Age, years61 ± 16
Men123 (65.7)
Body mass index26.8 ± 5.3
Current smoker22 (11.7)
Former smoker10 (5.3)
History of lung disease19 (10.1)
Image characteristicsN=197
CT scanner
Aquilion ONE73 (37.0)
GE Revolution13 (6.6)
GE Discovery CT750 HD37 (18.8)
LightSpeed VCT36 (18.3)
Brilliance iCT28 (14.2)
Unknown10 (5.1)
Non-contrast167 (84.8)
CT pulmonary angiography30 (15.2)
ECG-gated35 (17.8)

The data presented in the table are as n(%) or mean ± SD

Table 2

Dataset information.

CohortNo. of patientsNo. of lesionsNo. of lesion slices
Ground-glass opacityHigh opacity
Training19731560153756933
External ValidationCentro Cardiologico Monzino171005343961834
MosMedData [20]5020497850

Imaging Protocol

Chest CT scans were performed with different multi-slice CT systems: Aquilion ONE (Toshiba Medical Systems, Otawara, Japan); GE Revolution, GE Discovery CT750 HD, or LightSpeed VCT (GE Healthcare, Milwaukee, WI, USA); and Brilliance iCT (Philips Healthcare, Cleveland, OH, USA). Parameters used for scans without intravenous contrast included a peak x-ray tube voltage of 120 kV, automatic tube current modulation (300–500 mAs), and slice thickness of 0.625 to 1.25 mm. The protocol for contrast-enhanced included a peak x-ray tube voltage of 120 kV, automatic tube current modulation (500–650 mAs), and slice thickness of 0.625 to 1.0 mm. A total of 80–100 ml iodinated contrast material (Iomeron 400, Bracco Imaging SpA, Milan, Italy; or Ominpaque 350, GE Healthcare, United States) was injected intravenously at a rate of 5 ml/s and followed by 20–30ml of saline chaser at a rate of 4–5 ml/s. Images were reconstructed using standard lung filters specific to each CT vendor. All scans were obtained in the supine position during inspiratory breath-hold.

Ground truth generation

Images were analyzed at the Cedars-Sinai Medical Center core laboratory by two physicians (K.G. and A.L.) with 3 and 8 years of experience in chest CT, respectively, and who were blinded to clinical data. A standard lung window (width of 1500 Hounsfield units [HU] and level of −400 HU) was used. Lung abnormalities were segmented using semi-automated research software (Fusion-Quant Lung v1.0, Cedars-Sinai Medical Center, Los Angeles, CA, USA). These included ground glass opacities (GGO), consolidation, or pleural effusion according to the Fleischner Society lexicon. Consolidation and pleural effusion were collectively segmented as high-opacity to facilitate the training of the model due to a limited number of slices involving these lesions. Chronic lung abnormalities such as emphysema or fibrosis were excluded from segmentation, based on correlation with previous imaging and/or a consensus reading. GGO was defined as hazy opacities that did not obscure the underlying bronchial structures or pulmonary vessels; consolidation as opacification obscuring the underlying bronchial structures or pulmonary vessels; and pleural effusion as a fluid collection in the pleural cavity. Total pneumonia volume was calculated by summing the volumes of the GGO and consolidation components. Total pneumonia burden was calculated as: total pneumonia volume/total lung volume×100%. Difficult cases of quantitative analysis were resolved by consensus.

Data preparation for convolutional neural networks

First, the input slices were cropped to the body region and then uniformly resized to 256×256 in order to optimize the computation cost and performance Fig. 1. Further, the data was randomly augmented with the rotation of [−10, +10] degrees, translation of up to 10 pixels in x- and y-directions, and scaling of [0.9,1.05] times. Finally, the data were homogenized by clipping the Hounsfield units between −1024 to 600 (Lung region) and normalizing between 0 and 1 using Min-Max scaling.
Fig. 1

Framework for data preparation.

First, input slices were restricted to the body region and then uniformly resized to 256×256. Further, the data was randomly augmented with the rotation of [−10, +10] degrees, translation of up to 10 pixels in x- and y-directions, and scaling of [0.9, 1.05] times. Finally, data were homogenized by clipping the Hounsfield units between −1024 to 600 (Lung region) and normalizing between 0 and 1 using Min-Max scaling.

Model architecture

The model architecture was based on hierarchical multi-scale attention for semantic segmentation with the attention head looking through several adjacent slices above and below the current one to improve lesion recognition [24]. As shown in Fig. 2, the architecture was divided into two branches: the main segmentation branch and the attention branch. The main segmentation branch, consisting of a dense block followed by a segmentation head, extracted larger and easy-to-classify lesions[14]. The attention branch comprised a sequential processor: a ConvLSTM followed by a segmentation head and an attention head[23]. The ConvLSTM block allowed to imitate a radiologist reviewing adjacent slices of a CT scan to detect lung abnormalities and ensure appropriate annotation. The attention head synergized with the ConvLSTM to correct borderline misclassifications of the main segmentation branch. The segmentation head comprised three blocks: the first two blocks consisted of a 3×3 convolutional layer followed by a batch normalization layer and a LeakyRelu activation layer, while the final block was a 1×1 convolutional layer[27]. The attention head was identical to the segmentation head in structure with the only difference being followed by an additional Sigmoid layer. A single slice indented for the segmentation was used as an input by the main segmentation branch, whereas the attention branch incorporated additional 5 slices above and below the target one. Thus, the ConvLSTM block processed sequentially a total of 11 slices and conveyed the final output to the attention and segmentation head of the attention branch. The output of the attention head was then multiplied with the respective semantic feature maps of the two segmentation heads.
Fig. 2

Framework of the proposed method.

Network Training

The entire model was initialized using Kaiming He initialization, except for the dense block, where the pre-trained weights from ImageNet were used [13,16]. Region Mutual Information (RMI) loss was used as the cost function[30]. The model parameters were optimized using a Stochastic Gradient Descent optimizer with a momentum of 0.9, the initial learning rate of 103, and weight decay of 106. The learning rate was gradually decreased using Reduce on Plateau technique with a factor of 0.1 and patience of 10. This learning rate scheduler kept track of validation loss. In cases, where no improvement was seen for a “patience” number of epochs, the learning rate was reduced by the given “factor”. The training batch size was 32. The training was stopped as soon as the learning rate reached 107.

Cross-validation (repeated testing)

To perform a robust non-biased evaluation of the framework, 5-fold cross-validation was used, using 5 separate models and 5 exclusive hold-out sets each of 20%. The whole cohort of 197 cases was split into 5 subsets Fig. 3. For each fold of the 5-fold cross-validation, the following datasets were used: (1) training dataset (125 or 126 cases) was used to train the ConvLSTM; (2) validation dataset (32 cases) was defined to tune the network, select optimal hyperparameters and verify there was no over-fitting; (3) test dataset (39 or 40 cases) was used for the evaluation of the method. The final results were then concatenated from 5 separated subsets. Thus, the overall test population was 197 with 5 different models.
Fig. 3

Cross-Validation

For each fold of the 5-fold cross-validation, the following datasets were used: (1) training dataset (125 or 126 cases); (2) validation dataset (32 cases); (3) test dataset (39 or 40 cases).

Evaluation and statistical analysis

The primary end-point of this study was the performance of the deep learning method compared to the evaluation by the expert reader. Statistical analysis was performed using the SPSS software, version 23 (IBM SPSS Statistics, New York, USA). Continuous variables were expressed as mean ± standard deviation or median and interquartile range (IQR). Deep learning and expert quantifications were systematically compared using the Spearman correlation coefficient, Bland-Altman plots, and paired Wilcoxon rank-sum test. A p-value of < 0.05 was considered statistically significant. Deep learning performance was also evaluated with the Dice similarity coefficient (DSC), as a measure of overlap between expert and deep learning. This is computed by DSC = 2TP/(2TP + FP + FN), where TP (true positive) is the number of correctly positively classified voxels, FP (false positive) is the number of incorrectly positively classified, and FN (false negative) is the number of incorrectly negatively annotated classified voxels.

Implementation

The code was written in Pytorch (v1.7.1) deep learning framework and incorporated in research CT lung analysis software (Deep Lung) written in C++. The training was performed on Nvidia Titan RTX 24GB GPU on a 10th generation Intel Core i9 CPU. Deep Lung can be used with or without the GPU acceleration.

Results

Lesion quantification in the training cohort

A comparison between expert and automatic segmentation of the lung and its lesions is presented in Figs. 4 and 5. Across the 5-fold test dataset, a mean DSC of 0.876 ± 0.005 was obtained for the lesion mask differentiating between GGO and high-opacities. The median GGO volume was 330.1 mL (IQR 85.8 – 730.5 mL) and 326.1 mL (IQR 81.7 – 731.5 mL; p=0.849) for expert and automatic segmentations, respectively. An excellent correlation of 0.978 was obtained (p<0.001) and the Bland-Altman analysis demonstrated a low bias of 11.5mL Fig. 6(A–B). Similarly, no significant differences between expert and automatic segmentation were found for high-opacities volumes (6.4 mL [IQR 0 – 151.1 mL] vs 12.1 [IQR 0.3 – 88.1 mL]; p=0.658); with excellent correlation (r=0.981, p<0.001) and a non-significant bias of 19.9 mL Fig. 6(C–D).
Fig. 4

Comparison of expert and automatic lung lesion segmentation.

Blue represents ground-glass opacity. The dice score coefficient for this patient was 0.857.

Fig. 5

Comparison of expert and automatic lung lesion segmentation.

Blue represents ground-glass opacity and yellow represents consolidations. The dice score coefficient for this patient was 0.792.

Fig. 6

Bland-Altman plots and Spearman correlation for volumes of ground-glass opacity (A-B) and high-opacity (C-D) between expert and automatic quantification in a testing cohort.

Lesion quantification in the external validation cohort

A mean DSC of 0.767 ± 0.009 was obtained for the lesion mask differentiating between GGO and high-opacities in the external validation dataset using the model with the highest mean DSC. The median GGO volume was 74.9 mL (IQR 25.8 – 147.52 mL) and 90.2 mL (IQR 30.2 – 175.2 mL; p=0.849) for expert and automatic segmentations, respectively. An excellent correlation of 0.989 was obtained (p<0.001) and the Bland-Altman analysis demonstrated a low bias of 6.8mL (Supplementary Fig. i A–B). Similarly, no significant differences between expert and automatic segmentation were found for high-opacities volumes (0 mL [IQR 0 – 0 mL] vs 0 mL [IQR 0.0 – 1.3 mL]; p=0.658); with excellent correlation (r = 0.996, p<0.001) and non-significant bias of 0.2 mL (Supplementary Fig. i C–D).

Lesion quantification in non-contrast and contrast-enhanced CT

There were no significant differences between mean DSC calculated for segmentation from non-contrast (0.886 ± 0.034) and contrast-enhanced (0.829 ± 0.029; p<0.001) CT scans. The median difference between expert and automatic measurements of GGO (14.7 mL [IQR −34.2 – 23.3 mL] vs −10.3 mL [IQR −62.3 – 16.4 mL]; p=0.841) and high opacities (−11.3 mL [IQR −48.7 – 18.3 mL] vs 0.0 mL [IQR −21.3 – 0.5 mL]; p=0.667) were comparable between scanning protocols.

Computational performance

Minimal system requirements for the DeepLung are 8GB of RAM and 4GB of GRAM, thus allowing to run the segmentation in the lower-end hardware Table 3. Computations for a CT scan comprising 120 slices were performed under 2 seconds on a personal computer equipped with NVIDIA Titan RTX (24GB DDR6) GPU and under 24 seconds using the unit with Nvidia Quadro K1200 (4GB DDR5) GPU. Time required for the computations without GPU acceleration was 150 seconds using Intel Core i7–6800K (3.4GHz) Manual segmentation, depending on the case complexity, required 15–20 minutes.
Table 3

Computation times using different hardware components.

HardwareNo. of Slices
120500
GPU 24GB (Titan RTX)1.8 secs5.5 secs
GPU 4GB (Quadro K1200)23.5 secs98 secs
CPU 3.4 GHZ (i7–6800K)145 secs605 secs

Discussion

We developed and evaluated a novel deep-learning ConvLSTM network approach for fully automatic quantification of COVID-19 pneumonia burden from both non-contrast and contrast-enhanced chest CT. To the best of our knowledge, ConvLSTM has not been applied before for segmentation of medical imaging data. We demonstrated that automatic pneumonia burden quantification by the proposed method shows strong agreement with expert manual measurements and rapid performance, suitable for clinical deployment. While vaccines have been developed to protect from COVID-19, the incidental findings of COVID-19 abnormalities due to imperfect vaccination rates and new strains will be a mainstay of medical practice. This method will provide a ‘real-time’ COVID-19 disease detection to the physician and allow image-guidance of novel therapies in severe cases. The evolution of deep learning applications for COVID-19 is reflecting the changing role of CT imaging during the pandemic. Initially, when RT-PCR testing was unavailable or delayed, chest CT was used as a surrogate tool to identify suspected COVID-19 cases [2]. AI-assisted image analysis could improve the diagnostic accuracy of junior doctors in differentiating COVID-19 from other chest diseases including community-acquired pneumonia and facilitate prompt isolation of patients with suspected SARS-CoV-2 infection [29,18]. Currently, when RT-PCR testing is widely available with timely results, rapid quantification of pneumonia burden from chest CT as proposed here can be used for the management of the patients with COVID-19. As demonstrated in prior investigations, increasing attenuation of GGO and a higher proportion of consolidation in total pneumonia burden had prognostic value, thus under-scoring the importance of utilizing all CT information for training the patients [11,12]. Manual segmentation of the lung lesions is, however, challenging and prohibitively time-consuming task due to complex appearances and ambiguous boundaries of the opacities [25]. To automate the segmentation of respective lung lesions in COVID-19, several different segmentation networks have been introduced [6,29,3,22]. Whereas most of them tend to consume a lot of memory in storing the intermediate features for skip connections, it may be favorable to use several input slices in order to improve the performance in semantic segmentation tasks within data of the spatiotemporal nature [7,19]. We propose the application of ConvLSTM, presenting the potential to outperform other neural networks in capturing the spatiotemporal correlations, due to its capability of preserving relevant features with simultaneous dismission of irrelevant ones in the form of the feedback loop for the memory-sparing strategy and holistic analysis of the images [23]. It has been found that ConvLSTM localized at the input end allowed to effectively capture the global information and optimize the model performance. Automated segmentation of lung lesions with ConvLSTM networks offers a solution to generate big data with limited human resources and minimal hardware requirements. Since results of segmentation are presented to the human reader for visual inspection, eventual corrections enable the implementation of a human-in-the-loop strategy to reduce the annotation effort and provide high-volume training datasets to improve the performance of deep-learning models [25]. Furthermore, objective and repeatable quantification of pneumonia burden might aid the evaluation of the disease progression and assist the tomographic monitoring of different treatment responses. Our study had several limitations. First, different patient profiles and treatment protocols between countries may have resulted in heterogeneity in COVID-19 pneumonia severity. Second, most of the CT scans were acquired during the hospital admission, therefore availability of the slices with high-opacity (consolidations and plural effusion), representing a peak stage of the disease, was limited. Finally, training and external validation datasets comprised a relatively low number of patients manually segmented by two expert readers; however, to mitigate this we have utilized repeated testing allowing us to evaluate expected average performance of the model.

Conclusion

In conclusion, we propose and evaluate a deep learning method based on convolutional LSTM for fully automated quantification of pneumonia burden in COVID-19 patients from both non-contrast and contrast-enhanced CT datasets. The proposed method provided rapid segmentation of lung lesions with strong agreement with manual segmentation and may represent a robust tool to generate the big data with an accuracy similar to the expert readers.
  18 in total

1.  Quantitative Burden of COVID-19 Pneumonia on Chest CT Predicts Adverse Outcomes: A Post-Hoc Analysis of a Prospective International Registry.

Authors:  Kajetan Grodecki; Andrew Lin; Sebastien Cadet; Priscilla A McElhinney; Aryabod Razipour; Cato Chan; Barry Pressman; Peter Julien; Pal Maurovich-Horvat; Nicola Gaibazzi; Udit Thakur; Elisabetta Mancini; Cecilia Agalbato; Roberto Menè; Gianfranco Parati; Franco Cernigliaro; Nitesh Nerlekar; Camilla Torlasco; Gianluca Pontone; Piotr J Slomka; Damini Dey
Journal:  Radiol Cardiothorac Imaging       Date:  2020-10-01

2.  Chest CT Findings in Coronavirus Disease-19 (COVID-19): Relationship to Duration of Infection.

Authors:  Adam Bernheim; Xueyan Mei; Mingqian Huang; Yang Yang; Zahi A Fayad; Ning Zhang; Kaiyue Diao; Bin Lin; Xiqi Zhu; Kunwei Li; Shaolin Li; Hong Shan; Adam Jacobi; Michael Chung
Journal:  Radiology       Date:  2020-02-20       Impact factor: 11.105

3.  COVID-19 lung CT image segmentation using deep learning methods: U-Net versus SegNet.

Authors:  Adnan Saood; Iyad Hatem
Journal:  BMC Med Imaging       Date:  2021-02-09       Impact factor: 1.930

4.  A meta-analysis of accuracy and sensitivity of chest CT and RT-PCR in COVID-19 diagnosis.

Authors:  Fatemeh Khatami; Mohammad Saatchi; Seyed Saeed Tamehri Zadeh; Zahra Sadat Aghamir; Alireza Namazi Shabestari; Leonardo Oliveira Reis; Seyed Mohammad Kazem Aghamir
Journal:  Sci Rep       Date:  2020-12-28       Impact factor: 4.379

5.  Chest CT score in COVID-19 patients: correlation with disease severity and short-term prognosis.

Authors:  Marco Francone; Franco Iafrate; Giorgio Maria Masci; Simona Coco; Francesco Cilia; Lucia Manganaro; Valeria Panebianco; Chiara Andreoli; Maria Chiara Colaiacomo; Maria Antonella Zingaropoli; Maria Rosa Ciardi; Claudio Maria Mastroianni; Francesco Pugliese; Francesco Alessandri; Ombretta Turriziani; Paolo Ricci; Carlo Catalano
Journal:  Eur Radiol       Date:  2020-07-04       Impact factor: 5.315

6.  Systematic review with meta-analysis of the accuracy of diagnostic tests for COVID-19.

Authors:  Beatriz Böger; Mariana M Fachi; Raquel O Vilhena; Alexandre F Cobre; Fernanda S Tonin; Roberto Pontarolo
Journal:  Am J Infect Control       Date:  2020-07-10       Impact factor: 2.918

7.  Artificial Intelligence Augmentation of Radiologist Performance in Distinguishing COVID-19 from Pneumonia of Other Origin at Chest CT.

Authors:  Harrison X Bai; Robin Wang; Zeng Xiong; Ben Hsieh; Ken Chang; Kasey Halsey; Thi My Linh Tran; Ji Whae Choi; Dong-Cui Wang; Lin-Bo Shi; Ji Mei; Xiao-Long Jiang; Ian Pan; Qiu-Hua Zeng; Ping-Feng Hu; Yi-Hui Li; Fei-Xian Fu; Raymond Y Huang; Ronnie Sebro; Qi-Zhi Yu; Michael K Atalay; Wei-Hua Liao
Journal:  Radiology       Date:  2020-04-27       Impact factor: 11.105

8.  Using Artificial Intelligence to Detect COVID-19 and Community-acquired Pneumonia Based on Pulmonary CT: Evaluation of the Diagnostic Accuracy.

Authors:  Lin Li; Lixin Qin; Zeguo Xu; Youbing Yin; Xin Wang; Bin Kong; Junjie Bai; Yi Lu; Zhenghan Fang; Qi Song; Kunlin Cao; Daliang Liu; Guisheng Wang; Qizhong Xu; Xisheng Fang; Shiqin Zhang; Juan Xia; Jun Xia
Journal:  Radiology       Date:  2020-03-19       Impact factor: 11.105

9.  Epicardial adipose tissue is associated with extent of pneumonia and adverse outcomes in patients with COVID-19.

Authors:  Kajetan Grodecki; Andrew Lin; Aryabod Razipour; Sebastien Cadet; Priscilla A McElhinney; Cato Chan; Barry D Pressman; Peter Julien; Pal Maurovich-Horvat; Nicola Gaibazzi; Udit Thakur; Elisabetta Mancini; Cecilia Agalbato; Robert Menè; Gianfranco Parati; Franco Cernigliaro; Nitesh Nerlekar; Camilla Torlasco; Gianluca Pontone; Piotr J Slomka; Damini Dey
Journal:  Metabolism       Date:  2020-11-19       Impact factor: 13.934

10.  CT image visual quantitative evaluation and clinical classification of coronavirus disease (COVID-19).

Authors:  Kunwei Li; Yijie Fang; Wenjuan Li; Cunxue Pan; Peixin Qin; Yinghua Zhong; Xueguo Liu; Mingqian Huang; Yuting Liao; Shaolin Li
Journal:  Eur Radiol       Date:  2020-03-25       Impact factor: 5.315

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.