Literature DB >> 32683550

From community-acquired pneumonia to COVID-19: a deep learning-based method for quantitative analysis of COVID-19 on thick-section CT scans.

Zhang Li1,2, Zheng Zhong3, Yang Li1,2, Tianyu Zhang4,5, Liangxin Gao6, Dakai Jin7, Yue Sun8, Xianghua Ye9, Li Yu10, Zheyu Hu11, Jing Xiao6, Lingyun Huang12, Yuling Tang13.   

Abstract

OBJECTIVE: To develop a fully automated AI system to quantitatively assess the disease severity and disease progression of COVID-19 using thick-section chest CT images.
METHODS: In this retrospective study, an AI system was developed to automatically segment and quantify the COVID-19-infected lung regions on thick-section chest CT images. Five hundred thirty-one CT scans from 204 COVID-19 patients were collected from one appointed COVID-19 hospital. The automatically segmented lung abnormalities were compared with manual segmentation of two experienced radiologists using the Dice coefficient on a randomly selected subset (30 CT scans). Two imaging biomarkers were automatically computed, i.e., the portion of infection (POI) and the average infection HU (iHU), to assess disease severity and disease progression. The assessments were compared with patient status of diagnosis reports and key phrases extracted from radiology reports using the area under the receiver operating characteristic curve (AUC) and Cohen's kappa, respectively.
RESULTS: The dice coefficient between the segmentation of the AI system and two experienced radiologists for the COVID-19-infected lung abnormalities was 0.74 ± 0.28 and 0.76 ± 0.29, respectively, which were close to the inter-observer agreement (0.79 ± 0.25). The computed two imaging biomarkers can distinguish between the severe and non-severe stages with an AUC of 0.97 (p value < 0.001). Very good agreement (κ = 0.8220) between the AI system and the radiologists was achieved on evaluating the changes in infection volumes.
CONCLUSIONS: A deep learning-based AI system built on the thick-section CT imaging can accurately quantify the COVID-19-associated lung abnormalities and assess the disease severity and its progressions. KEY POINTS: • A deep learning-based AI system was able to accurately segment the infected lung regions by COVID-19 using the thick-section CT scans (Dice coefficient ≥ 0.74). • The computed imaging biomarkers were able to distinguish between the non-severe and severe COVID-19 stages (area under the receiver operating characteristic curve 0.97). • The infection volume changes computed by the AI system were able to assess the COVID-19 progression (Cohen's kappa 0.8220).

Entities:  

Keywords:  Artificial intelligence; COVID-19; Deep learning; Disease progression

Mesh:

Year:  2020        PMID: 32683550      PMCID: PMC7368602          DOI: 10.1007/s00330-020-07042-x

Source DB:  PubMed          Journal:  Eur Radiol        ISSN: 0938-7994            Impact factor:   5.315


Introduction

Coronavirus disease 2019 (COVID-19) has rapidly spread all over the world since the end of 2019, and 1,436,198 cases have been confirmed as COVID-19 to date (9 April 2020) [1]. Reverse-transcription polymerase chain reaction (RT-PCR) is used as the standard diagnostic method. However, it suffers from low sensitivities as reported in [2, 3]. Computed tomography (CT) imaging is often adopted to confirm the COVID-19 in China and some European countries, e.g., Netherlands. CT plays a key role in the diagnosis and treatment assessment of COVID-19 due to its high sensitivity [2, 4]. The explosive growing number of COVID-19 patients requires the automated AI-based computer-aided diagnosis (CAD) systems that can accurately and objectively detect the disease-infected lung regions and assess the severity and the progressions. Recently, several deep learning–based AI systems were developed to differentiate the COVID-19 and community-acquired pneumonia (CAP) [5] or other viral pneumonia [6, 7], and to quantify the infection regions [8-11]. However, all these previous AI systems built upon the high-resolution thin-section CT images, which have high radiation doses and require higher costs. In contrast, the thick-section CT images from affordable CT scanners have relatively low radiation doses and are popularly used in hospitals worldwide, especially in primary care. Hence, it is worthwhile to develop an AI-based CAD system using thick-section CT images. In this study, we developed a fully automated AI system to quantify COVID-19-associated lung abnormalities and assess the disease severity and the disease progressions using thick-section chest CT images. To the best of our knowledge, this is the first AI-based study to quantitatively assess the COVID-19 severity and disease progression using the thick-section CT images.

Materials and methods

Patients

This study was approved by the Ethics of Committees of the First Hospital of Changsha, Hunan, China. Informed consent for this retrospective study was waived. Five hundred forty-eight CT scans from 204 patients diagnosed with COVID-19 (RT-PCR test positive) were retrospectively reviewed for the period from 23 January 2020 to 12 February 2020 in the First Hospital of Changsha, which is the only appointed hospital healing COVID-19 patients in Changsha city, Hunan province, China. Eight patients under 18 years old were excluded from this study. The characteristics of the rest 196 adult patients were summarized in Table 1. According to the guideline of 2019-nCoV (trial version 7) issued by the China National Health Commission [12], the severity of COVID-19 includes mild, common, severe, and critical types. Since there were few mild and critical cases, we categorized all the CT scans into the severe group (including severe and critical) and non-severe group (mild and common). In total, we had 79 severe CT scans from 32 patients, and 452 general CT scans from 164 patients. It should be noticed that some patients were in the non-severe phase when they entered the hospital, but may develop into a severe phase during treatment. All the COVID-19 patients were used to test the AI system performance.
Table 1

Characteristics of COVID-19 patients in this study

All patientsSevere patientsNon-severe patientsp value
No.19632164-
Age (year)47 ± 1556 ± 1445 ± 14< 0.001
Male96 (49%)14 (44%)82 (50%)0.52
Exams53195436-
Patients with multiple exams162 (83%)31 (97%)131 (80%)-

Note: Values in parentheses are the percentage. Ages are reported as mean ± standard deviation. COVID-19, coronavirus disease 2019

Characteristics of COVID-19 patients in this study Note: Values in parentheses are the percentage. Ages are reported as mean ± standard deviation. COVID-19, coronavirus disease 2019 To train the lung abnormalities’ segmentation deep learning model, another multi-center pneumonia dataset was collected consisting of 531 CT scans with manual annotations. The informed consent waiver of the training data was approved by the Ethics of Committees of multiple institutes.

CT protocol

All COVID-19 patients underwent CT scanning using the GE Brivo CT325 scanner (General Electric). The scanning protocol was as follows: 120 kV; adaptive tube current (30–70 mAs); pitch = 0.99–1.22 mm; slice thickness = 10 mm; field of view, 350 mm2; matrix, 512 × 512; and breath-hold at full inspiration. CT images were reconstructed with 5-mm slice thickness and the soft reconstruction kernel. The effective dose for all CT scans used in the study was 0.62 msv. Note that the radiation dose (CTDI = 3.43 mGy) from the thick-section CT imaging is reasonably lower than the conventional high-resolution chest CT imaging (Siemens SOMATOM go. Top, CTDI = 6.03 mGy, effective dose = 2.94msv). For the multi-center pneumonia dataset, the 538 CT scans were from Siemens, Hitachi, GE, Philips, and United Imaging scanners. Slice thickness ranged from 1.0 to 5.0 mm. (Details of the CT imaging protocols for this multi-center pneumonia dataset are presented in the Table of the supplemental material.)

Deep learning model for lung abnormality segmentation

We developed a 2.5D-based deep learning model to segment the pneumonia infection regions using the UNet [13] structure equipped with the Resnet 34 backbone [14]. It is able to integrate the high-resolution information into the axial view with the coarse continuity information along the vertical view. We also trained a standard 2D UNet to segment the lung fields in thick-section CT scans. The model was trained on another multi-center annotated CAP CT dataset knowing that COVID-19 shares similar abnormal lung patterns to other pneumonia such as ground glass opacity (GGO), consolidation, and bilateral infiltration. (Details of deep learning model learning are presented in the supplemental material.) To quantitatively evaluate the accuracy of segmentation on our test dataset, two radiologists with 20 and 15 years experience (Z. Z. and Y. X.), who were blind to each other, manually delineated the infection regions of interests (ROIs) to serve as the reference standard. We randomly selected 30 CT scans of 30 patients (3 severe and 27 non-severe) and quantitatively evaluated the accuracy of the infection region segmentation on this subset.

Imaging biomarker computation

Based on the lung field and infection region segmentation masks, we computed the quantitative imaging biomarkers for COVID-19, i.e., the portion of infection (POI) and the average infection HU (iHU). Specifically, we computed the POI as the infection volume divided by the total lung volume in physical unit, and the iHU as the average HU values in the infection regions. The computed POI and iHU are consistent with the latest version (the seventh) of COVID-19 diagnostic guidelines released by the National Health Commission of China [12]. The guideline states that the POI is one of the principles to differentiate the severe and non-severe patients. It also reports that lung findings in chest CT may start from small subpleural GGO to crazy paving pattern and consolidation when patients conditions getting worse, which corresponds to the increase in iHU changes. The POI and iHU were applied to distinguish the COVID-19 severity. Moreover, the changes in POI and iHU in patient’s longitudinal CT scans were calculated to evaluate the COVID-19 progression. We calculated the changes in the POI and iHU for each consecutive CT scan pair of the patients. The key phrases extracted from patients’ radiology reports were used as a ground-truth reference. The correspondence of the computed biomarker changes with radiologists’ assessment was described in Table 2. To measure the agreement between the AI-computed imaging biomarker changes and the radiologists’ assessment, we first binarize the biomarker changes. The value 1 (or 0) represented the increasing (or decreasing) of biomarkers and its corresponded phrases of radiology reports. The AI-based assessment of disease severity and progression was compared with patients’ diagnosis status extracted from clinical and radiology reports on 184 and 82 cases, respectively.
Table 2

Correspondence between the imaging biomarker changes and radiology reports

Phrase from radiology reportsImaging biomarker changesNo. of phrase occurrences
“Infection region has (partially) expanded”Increasing of POI86
“Infection region has (partially) contracted”Decreasing of POI98
“Density of infection region has (partially) increased”Increasing of iHU43
“Density of infection region has (partially) decreased”Decreasing of iHU39

Note: Opposite phrase (partially expanded and partially contracted) that exits in six patients were excluded in this table. POI, portion of infection; iHU, average infection Hounsfield unit

Correspondence between the imaging biomarker changes and radiology reports Note: Opposite phrase (partially expanded and partially contracted) that exits in six patients were excluded in this table. POI, portion of infection; iHU, average infection Hounsfield unit

Statistical analysis and evaluation metrics

Statistical analysis was performed by SAS (version 9.4) and Matlab (version 2018b). Sensitivity and specificity were calculated using specific cutoffs by using the Youden index generated from the receiver operating characteristic curve (ROC). Cohen’s kappa statistic was used to measure the agreement between the disease progress assessment from AI and radiologists. χ2 test was used to compare the differences among different groups. A two-sided p value less than 0.05 was considered to be statistically significant. The Dice coefficient was computed to evaluate the agreement between the automatic infection region segmentation and the manual infection delineations by radiologists.

Results

Segmentation of lung infection region

Examples of the infection region segmentation for severe and non-severe patients in CT were shown in Figs. 1 and 2. The average Dice coefficient between our method and two radiologists was 0.74 ± 0.28 (median = 0.79, ZZ) and 0.76 ± 0.29 (median = 0.84, YX), respectively. The inter-observer variability between the two radiologists was also assessed using the Dice coefficient, which is 0.79 ± 0.25 (media = 0.85).
Fig. 1

Lesion segmentation for three consecutive axial CTs from a severe patient. First row: original image; second row: lesion segmentation image

Fig. 2

Lesion segmentation for three consecutive axial CTs from a non-severe patient. First row: original image; second row: lesion segmentation image

Lesion segmentation for three consecutive axial CTs from a severe patient. First row: original image; second row: lesion segmentation image Lesion segmentation for three consecutive axial CTs from a non-severe patient. First row: original image; second row: lesion segmentation image

Assessment of severe and non-severe COVID-19

Based on the clinical diagnosis reports, 79 CT scans had been identified to belong to the severe group, while 452 scans were in the non-severe group. Figure 3 shows the box-plot of the computed POI and iHU for severe and non-severe groups. Note that both the POI and iHU show a significant difference between severe and non-severe groups with a p value of < 0.001.
Fig. 3

Box-plot of POI (a) and iHU (b) for the severe and non-severe patients. POI, portion of infection; iHU, average infection Hounsfield unit

Box-plot of POI (a) and iHU (b) for the severe and non-severe patients. POI, portion of infection; iHU, average infection Hounsfield unit Predictive probabilities were generated using the logistic regression model. Comparisons of different imaging biomarkers for assessment of severe and non-severe exams are shown in Table 3. Using the POI as input, the sensitivity and specificity for identifying the severe group are 92.4% and 90.5%, respectively. Using the iHU as input, the sensitivity and specificity for identifying the severe group are 91.1% and 41.6%, respectively. When combining the POI with iHU, the sensitivity and specificity for identifying the severe group are 93.7% and 88.1%, respectively. The ROC curves are shown in Fig. 4. The corresponding AUC values for using the iHU, POI, and POI + iHU are 0.69, 0.97, and 0.97, respectively. The odds of severity at 1sd increase of POI was 18.762 (95% CI, 10.056, 35.000) (p < 0.001) times higher than the baseline POI; the odds of severity at 1sd increase of iHU was 1.824 (95% CI, 1.430, 2.326) (p < 0.001) times higher than the odds of severity at baseline iHU. The Akaike information criterion (AIC) for POI, iHU, and POI + iHU is 174.877, 426.160, and 173.767, respectively.
Table 3

The performance of deep learning framework

Sensitivity %Specificity %AUCp value
POI92.41 (73 of 79) [84.85, 97.50]90.49 (409 of 452) [87.60, 92.79]0.97 [0.95, 0.98]< 0.001
iHU91.14 (72 of 79) [82.44, 96.15]41.59 (188 of 452) [36.95, 46.26]0.69 [0.63, 0.74]< 0.001
POI + iHU93.67 (74 of 79) [86.60, 97.73]88.05 (398 of 452) [85.07, 90.81]0.97 [0.94, 0.98]< 0.001

Note: Values in parentheses are the numbers for the percentage calculation. Values in brackets are 95% confidence intervals [95%CI, %]. AUC, area under the receiver operating characteristic curve; POI, portion of infection; iHU, average infection Hounsfield unit

Fig. 4

Receiver operating characteristic (ROC) curves of the model. AUC, area under the receiver operating characteristic curve; POI, portion of infection; iHU, average infection Hounsfield unit

The performance of deep learning framework Note: Values in parentheses are the numbers for the percentage calculation. Values in brackets are 95% confidence intervals [95%CI, %]. AUC, area under the receiver operating characteristic curve; POI, portion of infection; iHU, average infection Hounsfield unit Receiver operating characteristic (ROC) curves of the model. AUC, area under the receiver operating characteristic curve; POI, portion of infection; iHU, average infection Hounsfield unit

Assessment of disease progression

Figure 5 shows a qualitative example of the automatically segmented infection regions of a severe patient’s longitudinal CT scans. Cohen’s kappa statistic is shown in Table 4. The very good and moderate agreements were achieved between two AI imaging biomarkers and radiologists’ assessment if we only consider the changes on whole lung level (ignoring the cases with a phrase of “partially changes”). The change in POI showed overall better agreement (very good and good) with radiologists’ assessment than iHU (moderate and fair).
Fig. 5

The lesion segmentation of six adjacent CT scans that taken from Jan. 27 to Feb. 12 for a severe patient. The red dot corresponds to the time given for the “severe” diagnosis and the green point corresponds to the time given for the “non-severe” diagnosis

Table 4

Correspondence between the imaging biomarker changes and radiology reports

No. of casesObserved agreementCohen’s kappaKappa errorStrength of agreement
Infection region changes1420.91550.82200.0492Very good
Infection region changes (including partially changes)1840.85330.70440.0525Good
Intensity changes540.73210.46430.1184Moderate
Intensity changes (including partially changes)820.68290.37180.1018Fair

Note: Strength of agreement: 0–0.20, poor; 0.21–0.40, fair; 0.41–0.60, moderate; 0.61–0.80, good; 0.81–1.00, very good

The lesion segmentation of six adjacent CT scans that taken from Jan. 27 to Feb. 12 for a severe patient. The red dot corresponds to the time given for the “severe” diagnosis and the green point corresponds to the time given for the “non-severe” diagnosis Correspondence between the imaging biomarker changes and radiology reports Note: Strength of agreement: 0–0.20, poor; 0.21–0.40, fair; 0.41–0.60, moderate; 0.61–0.80, good; 0.81–1.00, very good

Discussion

In this study, we developed and evaluated an AI system for quantitative analysis of coronavirus disease 2019 (COVID-19) from thick-section chest CT scans. Our findings can be summarized as follows: (1) The deep learning model that trained on a multi-center CAP CT dataset could be directly applied for segmenting the lung abnormalities in COVID-19 patients; (2) the portion of infection (POI) and the average infection HU (iHU), with the area under the receiver operating characteristic curve (AUC) of 0.97 (95% CI 0.95, 0.98) and 0.69 (95% CI 0.63, 0.74), showed significant difference (p value < 0.001) in severe and non-severe COVID-19 states; (3) POI showed very good agreement (κ = 0.8220) with the radiologist reports on evaluating the changes in infection volumes on the whole lung level. Though high-resolution CT is shown to have high sensitivity in the detection of COVID-19, both cost and radiation doses are relatively high. In contrast, our study for the first time shows that an AI system can efficiently segment and quantify the COVID-19 lung infections in thick-section CT images (with relatively low radiation doses). This would benefit the developing or low-income countries, where the quantification of COVID-19 severity and the triage can be determined effectively using thick-section CT volumes of affordable CT scanners. Our diagnosis system is a multi-stage AI system. The key step is to extract infection region. It is interesting that these processing modules are trained using CAP cases while the detection and segmentation accuracy is still closed to the radiologist level. The Dice coefficients between the COVID-19-infected region segmentation of the AI system and two experienced radiologists were 0.74 ± 0.28 and 0.76 ± 0.29, respectively, which were close to the inter-observer agreement, i.e., 0.79 ± 0.25. Among our computed imaging biomarkers, only the POI shows high sensitivity and specificity for differentiating the severe from non-severe COVID-19 groups. This indicates the POI is an effective imaging biomarker to assess the severity of COVID-19 patients. Although iHU value is also able to reflect infection progress, however, it is affected by several other disease irrelevant factors, such as the reconstruction slice thickness and the respiration status [15, 16]. For instance, consolidation on HRCT images might be displayed as GGO on thick-section CT images. The changes in volume and density of infected region are two key indicators that used by radiologists for COVID-19 progression assessment. However, it is time-consuming (or even impractical) for radiologists to produce quantitative measurements for this longitudinal analysis. Our AI system provided a quantitative and objective measurement, i.e., the POI, which shows strong agreement with radiologists’ qualitative judgments. More importantly, the AI-based longitudinal disease quantification is precise, reproducible, and fast, which can reduce the reading time of radiologists for COVID-19 in each patient and improve the quality of the disease progression assessment [10]. This study has several limitations. Firstly, we only evaluated changes in imaging biomarkers at the whole lung level in certain phrase. Although our model can compute the biomarkers at the lobe level, the standard phrases from the radiology reports were mostly at the whole lung level. Furthermore, some phrases in the reports like “lesion absorption” might respond to either infection region decreasing or HU value reduction. Thus, it needs more sophisticated and precise analysis evaluating our model in the future. Secondly, motion artifacts due to respiration and heart motion may cause false-positive segmentation in the AI system. We noticed that some false-positive segmentation affected the longitudinal infection evaluations (Fig. 6). One possible solution is to identify the motion artifacts before applying the infection segmenting. Finally, our model only tested COVID-19-positive patients. A recent study has shown that a deep learning–based AI classification model can detect the COVID-19 and distinguish it from the community-acquired pneumonia and other non-pneumonic lung diseases using thin-section HRCT [5]. As the next step, it would be interesting to see if our model can also differentiate pneumonia caused by COVID-19 and other factors using the thick-section CT imaging.
Fig. 6

The false-positive segmentation from an exam with motion artifacts

The false-positive segmentation from an exam with motion artifacts In conclusion, a deep learning–based AI system is developed to quantify COVID-19 abnormal lung patterns and assess the disease severity and the progression using thick-section chest CT images. The imaging biomakers computed from the AI system could be used for reproducing several findings of infection change from the reports by radiologists. These results demonstrate that the deep learning–based tool has the ability to help radiologists on diagnosing and follow-up treatment for COVID-19 patients based on CT scans. (DOCX 521 kb)
  10 in total

1.  Influence of slice thickness and reconstruction kernel on the computed tomographic attenuation of coronary atherosclerotic plaque.

Authors:  Stephan Achenbach; Kerstin Boehmer; Tobias Pflederer; Dieter Ropers; Martin Seltmann; Michael Lell; Katharina Anders; Axel Kuettner; Michael Uder; Werner G Daniel; Mohamed Marwan
Journal:  J Cardiovasc Comput Tomogr       Date:  2010-01-21

2.  Density patterns in the normal lung as determined by computed tomography.

Authors:  L J Rosenblum; R A Mauceri; D E Wellenstein; F D Thomas; D A Bassano; B N Raasch; C C Chamberlain; E R Heitzman
Journal:  Radiology       Date:  1980-11       Impact factor: 11.105

3.  Serial Quantitative Chest CT Assessment of COVID-19: A Deep Learning Approach.

Authors:  Lu Huang; Rui Han; Tao Ai; Pengxin Yu; Han Kang; Qian Tao; Liming Xia
Journal:  Radiol Cardiothorac Imaging       Date:  2020-03-30

4.  Sensitivity of Chest CT for COVID-19: Comparison to RT-PCR.

Authors:  Yicheng Fang; Huangqi Zhang; Jicheng Xie; Minjie Lin; Lingjun Ying; Peipei Pang; Wenbin Ji
Journal:  Radiology       Date:  2020-02-19       Impact factor: 11.105

5.  Correlation of Chest CT and RT-PCR Testing for Coronavirus Disease 2019 (COVID-19) in China: A Report of 1014 Cases.

Authors:  Tao Ai; Zhenlu Yang; Hongyan Hou; Chenao Zhan; Chong Chen; Wenzhi Lv; Qian Tao; Ziyong Sun; Liming Xia
Journal:  Radiology       Date:  2020-02-26       Impact factor: 11.105

6.  Chest CT for Typical Coronavirus Disease 2019 (COVID-19) Pneumonia: Relationship to Negative RT-PCR Testing.

Authors:  Xingzhi Xie; Zheng Zhong; Wei Zhao; Chao Zheng; Fei Wang; Jun Liu
Journal:  Radiology       Date:  2020-02-12       Impact factor: 11.105

7.  Deep learning-based model for detecting 2019 novel coronavirus pneumonia on high-resolution computed tomography.

Authors:  Jun Chen; Lianlian Wu; Jun Zhang; Liang Zhang; Dexin Gong; Yilin Zhao; Qiuxiang Chen; Shulan Huang; Ming Yang; Xiao Yang; Shan Hu; Yonggui Wang; Xiao Hu; Biqing Zheng; Kuo Zhang; Huiling Wu; Zehua Dong; Youming Xu; Yijie Zhu; Xi Chen; Mengjiao Zhang; Lilei Yu; Fan Cheng; Honggang Yu
Journal:  Sci Rep       Date:  2020-11-05       Impact factor: 4.379

8.  A deep learning algorithm using CT images to screen for Corona virus disease (COVID-19).

Authors:  Shuai Wang; Bo Kang; Jinlu Ma; Xianjun Zeng; Mingming Xiao; Jia Guo; Mengjiao Cai; Jingyi Yang; Yaodong Li; Xiangfei Meng; Bo Xu
Journal:  Eur Radiol       Date:  2021-02-24       Impact factor: 5.315

  10 in total
  32 in total

Review 1.  The role of chest computed tomography in the management of COVID-19: A review of results and recommendations.

Authors:  Molly D Wong; Theresa Thai; Yuhua Li; Hong Liu
Journal:  Exp Biol Med (Maywood)       Date:  2020-06-26

2.  An Explainable AI Approach for the Rapid Diagnosis of COVID-19 Using Ensemble Learning Algorithms.

Authors:  Houwu Gong; Miye Wang; Hanxue Zhang; Md Fazla Elahe; Min Jin
Journal:  Front Public Health       Date:  2022-06-21

3.  Novel Deep Learning Technique Used in Management and Discharge of Hospitalized Patients with COVID-19 in China.

Authors:  Qingcheng Meng; Wentao Liu; Pengrui Gao; Jiaqi Zhang; Anlan Sun; Jia Ding; Hao Liu; Ziqiao Lei
Journal:  Ther Clin Risk Manag       Date:  2020-12-08       Impact factor: 2.423

4.  Quantitative evaluation of COVID-19 pneumonia severity by CT pneumonia analysis algorithm using deep learning technology and blood test results.

Authors:  Tomohisa Okuma; Shinichi Hamamoto; Tetsunori Maebayashi; Akishige Taniguchi; Kyoko Hirakawa; Shu Matsushita; Kazuki Matsushita; Katsuko Murata; Takao Manabe; Yukio Miki
Journal:  Jpn J Radiol       Date:  2021-05-14       Impact factor: 2.374

5.  Six artificial intelligence paradigms for tissue characterisation and classification of non-COVID-19 pneumonia against COVID-19 pneumonia in computed tomography lungs.

Authors:  Luca Saba; Mohit Agarwal; Anubhav Patrick; Anudeep Puvvula; Suneet K Gupta; Alessandro Carriero; John R Laird; George D Kitas; Amer M Johri; Antonella Balestrieri; Zeno Falaschi; Alessio Paschè; Vijay Viswanathan; Ayman El-Baz; Iqbal Alam; Abhinav Jain; Subbaram Naidu; Ronald Oberleitner; Narendra N Khanna; Arindam Bit; Mostafa Fatemi; Azra Alizad; Jasjit S Suri
Journal:  Int J Comput Assist Radiol Surg       Date:  2021-02-03       Impact factor: 3.421

6.  Quantification of COVID-19 Opacities on Chest CT - Evaluation of a Fully Automatic AI-approach to Noninvasively Differentiate Critical Versus Noncritical Patients.

Authors:  Christoph Mader; Simon Bernatz; Sabine Michalik; Vitali Koch; Simon S Martin; Scherwin Mahmoudi; Lajos Basten; Leon D Grünewald; Andreas Bucher; Moritz H Albrecht; Thomas J Vogl; Christian Booz
Journal:  Acad Radiol       Date:  2021-03-06       Impact factor: 3.173

7.  What Can COVID-19 Teach Us about Using AI in Pandemics?

Authors:  Krzysztof Laudanski; Gregory Shea; Matthew DiMeglio; Mariana Rastrepo; Cassie Solomon
Journal:  Healthcare (Basel)       Date:  2020-12-01

Review 8.  Medical imaging and computational image analysis in COVID-19 diagnosis: A review.

Authors:  Shahabedin Nabavi; Azar Ejmalian; Mohsen Ebrahimi Moghaddam; Ahmad Ali Abin; Alejandro F Frangi; Mohammad Mohammadi; Hamidreza Saligheh Rad
Journal:  Comput Biol Med       Date:  2021-06-23       Impact factor: 6.698

9.  A novel CT-based radiomics in the distinction of severity of coronavirus disease 2019 (COVID-19) pneumonia.

Authors:  Zongyu Xie; Haitao Sun; Jian Wang; Chunhong Hu; Weiqun Ao; He Xu; Shuhua Li; Cancan Zhao; Yuqing Gao; Xiaolei Wang; Tongtong Zhao; Shaofeng Duan
Journal:  BMC Infect Dis       Date:  2021-06-25       Impact factor: 3.090

Review 10.  COVID-19 pathways for brain and heart injury in comorbidity patients: A role of medical imaging and artificial intelligence-based COVID severity classification: A review.

Authors:  Jasjit S Suri; Anudeep Puvvula; Mainak Biswas; Misha Majhail; Luca Saba; Gavino Faa; Inder M Singh; Ronald Oberleitner; Monika Turk; Paramjit S Chadha; Amer M Johri; J Miguel Sanches; Narendra N Khanna; Klaudija Viskovic; Sophie Mavrogeni; John R Laird; Gyan Pareek; Martin Miner; David W Sobel; Antonella Balestrieri; Petros P Sfikakis; George Tsoulfas; Athanasios Protogerou; Durga Prasanna Misra; Vikas Agarwal; George D Kitas; Puneet Ahluwalia; Raghu Kolluri; Jagjit Teji; Mustafa Al Maini; Ann Agbakoba; Surinder K Dhanjil; Meyypan Sockalingam; Ajit Saxena; Andrew Nicolaides; Aditya Sharma; Vijay Rathore; Janet N A Ajuluchukwu; Mostafa Fatemi; Azra Alizad; Vijay Viswanathan; Pudukode R Krishnan; Subbaram Naidu
Journal:  Comput Biol Med       Date:  2020-08-14       Impact factor: 4.589

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.