| Literature DB >> 36044215 |
Jong Seok Ahn1, Shadi Ebrahimian2,3, Shaunagh McDermott2, Sanghyup Lee1, Laura Naccarato2, John F Di Capua2, Markus Y Wu2, Eric W Zhang2, Victorine Muse2, Benjamin Miller2,4, Farid Sabzalipour2,4, Bernardo C Bizzo2,4, Keith J Dreyer2,4, Parisa Kaviani2, Subba R Digumarthy2, Mannudeep K Kalra2,4.
Abstract
Importance: The efficient and accurate interpretation of radiologic images is paramount. Objective: To evaluate whether a deep learning-based artificial intelligence (AI) engine used concurrently can improve reader performance and efficiency in interpreting chest radiograph abnormalities. Design, Setting, and Participants: This multicenter cohort study was conducted from April to November 2021 and involved radiologists, including attending radiologists, thoracic radiology fellows, and residents, who independently participated in 2 observer performance test sessions. The sessions included a reading session with AI and a session without AI, in a randomized crossover manner with a 4-week washout period in between. The AI produced a heat map and the image-level probability of the presence of the referrable lesion. The data used were collected at 2 quaternary academic hospitals in Boston, Massachusetts: Beth Israel Deaconess Medical Center (The Medical Information Mart for Intensive Care Chest X-Ray [MIMIC-CXR]) and Massachusetts General Hospital (MGH). Main Outcomes and Measures: The ground truths for the labels were created via consensual reading by 2 thoracic radiologists. Each reader documented their findings in a customized report template, in which the 4 target chest radiograph findings and the reader confidence of the presence of each finding was recorded. The time taken for reporting each chest radiograph was also recorded. Sensitivity, specificity, and area under the receiver operating characteristic curve (AUROC) were calculated for each target finding.Entities:
Mesh:
Year: 2022 PMID: 36044215 PMCID: PMC9434361 DOI: 10.1001/jamanetworkopen.2022.29289
Source DB: PubMed Journal: JAMA Netw Open ISSN: 2574-3805
Figure 1. Different Display Modes Available for the Artificial Intelligence Output
Shown are the color heat map (A), grayscale contour map (B), combined map (C), and single-color map (D).
Distribution of the Findings
| Findings | Total, No. (%) |
|---|---|
| Origin of chest radiographs (n = 497) | |
| Massachusetts General Hospital | 250 (50.3) |
| The Medical Information Mart for Intensive Care | 247 (49.7) |
| Target findings (n = 538) | |
| Nodule | 114 (21.2) |
| Pleural effusion | 149 (27.7) |
| Pneumonia | 195 (36.2) |
| Pneumothorax | 80 (14.9) |
| Nontarget findings (n = 193) | |
| Extra pleural | 7 (3.6) |
| Extra thoracic | 181 (93.8) |
| Bone fractures | 5 (2.6) |
| Target findings on chest radiographs (n = 497) | |
| 0 | 146 (29.4) |
| 1 | 198 (39.8) |
| 2 | 120 (24.2) |
| 3 | 32 (6.4) |
| 4 | 1 (0.2) |
| Chest radiographs with findings (target and nontarget) (n = 497) | |
| 0 | 105 (21.1) |
| 1 | 174 (35.1) |
| 2 | 108 (21.7) |
| ≥3 | 110 (22.1) |
Figure 2. Receiver Operating Characteristic Curves of a Deep-Learning Artificial Intelligence (AI) Algorithm for the Target Findings and Comparison Against the Reader Performance
Graphs show data for nodules (A), pleural effusions (B), pneumonia (C), and pneumothorax (D). Diagonal lines denote lines of regression. A indicates attending radiologist; F, fellow; R, resident.
Sensitivity, Specificity, and AUROC of Individual Readers
| Findings and readers | Sensitivity (95% CI) | Specificity (95% CI) | AUROC (95% CI) | |||
|---|---|---|---|---|---|---|
| Without AI | With AI | Without AI | With AI | Without AI | With AI | |
| Nodules | ||||||
| AI | NA | 0.816 (0.732-0.882) | NA | 0.731 (0.684-0.775) | NA | 0.858 (0.819-0.897) |
| Attending radiologist 1 | 0.746 (0.656-0.823) | 0.765 (0.674-0.838) | 0.859 (0.820-0.892) | 0.846 (0.806-0.881) | 0.799 (0.748-0.850) | 0.801 (0.751-0.851) |
| Attending radiologist 2 | 0.518 (0.422-0.612) | 0.486 (0.396-0.587) | 0.898 (0.863-0.927) | 0.958 (0.933-0.976) | 0.706 (0.645-0.766) | 0.723 (0.662-0.784) |
| Fellow 1 | 0.456 (0.363-0.552) | 0.640 (0.545-0.728) | 0.945 (0.917-0.966) | 0.851 (0.812-0.885) | 0.699 (0.637-0.760) | 0.773 (0.687-0.799) |
| Fellow 2 | 0.482 (0.388-0.578) | 0.632 (0.536-0.72) | 0.909 (0.875-0.936) | 0.919 (0.887-0.944) | 0.639 (0.632-0.755) | 0.773 (0.716-0.829) |
| Resident 1 | 0.640 (0.545-0.728) | 0.553 (0.457-0.646) | 0.880 (0.843-0.911) | 0.898 (0.863-0.927) | 0.757 (0.701-0.814) | 0.723 (0.664-0.782) |
| Resident 2 | 0.561 (0.465-0.654) | 0.693 (0.600-0.776) | 0.822 (0.780-0.859) | 0.812 (0.769-0.85) | 0.689 (0.630-0.749) | 0.749 (0.695-0.804) |
| Mean | 0.567 (0.524-0.611) | 0.629 (0.586-0.671) | 0.885 (0.858-0.913) | 0.881 (0.852-0.909) | 0.724 (0.700-0.748) | 0.752 (0.729-0.775) |
| Pneumonia | ||||||
| AI | NA | 0.887 (0.834-0.928) | NA | 0.728 (0.675-0.778) | NA | 0.880 (0.849-0.911) |
| Attending radiologist 1 | 0.785 (0.720-0.84) | 0.662 (0.590-0.728) | 0.864 (0.820-0.901) | 0.904 (0.865-0.935) | 0.825 (0.784-0.765) | 0.783 (0.738-0.828) |
| Attending radiologist 2 | 0.646 (0.575-0.713) | 0.662 (0.590-0.728) | 0.838 (0.791-0.877) | 0.877 (0.835-0.912) | 0.742 (0.696-0.789) | 0.770 (0.724-0.815) |
| Fellow 1 | 0.728 (0.660-0.789) | 0.856 (0.799-0.902) | 0.861 (0.817-0.898) | 0.705 (0.650-0.756) | 0.795 (0.752-0.938) | 0.783 (0.741-0.825) |
| Fellow 2 | 0.467 (0.395-0.539) | 0.641 (0.569-0.708) | 0.947 (0.915-0.969) | 0.877 (0.835-0.912) | 0.707 (0.657-0.757) | 0.759 (0.713-0.805) |
| Resident 1 | 0.703 (0.633-0.766) | 0.713 (0.644-0.775) | 0.825 (0.777-0.866) | 0.838 (0.791-0.877) | 0.764 (0.719-0.809) | 0.776 (0.731-0.820) |
| Resident 2 | 0.708 (0.638-0.770) | 0.779 (0.715-0.836) | 0.838 (0.791-0.877) | 0.791 (0.741-0.836) | 0.773 (0.728-0.817) | 0.786 (0.743-0.829) |
| Mean | 0.673 (0.632-0.714) | 0.719 (0.679-0.758) | 0.862 (0.832-0.892) | 0.832 (0.799-0.865) | 0.768 (0.749-0.786) | 0.776 (0.758-0.794) |
| Pleural effusion | ||||||
| AI | NA | 0.872 (0.808-0.921) | NA | 0.960 (0.933-0.978) | NA | 0.983 (0.974-0.992) |
| Attending radiologist 1 | 0.906 (0.847-0.948) | 0.953 (0.906-0.981) | 0.954 (0.926-0.973) | 0.931 (0.899-0.955) | 0.930 (0.900-0.960) | 0.942 (0.917-0.967) |
| Attending radiologist 2 | 0.933 (0.880-0.967) | 0.913 (0.855-0.953) | 0.899 (0.863-0.929) | 0.928 (0.896-0.953) | 0.916 (0.887-0.946) | 0.921 (0.890-0.951) |
| Fellow 1 | 0.906 (0.847-0.948) | 0.953 (0.906-0.981) | 0.885 (0.847-0.917) | 0.885 (0.847-0.917) | 0.896 (0.862-0.929) | 0.916 (0.887-0.945) |
| Fellow 2 | 0.872 (0.808-0.921) | 0.893 (0.831-0.937) | 0.948 (0.919-0.969) | 0.943 (0.913-0.965) | 0.910 (0.876-0.944) | 0.918 (0.886-0.950) |
| Resident 1 | 0.846 (0.777-0.900) | 0.799 (0.725-0.860) | 0.954 (0.926-0.973) | 0.977 (0.955-0.990) | 0.900 (0.864-0.936) | 0.888 (0.848-0.927) |
| Resident 2 | 0.872 (0.808-0.921) | 0.859 (0.793-0.911) | 0.931 (0.899-0.955) | 0.960 (0.933-0.978) | 0.903 (0.869-0.938) | 0.909 (0.875-0.944) |
| Mean | 0.889 (0.862-0.917) | 0.895 (0.868-0.922) | 0.928 (0.906-0.951) | 0.937 (0.916-0.959) | 0.909 (0.896-0.923) | 0.916 (0.902-0.929) |
| Pneumothorax | ||||||
| AI | NA | 0.988 (0.932-1.000) | NA | 0.986 (0.969-0.995) | NA | 0.999 (0.997-1.000) |
| Attending radiologist 1 | 0.975 (0.913-0.997) | 1.000 (0.955-1.000) | 0.990 (0.976-0.997) | 0.981 (0.963-0.992) | 0.977 (0.952-1.000) | 0.984 (0.968-1.000) |
| Attending radiologist 2 | 0.812 (0.710-0.891) | 0.975 (0.913-0.997) | 0.983 (0.966-0.993) | 0.990 (0.976-0.997) | 0.893 (0.841-0.945) | 0.977 (0.952-1.000) |
| Fellow 1 | 0.713 (0.600-0.808) | 0.950 (0.877-0.986) | 0.995 (0.983-0.999) | 0.990 (0.976-0.997) | 0.849 (0.788-0.910) | 0.958 (0.924-0.992) |
| Fellow 2 | 0.775 (0.668-0.861) | 0.938 (0.860-0.979) | 0.993 (0.979-0.999) | 0.990 (0.976-0.997) | 0.879 (0.823-0.935) | 0.958 (0.924-0.992) |
| Resident 1 | 0.787 (0.682-0.871) | 0.963 (0.894-0.992) | 0.988 (0.972-0.996) | 0.978 (0.959-0.99) | 0.883 (0.828-0.937) | 0.965 (0.936-0.993) |
| Resident 2 | 0.688 (0.574-0.787) | 0.963 (0.894-0.992) | 0.978 (0.959-0.99) | 0.988 (0.972-0.996) | 0.829 (0.766-0.891) | 0.969 (0.940-0.997) |
| Mean | 0.792 (0.756-0.827) | 0.965 (0.949-0.981) | 0.988 (0.978-0.997) | 0.986 (0.976-0.996) | 0.885 (0.863-0.907) | 0.969 (0.957-0.980) |
Abbreviations: AI, artificial intelligence; AUROC, area under the receiver operating characteristic curve; NA, not applicable.
Denotes metrics with statistically significant differences between the AUROCs, sensitivities, and specificities with and without AI (P < .05).
The mean values represent mean reader performance and do not include stand-alone AI performance.
Summary of Artificial Intelligence Stand-alone Performance for Detection of the 4 Target Findings in Chest Radiographs With and Without Nontarget Findings
| Target findings | Detected findings, No. | AUROC (95% CI) | Sensitivity (95% CI) | Specificity (95% CI) |
|---|---|---|---|---|
| Nodule | ||||
| Without extra findings | 70 | 0.870 (0.823-0.918) | 0.829 (0.720-0.908) | 0.740 (0.678-0.796) |
| With extra findings | 44 | 0.842 (0.778-0.906) | 0.795 (0.647-0.902) | 0.718 (0.640-0.787) |
| NA | .49 | .66 | .63 | |
| Pneumonia | ||||
| Without extra findings | 105 | 0.931 (0.899-0.963) | 0.895 (0.820-0.947) | 0.854 (0.796-0.901) |
| With extra findings | 90 | 0.777 (0.712-0.841) | 0.878 (0.792-0.937) | 0.509 (0.412-0.606) |
| NA | <.001 | .70 | <.001 | |
| Pleural effusion | ||||
| Without extra findings | 50 | 0.985 (0.971-0.999) | 0.860 (0.733-0.942) | 0.980 (0.953-0.993) |
| With extra findings | 99 | 0.975 (0.958-0.991) | 0.879 (0.798-0.936) | 0.911 (0.838-0.958) |
| NA | .34 | .75 | .003 | |
| Pneumothorax | ||||
| Without extra findings | 38 | 0.999 (0.997-1.000) | 0.974 (0.862-0.999) | 0.996 (0.979-1.000) |
| With extra findings | 42 | 0.999 (0.997-1.000) | 1.000 (0.916-1.000) | 0.968 (0.928-0.990) |
| NA | .93 | .29 | .02 | |
| Total | ||||
| Without extra findings | 297 | 0.955 (0.933-0.976) | 0.917 (0.868-0.952) | 0.857 (0.775-0.918) |
| With extra findings | 200 | 0.898 (0.838-0.958) | 0.969 (0.928-0.99) | 0.585 (0.421-0.737) |
| NA | .08 | .04 | <.001 |
Abbreviations: AUROC, area under the receiver operating characteristic curve; NA, not applicable.