Literature DB >> 33828217

Diagnostic accuracy of deep learning in medical imaging: a systematic review and meta-analysis.

Ravi Aggarwal1, Viknesh Sounderajah1, Guy Martin1, Daniel S W Ting2, Alan Karthikesalingam1, Dominic King1, Hutan Ashrafian3, Ara Darzi1.   

Abstract

Deep learning (DL) has the potential to transform medical diagnostics. However, the diagnostic accuracy of DL is uncertain. Our aim was to evaluate the diagnostic accuracy of DL algorithms to identify pathology in medical imaging. Searches were conducted in Medline and EMBASE up to January 2020. We identified 11,921 studies, of which 503 were included in the systematic review. Eighty-two studies in ophthalmology, 82 in breast disease and 115 in respiratory disease were included for meta-analysis. Two hundred twenty-four studies in other specialities were included for qualitative review. Peer-reviewed studies that reported on the diagnostic accuracy of DL algorithms to identify pathology using medical imaging were included. Primary outcomes were measures of diagnostic accuracy, study design and reporting standards in the literature. Estimates were pooled using random-effects meta-analysis. In ophthalmology, AUC's ranged between 0.933 and 1 for diagnosing diabetic retinopathy, age-related macular degeneration and glaucoma on retinal fundus photographs and optical coherence tomography. In respiratory imaging, AUC's ranged between 0.864 and 0.937 for diagnosing lung nodules or lung cancer on chest X-ray or CT scan. For breast imaging, AUC's ranged between 0.868 and 0.909 for diagnosing breast cancer on mammogram, ultrasound, MRI and digital breast tomosynthesis. Heterogeneity was high between studies and extensive variation in methodology, terminology and outcome measures was noted. This can lead to an overestimation of the diagnostic accuracy of DL algorithms on medical imaging. There is an immediate need for the development of artificial intelligence-specific EQUATOR guidelines, particularly STARD, in order to provide guidance around key issues in this field.

Entities:  

Year:  2021        PMID: 33828217      PMCID: PMC8027892          DOI: 10.1038/s41746-021-00438-z

Source DB:  PubMed          Journal:  NPJ Digit Med        ISSN: 2398-6352


Introduction

Artificial Intelligence (AI), and its subfield of deep learning (DL)[1], offers the prospect of descriptive, predictive and prescriptive analysis, in order to attain insight that would otherwise be untenable through manual analyses[2]. DL-based algorithms, using architectures such as convolutional neural networks (CNNs), are distinct from traditional machine learning approaches. They are distinguished by their ability to learn complex representations in order to improve pattern recognition from raw data, rather than requiring human engineering and domain expertise to structure data and design feature extractors[3]. Of all avenues through which DL may be applied to healthcare; medical imaging, part of the wider remit of diagnostics, is seen as the largest and most promising field[4,5]. Currently, radiological investigations, regardless of modality, require interpretation by a human radiologist in order to attain a diagnosis in a timely fashion. With increasing demands upon existing radiologists (especially in low-to-middle-income countries)[6-8], there is a growing need for diagnosis automation. This is an issue that DL is able to address[9]. Successful integration of DL technology into routine clinical practice relies upon achieving diagnostic accuracy that is non-inferior to healthcare professionals. In addition, it must provide other benefits, such as speed, efficiency, cost, bolstering accessibility and the maintenance of ethical conduct. Although regulatory approval has already been granted by the Food and Drug Administration for select DL-powered diagnostic software to be used in clinical practice[10,11], many note that the critical appraisal and independent evaluation of these technologies are still in their infancy[12]. Even within seminal studies in the field, there remains wide variation in design, methodology and reporting that limits the generalisability and applicability of their findings[13]. Moreover, it is noted that there has been no overarching medical specialty-specific meta-analysis assessing diagnostic accuracy of DL performance, particularly in ophthalmology, respiratory medicine and breast surgery, which have the most diagnostic studies to date[13]. Therefore, the aim of this review is to (1) quantify the diagnostic accuracy of DL in speciality-specific radiological imaging modalities to identify or classify disease, and (2) to appraise the variation in methodology and reporting of DL-based radiological diagnosis, in order to highlight the most common flaws that are pervasive across the field.

Results

Search and study selection

Our search identified 11,921 abstracts, of which 9484 were screened after duplicates were removed. Of these, 8721 did not fulfil inclusion criteria based on title and abstract. Seven hundred sixty-three full manuscripts were individually assessed and 260 were excluded at this step. Five hundred three papers fulfilled inclusion criteria for the systematic review and contained data required for sensitivity, specificity or AUC. Two hundred seventy-three studies were included for meta-analysis, 82 in ophthalmology, 115 in respiratory medicine and 82 in breast cancer (see Fig. 1). These three fields were chosen to meta-analyse as they had the largest numbers of studies with available data. Two hundred twenty-four other studies were included for qualitative synthesis in other medical specialities. Summary estimates of imaging and speciality-specific diagnostic accuracy metrics are described in Table 1. Units of analysis for each speciality and modality are indicated in Tables 2–4.
Fig. 1

PRISMA flow diagram of included studies.

PRISMA (preferred reporting items for systematic reviews and meta-analyses) flow diagram of included studies.

Table 1

Summary estimates of pooled speciality and imaging modality specific diagnostic accuracy metrics.

Imaging modalityDiagnosisAUC95% CII2Sensitivity95% CII2Specificity95% CII2PPV95% CII2NPV95% CII2Accuracy95% CII2F1 score95% CII2
Ophthalmology imaging
RFPDR0.9390.920–0.95899.90.9760.975–0.97799.90.9020.889–0.91699.70.3890.166–0.61299.71190.60.9270.899–0.95596.3
RFPAMD0.9630.948–0.97999.30.9730.971–0.97499.90.9240.896–0.95299.60.7970.719–0.87599.9
RFPGlaucoma0.9330.924–0.94299.60.8830.862–0.90499.90.9180.898–0.93899.70.8810.847–0.91597.7
RFPROP0.960.913–1.00899.50.9070.749–1.06699.8
OCTDR10.999–1.098.10.9540.937–0.97298.90.9930.991–0.99498.20.970.959–0.98197.5
OCTAMD0.9690.955–0.98399.40.9970.996–0.99799.70.9320.914–0.95098.90.9360.906–0.96599.6
OCTGlaucoma0.9640.941–0.98677.7
Respiratory imaging
CTLung nodules0.9370.924–0.949970.860.831–0.89099.70.8960.871–0.92199.20.7850.711–0.85899.20.8890.870–0.90898.40.790.747–0.83497.9
CTLung cancer0.8870.847–0.92895.90.8370.780–0.89494.60.8260.735–0.91898.10.8270.784–0.87081.7
X-rayNodules0.8840.842–0.92599.60.750.634–0.866990.9440.912–0.97698.40.860.736–0.98499.80.8940.842–0.94581.4
X-rayMass0.8640.827–0.90199.70.8010.683–0.91999.7
X-rayAbnormal0.9170.869–0.96699.90.8730.762–0.98599.90.8940.860–0.92998.70.850.567–1.1331000.8590.736–0.983990.760.558–0.96299.7
X-rayAtelectasis0.8240.783–0.86699.7
X-rayCardiomegaly0.9050.871–0.93899.7
X-rayConsolidation0.8750.800–0.94999.90.9140.816–1.01399.50.7510.637–0.86698.60.8970.828–0.96696.4
X-rayPulmonary oedema0.8930.843–0.94499.9
X-rayEffusion0.9060.862–0.95099.8
X-rayEmphysema0.8850.855–0.91699.7
X-rayFibrosis0.8340.796–0.87299.7
X-rayHiatus hernia0.8940.858–0.93099.8
X-rayInfiltration0.7240.682–0.76799.6
X-rayPleural thickening0.8160.762–0.87099.8
X-rayPneumonia0.8450.782–0.90799.90.9510.936–0.96596.30.7160.480–0.9531000.6810.367–0.9951000.7630.559–0.9681000.8890.838–0.94197.6
X-rayPneumothorax0.910.863–0.95799.90.7180.433–1.0041000.9180.870–0.96599.90.4960.369–0.623100
X-rayTuberculosis0.9790.978–0.98199.60.9980.997–0.99999.610.999–1.00095.30.940.921–0.95984.6
Breast imaging
MMGBreast cancer0.8730.853–0.89498.80.8510.779–0.92399.90.8820.859–0.90597.20.9050.880–0.93097.9
UltrasoundBreast cancer0.9090.881–0.93691.70.8530.815–0.89193.90.9010.870–0.93196.60.8040.727–0.88093.70.9220.851–0.99297.20.8730.841–0.90687.50.8550.803–0.90687.9
MRIBreast cancer0.8680.850–0.88627.80.7860.710–0.86180.50.7880.697–0.88086.2
DBTBreast cancer0.9080.880–0.93763.20.8310.675–0.98897.60.9180.905–0.9300
Table 2

Characteristics of ophthalmic imaging studies.

StudyModelProspective?Test setPopulationTest datasetsType of internal validationExternal validationReference standardAI vs clinician?Imaging modalityPathology
Abramoff et al. 2016AlexNet/VGGNo1748PhotographsMessidor-2NRNoExpert consensusNoRetinal fundus photographyReferable DR
Abramoff et al. 2018[14]AlexNet/VGGYes819PatientsProspective cohort from 10 primary care practice sites in USANRYesExpert consensusNoRetinal fundus photographyMore than mild DR
Ahn et al. 2018(a) Inception-v3; (b) customised CNNNo(a) 464; (b) 464ImagesKim’s Eye Hospital, KoreaRandom splitNoExpert consensusNoRetinal fundus photographyEarly and advanced glaucoma
Ahn et al. 2019ResNet50No219PhotographsKim’s Eye Hospital, KoreaRandom splitNoExpert consensusNoRetinal fundus photographsPseudopapilloedema
Al-Aswad et al. 2019[46]Pegasus (ResNet50)No110PhotographsSingapore Malay Eye StudyRandom splitNoExisting diagnosis from source dataYesRetinal fundus photographsGlaucoma
Alqudah et al. 2019[22]AOCT-NETNo1250ScansFarsiu Ophthalmology 2013Hold-out methodYesNRNoOCT(a) AMD; (b) DME
Arcadu et al. 2019Inception-v3No(a) 1237; (b) 1798ImagesRISE/RIDE trialsRandom splitNoExpert consensusNoRetinal fundus photography(a) DME—central subfield thickness >400 µm; (b) DME—central fovea thickness >400 µm
Asaoka et al. 2016Deep feed-forward neural network with stacked denoising autoencoderNo279EyesUniversity of Tokyo Hospital, TokyoRandom splitNoOther imaging techniqueNoVisual FieldsPreperimetric open-angle glaucoma
Asaoka et al. 2019Customised CNNNo196ImagesUniversity of Tokyo Hospital, TokyoRandom splitNoExpert consensusNoOCTEarly open-angle glaucoma
Asaoka et al. 2019[23]ResNet50No(a) 205; (b) 171Scans(a) Iinan Hospital; (b) Hiroshiuma UniversityNRYesExpert consensusNoOCTGlaucoma
Bellemo et al. 2019[15]VGG/ResNetYes3093EyesKitwe Central Hospital Eye Unit, ZambiaNAYesExpert consensusNoRetinal fundus photography(a) Referable DR; (b) vision-threatening DR; (c) DME
Bhatia et al. 2019[24]VGG-16No(a) 4686; (b) 384; (c) 148; (d) 100; (e) 135; (f) 135; (g) 148; (h) 100Scans(a) Shiley Eye Institute of the UCDS; (b) Devers Eye Institute; (c) Noor Eye Hospital; (d) Ophthalmica Ophthalmology Greece; (e) Cardiff University; (f) Cardiff University; (g) Noor Eye Hospital; (h) Ophthalmica Ophthalmology GreeceNAYes(a) Expert consensus; (b) NR; (c) NR; (d) NR; (e) Expert consensus + further imaging; (f) expert consensus + further imaging; (g) NR; (h) NRNoOCT(a) Abnormal scan; (b–f) AMD; (g–h) DME
Brown et al. 2018[47]Inception-v1 and U-NetNo100Photographsi-ROPHold-out methodNoExpert consensusYesRetinal fundus photographyPlus disease in ROP
Burlina et al. 2017[49]DCNNNo5664ImagesAREDS 4 datasetNRNoExpert consensusYesRetinal fundus photographyAMD-AREDS 4 step
Burlina et al. 2018[48]ResNet50No5000ImagesAREDSRandom splitNoReading centre graderNoRetinal fundus photographsReferable AMD
Burlina et al. 2018[50]AlexNetNo13,480PhotographsNIH AREDSNRNoReading centre graderYesRetinal fundus photographyReferable AMD
Burlina et al. 2018ResNet50No(a) 6654; (b) 58,978Images(a) AREDS 9 dataset; (b) AREDS 4 datasetNRNoReading centre graderYesRetinal fundus photography(a) AMD-AREDS 4 step; (b) AMD-AREDS 9 step
Chan et al. 2018[25]AlexNet, VGGNet, GoogleNetNo4096ImagesSERINRYesReading centre graderNoOCTDME
Choi et al. 2017VGG-19No(a) 3000; (b) 3000PhotographsSTARE databaseRandom splitNoExpert consensusNoRetinal fundus photographs(a) DR; (b) AMD
Christopher et al. 2018[16](a) VGG-16; (b) Inception-v3; (c) ResNet50Yes1482ImagesADAGES and DIGSRandom splitNoExpert consensusNoRetinal fundus photographyGlaucomatous optic neuropathy
Das et al. 2019VGG-16No1000ImagesUCSDHold-out methodNoExpert consensusNoOCTDME
De Fauw et al. 2018[51](a) U-Net (b) customised CNNNo(a) 997; (b) 116(a) Scans (Topcon device); (b) scans (Spectralis device)Moorfields, LondonRandom splitNoFollow upYesOCTUrgent referral eye disease
ElTanboly et al. 2016Deep fusion classification network (DFCN)No12OCT scansHold-out methodNoNRNoOCTEarly DR
Gargeya et al. 2017[26]CNNNo(a) 15,000 (b) 1748; (c) 463Photographs(a) EyePACS-1; (b) Messidor-2; (c) E-OpthmaRandom splitYesExpert consensusNoRetinal fundus photographyDR
Gomez-Valverde et al. 2019[52]VGG-19No494PhotographsESPERANZARandom splitNoExpert consensusYesRetinal fundus photographsGlaucoma suspect or glaucoma
Grassman et al. 2018[27]Ensemble:random forestNo(a) 12,019; (b) 5555Images(a) AREDS dataset; (b) KORA datasetRandom splitYesReading centre graderNoRetinal fundus photographyAMD-AREDS 9 step
Gulshan et al. 2019[17]Inception-v3Yes3049PhotographsProspectiveNAYesExpert consensusYesRetinal fundus photographsReferable DR
Gulshan et al. 2016[28]Inception-v3No(a) 8788; (b) 1745Photographs(a) EyePACS-1; (b) Messidor-2Random splitYesReading centre graderYesRetinal fundus photographyReferable DR
Hwang et al. 2019[29](a) ResNet50; (b) VGG-16; (c) Inception-v3; (d) ResNet50; (e) VGG-16; (f) Inception-v3No(a–c) 3872; (d–f) 750Images(a–c) Department of Ophthalmology of Taipei Veterans General Hospital; (d–f) External validationRandom splitYesExpert consensusYesOCTAMD-AREDS 4 step
Jammal et al. 2019[53]ResNet34No490ImagesRandomly drawn from test sampleNoReading centre graderYesRetinal fundus photographsGlaucomatous optic neuropathy
Kanagasingham et al. 2018[21]DCNNYes398PatientsPrimary Care Practice, Midland, Western AustraliaNAYesReading centre graderNoRetinal fundus photographyReferable DR
Karri et al. 2017GoogLeNetNo21ScansDuke UniversityRandom splitNoNRNoOCT(a) DME; (b) dry AMD
Keel et al. 2018[18]Inception-v3Yes93ImagesSt Vincent’s Hospital Melbourne and University Hospital Geelong, Barwon HealthNAYesReading centre graderNoRetinal fundus photographyReferable DR
Keel et al. 2019[30]CNNNo86,202PhotographsMelbourne Collaborative Cohort StudyHold-out methodYesExpert consensusNoRetinal fundus photographsNeovascular AMD
Kermany et al. 2018[54]Inception-v3No(a) 1000; (b–d) 500ScansShiley Eye Institute of the University of California San Diego, the California Retinal Research Foundation, Medical Centre Ophthalmology Associates, the Shanghai First People’s Hospital, and Beijing Tongren Eye CentreRandom splitNoConsensus involving experts and non-expertsYesOCT(a) Choroidal neovascularisation vs DME vs drusen vs normal; (b) choroidal neovascularisation; (c) DME; (d) AMD
Krause et al. 2018[31]CNNNo1958ImagesEyePACS-2Hold-out methodYesExpert consensusNoRetinal fundus photographsReferable DR
Lee et al. 2017VGG-16No2151ScansRandom splitNoRoutine clinical notesNoOCTAMD
Lee et al. 2019CNNNo200PhotographsSeoul National University HospitalHold-out methodNoOther imaging techniqueNoRetinal fundus photographsGlaucoma
Li et al. 2018[108]Inception-v3No8000ScansGuangdong (China)Random splitNoExpert gradersNoRetinal fundus photographyGlaucomatous optic neuropathy
Li et al. 2019[55]VGG-16No1000ImagesShiley Eye Institute of the University of California San Diego, the California Retinal Research Foundation, Medical Centre Ophthalmology Associates, the Shanghai First People’s Hospital, and Beijing Tongren Eye CentreRandom splitNoExpert consensusNoOCTChoroidal neovascularisation vs DME vs drusen Vs normal
Li et al. 2019OCT-NETNo859ScansWenzhou Medical UniversityRandom splitNoExpert gradersNoOCTEarly DR
Li et al. 2019[33]Inception-v3No800ImagesMessidor-2Random splitYesReading centre graderNoRetinal fundus photographsReferable DR
Li et al. 2019ResNet50No1635ImagesShanghai Zhongshan Hospital and the Shanghai First People’s HospitalRandom splitNoReading centre graderYesOCTDME
Lin et al. 2019[109]CC-CruiserYes—multicentre RCT350ImagesMulticentre RCTNANAExpert consensusYesSlit-lamp photographyChildhood cataracts
Li F et al. 2018VGG-15No300ImagesNRRandom splitNoNRNoVisual FieldsGlaucoma
Li Z et al. 2018[33]CNNNo35,201PhotographsNIEHS, SiMES, AusDiabRandom splitYesReading centre graderNoRetinal fundus photographsReferable DR
Liu et al. 2018[35]ResNet50No(a) 754; (b) 30Photographs(a) NR; (b) HRFRandom splitYesReading centre graderYesRetinal fundus photographsGlaucomatous optic discs
Liu et al. 2019[34]CNNNo(a) 28,569; (b) 20,466; (c) 12,718; (d) 9305; (e) 29,676; (f) 7877Photographs(a) Local Validation (Chinese Glaucoma Study Alliance); (b) Beijing Tongren Hospital; (c) Peking University Third Hospital; (d) Harbin Medical University First Hospital; (e) Handan Eye Study; (f) Hamilton Glaucoma CentreRandom splitYesConsensus involving experts and non-expertsNoRetinal fundus photographsGlaucomatous optic neuropathy
Long et al. 2017[56]DCNNNo57ImagesMultihospital clinical trialHold-out methodNoExpert consensusYesOcular imagesCongenital Cataracts
MacCormick et al. 2019[36]DenseNetNo(a) 130; (b) 159Images(a) ORIGA; (b) RIM-ONERandom splitYes(a) NR; (b) expert consensusNoRetinal fundus photographyGlaucomatous optic discs
Maetshke et al. 20193D CNNNo110OCT scansFivefold cross validationRandom splitNoFollow upNoOCTGlaucomatous optic neuropathy
Matsuba et al. 2018[57]DCNNNo111ImagesTsukazaki HospitalNRNoExpert consensus + further imagingYesRetinal fundus photography (optos)Exudative AMD
Medeiros et al. 2019ResNet34No6292ImagesDuke UniversityRandom splitNoFollow upNoRetinal fundus photographyGlaucomatous optic neuropathy
Motozawa et al. 2019CNNNo382ImagesKobe City Medical CentreRandom splitNoRoutine clinical notesNoOCTAMD
Muhammad et al. 2017AlexNetNo102ImagesNRNRNoExpert consensusNoOCTGlaucoma suspect or glaucoma
Nagasato et al. 2019VGG-16No466ImagesNRK-fold cross validationNoNRNoRetinal fundus photography (optos)Retinal vein occlusion
Nagasato et al. 2019[58]DNNNo322ScansTsukazaki Hospital and Tokushima University HospitalK-fold cross validationNoExpert gradersYesOCTRetinal vein occlusion
Nagasawa et al. 2019VGG-16No378ImagesTsukazaki Hospital and Tokushima University HospitalK-fold cross validationNoExpert gradersNoRetinal fundus photography (optos)Proliferative diabetic retinopathy
Ohsugi et al. 2017DCNNNo166ImagesTsukazaki HospitalRandom splitNoExpert consensusNoRetinal fundus photography (optos)Rhegmatogenous retinal detachment
Peng et al. 2019[59]Inception-v3No900ImagesAREDSRandom splitNoReading centre graderYesRetinal fundus photographyAge-related macular degeneration-AREDS 4 step
Perdomo et al. 2019OCT-NETNo2816ImagesSERI-CUHK data setRandom splitNoExpert gradersNoOCTDME
Phan et al. 2019DenseNet201No828ImagesYamanashi Koseiren HospitalNoExpert consensus + further imagingNoRetinal fundus photographyGlaucoma
Phene et al. 2019[37]Inception-v3No(a) 1205; (b) 9642; (c) 346Images(a) EyePACS, Inoveon, the United Kingdom Biobank, the Age-Related Eye Disease Study, and Sankara Nethralaya; (b) Atlanta Veterans Affairs (VA) Eye Clinic; (c) Dr. Shroff’s Charity Eye Hospital, New Delhi, IndiaRandom splitYesReading centre graderYesRetinal fundus photographsGlaucomatous optic neuropathy
Prahs et al. 2017GoogLeNetNo5358ImagesHeidelberg Eye Explorer, Heidelberg EngineeringRandom splitNoExpert gradersNoOCTInjection vs No injection for AMD
Raju et al. 2017CNNNo53,126ImagesEyePACS-1Random splitNoNRNoRetinal fundus photographyReferable DR
Ramachandran et al. 2018[38]Visiona intelligent diabetic retinopathy screening platformNo(a) 485; (b) 1200Photographs(a) ODEMS; (b) MessidorNAYesExpert gradersNoRetinal fundus photographsReferable DR
Raumviboonsuk et al. 2019[39]Inception-v4No(a–c) 25,348; (d) 24,332ImagesNational screening program for DR in ThailandNAYesExpert consensusYesRetinal fundus photography(a) Moderate non-proliferative DR or worse; (b) severe non-proliferative DR or worse; (c) proliferative DR; (d) referable DME
Redd et al. 2018Inception-v1 and U-NetNo4861ImagesMulticentre i-ROP studyNRNoExpert graders + further imagingNoRetinal fundus photographyPlus disease in ROP
Rogers et al. 2019[45]Pegasus (ResNet50)No94PhotographsEODATNAYesReading centre graderYesRetinal fundus photographsGlaucomatous optic neuropathy
Sandhu et al. 2018[19]Deep fusion SNCAEYes160ScansUniversity of WaikatoNANoClinical diagnosisNoRetinal fundus photographsNon-proliferative DR
Sayres et al. 2019[40]Inception-v4No2000ImagesEyePACS-2NAYesExpert consensusYesRetinal fundus photographsReferable DR
Shibata et al. 2018[60](a) ResNet; (b) VGG-16No110ImagesMatsue Red Cross HospitalRandom splitNoExpert consensusYesRetinal fundus photographyGlaucoma
Stevenson et al. 2019Inception-v3No(a) 2333; (b) 2283; (c) 2105PhotographsPublicly available databasesRandom splitNoExisting diagnosis from source dataNoRetinal fundus photographs(a) Glaucoma; (b) DR; (c) AMD
Ting et al. 2017[41]VGGNetNo(a) 71,896; (b) 15,798; (c) 3052; (d) 4512; (e) 1936; (f) 1052; (g) 1968; (h) 2302; (i) 1172; (j) 1254; (k) 7706; (l) 35,948; (m) 35,948Images(a) Singapore National Diabetic Retinopathy Screening Program 2014–2015; (b) Guangdong (China); (c) Singapore Malay Eye Study; (d) Singapore Indian Eye Study; (e) Singapore Chinese Eye Study; (f) Beijing Eye Study; (g) African American Eye Disease Study; (h) Royal Victoria Eye and Ear Hospital; (i) Mexican; (j) Chinese University of Hong Kong, (k, l) Singapore National Diabetic Retinopathy Screening Program 2014–2015Random splitYesExpert consensusNoRetinal fundus photographyReferable DR
Ting et al. 2019[42]VGGNetNo85,902ImagesCombined eight datasetsNAYesConsensus involving experts and non-expertsNoRetinal fundus photography(a) Any DR; (b) referable DR; (c) vision-threatening DR
Treder et al. 2017Inception-v3No100ScansNRHold-out methodNoNRNoOCTExudative AMD
van Grinsven et al. 2016[44](a) Ses CNN 60; (b) NSesCNN170No1200ImagesMessidorRandom splitYesExisting diagnosis from source dataYesRetinal fundus photographsRetinal haemorrhage
Verbraak et al. 2019[43]AlexNet/VGGNo1293ImagesNetherlands Star-SHLNAYesExpert consensusNoRetinal fundus photography(a) DR-vision-threatening; (b) DR- more than mild
Xu et al. 2017CNNNo200PhotographsKaggleRandom splitNoExisting diagnosis from source dataNoRetinal fundus photographsDR
Yang et al. 2019VGGNetNo500PhotographsIntelligent Ophthalmology Database of Zhejiang Society for Mathematical Medicine in ChinaHold-out methodNoExpert consensusNoRetinal fundus photographsReferable DR
Yoo et al. 2019VGG-19No900ScansProject MaculaRandom splitNoNRNo(a) OCT; (b) retinal fundus photographsAMD
Zhang et al. 2019[61]VGG-16No1742ImagesTelemed-R screeningRandom splitNoExpert consensusYesRetinal fundus photographsROP
Zheng et al. 2019[20]Inception-v3Yes102ScansJoint Shantou International Eye Centre of Shantou University and the Chinese University of Hong Kong (JSIEC)Hold-out methodNoNRNoOCTGlaucomatous optic neuropathy
Table 4

Characteristics of breast imaging studies.

StudyModelProspective?Test SetPopulationTest datasetsType of internal validationExternal validationReference standardAI vs clinician?Imaging modalityBody system/disease
Abdelsamea et al. 2019CNNNo118ImagesNRTenfold cross validationNoNRNoMammogramBreast cancer
Agnes et al. 2020Multiscale all CNNNo322Imagesmini-MIASRandom splitNoExisting labels from datasetNoMammogramBreast cancer
Akselrod-Ballin et al. 2017Faster R-CNNNo170ImagesMulticentre hospital data setRandom splitNoExpert readerNoMammogramBreast cancer
Al-Antari et al. 2018YOLONo410ImagesINbreastRandom splitNoExpert reader, histology,NoMammogramBreast cancer
Al-Antari et al. 2018DBNNo150ImagesDDSMRandom splitNoFollow up, histology, expert readerNoMammogramBreast cancer
Al-Masni et al. 2018YOLONo120ImagesDDSMRandom splitNoFollow up, histology, expert readerNoMammogramBreast cancer
Antropova et al. 2017VGG-19No(a) 690; (b) 245; (c) 1125(a) Lesions; (b) images; (c) lesionsPrivateRandom splitNoHistologyNo(a) MRI; (b) mammogram; (c) ultrasoundBreast cancer
Antropova et al. 2018VGGNetNo138LesionsUniversity of ChicagoRandom splitNoHistologyNoMRIBreast cancer
Antropova et al. 2018VGGNetNo141LesionsUniversity of ChicagoRandom splitNoHistologyNoMRIBreast cancer
Arevalo et al. 2016CNN3No736ImagesBreast Cancer Digital Repository (BCDR), PortugalStratified SamplingNoHistologyNoMammogramBreast cancer
Bandeira Diniz et al. 2018CNNNo(a) 200; (b) 288Images(a) DDSM Dense Breast; (b) DDSM Non Dense BreastRandom splitNoFollow up, histology, expert readerNoMammogramBreast cancer
Becker et al. 2017[91]dANNNo70ImagesBreast Cancer Digital Repository (BCDR)Random splitYesExpert readerYesMammogramBreast cancer
Becker et al. 2018[62]DNNNo192LesionsPrivateRandom splitNoFollow up, histologyYesUltrasoundBreast cancer
Bevilacqua et al. 2019VGG-SNo39ImagesNRNRNoNRNoDigital breast tomosynthesisBreast cancer
Byra et al. 2019[99]VGG-19No(a) 150; (b) 163; (c) 100Images(a) Moores Cancer Center, University of California; (b) UDIAT (c) OASBUDRandom splitNo(a) Follow up, histology; (b) expert reader; (c) expert reader, histology, follow upYesUltrasoundBreast cancer
Cai et al. 2019CNNNo99ImagesSYSUCC and Foshan, ChinaRandom splitNoHistologyNoMammogramBreast cancer
Cao et al. 2019SSD300 + ZFNetNo183LesionsSichuan Provincial People’s HospitalRandom splitNoExpert consensusNoUltrasoundBreast cancer
Cao et al. 2019NF-NetNo272LesionsSichuan Provincial People’s HospitalRandom splitNoHistologyNoUltrasoundBreast cancer
Cheng et al. 2016Stacked denoising autoencoderNo520LesionsTaipei Veterans General HospitalNRNoHistologyNoUltrasoundBreast Nodules
Chiao et al. 2019Mask R-CNNNo61ImagesChina Medical University HospitalRandom splitNoHistology, routine clinical reportNoUltrasoundBreast cancer
Choi et al. 2019[100]CNNNo253LesionsSamsung Medical Centre, SeoulNRNoFollow up, histologyYesUltrasoundBreast cancer
Chougrad et al. 2018Inception-v3No(a) 5316; (b) 600; (c) 200Images(a) DDSM; (b) Inbreast; (c) BCDRRandom splitNo(a) Follow up, histology, expert reader; (b) expert reader, histology; (c) clinical reportsNoMammogramBreast cancer
Ciritsis et al. 2019[92]dCNNNo(a) 101; (b) 43Images(a) Internal validation; (b) external validationRandom splitYesFollow up, histologyYesUltrasoundBreast cancer
Cogan et al. 2019[93]ResNet-101 Faster R-CNNNo124ImagesINbreastNAYesExpert reader, histology,NoMammogramBreast cancer
Dalmis et al. 2018U-NetNo66ImagesNRRandom splitNoFollow up, histologyNoMRIBreast cancer
Dalmis et al. 2019[101]DenseNetNo576LesionsRaboud University Medical CenterNRNoFollow up, histologyYesMRIBreast cancer
Dhungel et al. 2017CNNNo82ImagesINbreastRandom splitNoExpert reader, histology,NoMammogramBreast cancer
Duggento et al. 2019CNNNo378ImagesCurated Breast Imaging SubSet of DDSM (CBIS-DDSM)Random splitNoExpert readerNoMammogramBreast cancer
Fan et al. 2019Faster R-CNNNo182ImagesFudan University Affiliated Cancer CentreRandom splitNoHistologyNoDigital breast tomosynthesisBreast cancer
Fujioka et al. 2019[102]GoogleNetNo120LesionsPrivateRandom splitNoFollow up, histologyYesUltrasoundBreast cancer
Gao et al. 2018SD-CNNNo(a) 49; (b) 89(a) Lesions; (b) images(a) Mayo Clinic Arizona; (b) InbreastNRNo(a) Histology; (b) expert reader, histologyNo(a) Contrast enhanced digital mammogram; (b) mammogramBreast cancer
Ha et al. 2019CNNNo60ImagesColumbia University Medical CenterRandom splitNoFollow up, histologyNoMammogramDCIS
Han et al. 2017GoogleNetNo829LesionsSamsung Medical Centre, SeoulRandom splitNoHistologyNoUltrasoundBreast cancer
Herent et al. 2019ResNet50No168LesionsJournees Francophones de Radiologie 2018Random splitNoNRNoMRIBreast cancer
Hizukuri et al. 2018CNNNo194ImagesMie University HospitalRandom splitNoFollow up, histologyNoUltrasoundBreast cancer
Huyng et al. 2016AlexNetNo607ImagesUniversity of ChicagoNRNoHistologyNoMammogramBreast cancer
Jadoon et al. 2016CNN-DWNo2976ImagesIRMANRNoHistologyNoMammogramBreast cancer
Jiao et al. 2016CNNNo300ImagesDDSMRandom splitNoFollow up, histology, expert readerNoMammogramBreast cancer
Jiao et al. 2018(a) AlexNet; (b) parasitic metric learning layersNo(a) 150; (b) 150ImagesDDSMRandom splitNoFollow up, histology, expert readerNoMammogramBreast cancer
Jung et al. 2018RetinaNetNo(a) 410; (b) 222Images(a) Inbreast; (b) GURORandom splitNo(a) Expert reader; (b) histologyNoMammogramBreast cancer
Kim et al. 2012[103]ANNNo70LesionsKangwon National University College of MedicineRandom splitNoExpert consensusYesUltrasoundBreast cancer
Kim et al. 2018ResNetNo1238ImagesYonsei University Health SystemRandom splitNoFollow up, histologyNoMammogramBreast cancer
Kim et al. 2018VGG-16No340ImagesDDSMHold-out methodNoFollow up, histology, expert readerNoMammogramBreast cancer
Kooi et al. 2017CNNNo18,182ImagesNetherlands screening databaseRandom splitNoExpert reader, histology,NoMammogramBreast cancer
Kooi et al. 2017CNNNo1523ImagesNetherlands screening databaseRandom splitNoExpert reader, histology,NoMammogramBreast cancer
Kooi T et al. 2017CNNNo1804ImagesNetherlands screening databaseHold-out methodNoExpert reader, histology,NoMammogramBreast Cancer
Li et al. 2019DenseNet-IINo2042ImagesFirst Hospital of Shanxi Medical UniversityTenfold cross validationNoExpert readerNoMammogramBreast cancer
Li et al. 2019VGG-16No(a) 1854; (b) 1854ImagesNanfang HospitalFivefold cross validationNoFollow up, histologyNo(a) Digital breast tomosynthesis; (b) mammogramBreast cancer
Lin et al. 2014FCMNNNo65ImagesFar Eastern Memorial Hospital, TaiwanTenfold cross validationNoHistologyNoUltrasoundBreast cancer
McKinney et al. 2020[94]MobileNetV2 - ResNet-v2-50, ResNet-v1-50No(a) 25,856; (b) 3097Images(a) UK; (b) USARandom splitYesFollow up, histologyYesMammogramBreast cancer
Mendel et al. 2018VGG-19No(a) 78; (b) 78ImagesUniversity of ChicagoLeave-one-out methodNoFollow up, histologyNo(a) Mammogram; (b) digital breast tomosynthesisBreast cancer
Peng et al. 2016[95]ANNNo(a) 100; (b) 100Images(a) MIAS; (b) BancoWebHold-out methodYesExpert readerNoMammogramBreast cancer
Qi et al. 2019Inception-Resnet-v2No1359ImagesWest China Hospital, Sichuan UniversityRandom splitNoExpert consensusNoUltrasoundBreast cancer
Qiu et al. 2017CNNNo140ImagesPrivateRandom splitNoHistologyNoMammogramBreast cancer
Ragab et al. 2019AlexNetNo(a) 676; (b) 1581Images(a) Digital database for screening mammography (DDSM); (b) Curated Breast Imaging SubSet of DDSM (CBIS-DDSM)Random splitNoFollow up, histology, expert readerNoMammogramBreast cancer
Ribli et al. 2018[96]VGG-16No115ImagesINbreastNAYesExpert reader, histologyNoMammogramBreast cancer
Rodriguez-Ruiz et al. 2018[97]CNNNo240ImagesTwo datasets combinedNAYesExpert reader, histology, follow upYesMammogramBreast cancer
Rodriguez-Ruiz et al. 2019[98]CNNNo2642ImagesCombined nine datasetsNAYesFollow up, histologyYesMammogramBreast cancer
Samala et al. 2016DCNNNo94ImagesUniversity of MichiganRandom splitNoExpert readerNoDigital breast tomosynthesisBreast cancer
Samala et al. 2017DCNNNo907ImagesDDSM + privateRandom splitNoExpert readerNoMammogramBreast cancer
Samala et al. 2018DCNNNo94ImagesUniversity of MichiganRandom splitNoExpert readerNoDigital breast tomosynthesisBreast cancer
Samala et al. 2019AlexNetNo94ImagesUniversity of MichiganRandom splitNoExpert readerNoDigital breast tomosynthesisBreast cancer
Shen et al. 2019(a) VGG-16; (b) ResNet; (c) ResNet-VGGNo(a) 376; (b) 376; (c) 107Images(a) Curated Breast Imaging SubSet of DDSM (CBIS-DDSM); (b) Curated Breast Imaging SubSet of DDSM (CBIS-DDSM); (c) InbreastRandom splitNo(a) Histology; (b) histology; (c) expert readerNoMammogramBreast cancer
Shin et al. 2019VGG-16No(a) 600; (b) 40Images(a) Seoul National University Bundang Hospital; (b) UDIAT Diagnostic Centre of the Parc Taulí CorporationRandom splitNo(a) NR; (b) expert readerNoUltrasoundBreast cancer
Stoffel et al. 2018CNNNo33ImagesPrivateRandom splitNoSurgical confirmationYesUltrasoundPhyllodes tumour
Sun et al. 2017CNNNo758ImagesUniversity of Texas at El PasoRandom splitNoExpert readerNoMammogramBreast cancer
Tanaka et al. 2019VGG-19, Resnet152No154LesionsJapan Association of Breast and Thyroid SonologyRandom splitNoHistologyNoUltrasoundBreast cancer
Tao et al. 2019RefineNet + DenseNet121No253LesionsHuaxi Hospital and China-Japan Friendship HospitalRandom splitNoExpert readerNoUltrasoundBreast cancer
Teare et al. 2017Inception-v3No352ImagesDDSM + Zebra Mammography DatasetRandom splitNoFollow up, histologyNoMammogramBreast cancer
Truhn et al. 2018[104]CNNNo129LesionsRWTH Aachen University,Random splitNoFollow up, histologyYesMRIBreast cancer
Wang et al. 2016Inception-v3No74ImagesBreast Cancer Digital Repository (BCDR)Random splitNoExpert reader, histologyNoMammogramBreast cancer
Wang et al. 2016Stacked autoencoderNo204ImagesSun Yat-sen University Cancer Center (Guangzhou, China) and Nanhai Affiliated Hospital of Southern Medical University (Foshan, China)Hold-out methodNoHistologyNoMammogramBreast cancer
Wang et al. 2017CNNNo292ImagesUniversity of ChicagoRandom splitNoHistologyNoMammogramBreast cancer
Wang et al. 2018DNNNo292ImagesUniversity of ChicagoRandom splitNoHistologyNoMammogramBreast cancer
Wu et al. 2019[105]ResNet-22No(a) 401; (b) 1440ImagesNYUHold-out methodNoHistologyYesMammogramBreast cancer
Xiao et al. 2019Inception-v3, ResNet50, XceptionNo206ImagesThird Affiliated Hospital of Sun Yat-sen UniversityRandom splitNoSurgical confirmation, histologyNoUltrasoundBreast cancer
Yala et al. 2019[106]ResNet18No26,540ImagesMassachusetts General Hospital, Harvard Medical School,Random splitNoClinical reports, follow up, histologyYesMammogramBreast cancer
Yala et al. 2019[111]ResNet18No8751ImagesMassachusetts General Hospital, Harvard Medical School,Random splitNoClinical reports, follow up, histologyNoMammogramBreast cancer
Yap et al. 2018FCN-AlexNetNo(a) 306; (b) 163Lesions(a) Private; (b) UDIATNRNoExpert readerNoUltrasoundBreast cancer
Yap et al. 2019FCN-8sNo94LesionsTwo datasets combinedNRNoExpert readerNoUltrasoundBreast cancer
Yousefi et al. 2018DCNNNo28ImagesMGHRandom splitNoExpert consensusNoDigital breast tomosynthesisBreast cancer
Zhou et al. 2019[107]3D DenseNetNo307LesionsPrivateRandom splitNoFollow up, histologyYesMRIBreast cancer

PRISMA flow diagram of included studies.

PRISMA (preferred reporting items for systematic reviews and meta-analyses) flow diagram of included studies. Summary estimates of pooled speciality and imaging modality specific diagnostic accuracy metrics. Characteristics of ophthalmic imaging studies. Characteristics of respiratory imaging studies. Characteristics of breast imaging studies.

Ophthalmology imaging

Eighty-two studies with 143 separate patient cohorts reported diagnostic accuracy data for DL in ophthalmology (see Table 2 and Supplementary References 1). Optical coherence tomography (OCT) and retinal fundus photographs (RFP) were the two imaging modalities performed in this speciality with four main pathologies being diagnosed—diabetic retinopathy (DR), age-related macular degeneration (AMD), glaucoma and retinopathy of prematurity (ROP). Only eight studies[14-21] used prospectively collected data and 29 (refs. [14,15,17,18,21-45]) studies validated algorithms on external datasets. No studies provided a prespecified sample size calculation. Twenty-five studies[17,28,29,35,37,39,40,44-61] compared algorithm performance against healthcare professionals. Reference standards, definitions of disease and threshold for diagnosis varied greatly as did the method of internal validation used. There was high heterogeneity across all studies (see Table 2). Diabetic retinopathy: Twenty-five studies with 48 different patient cohorts reported diagnostic accuracy data for all, referable or vision-threatening DR on RFP. Twelve studies and 16 cohorts reported on diabetic macular oedema (DME) or early DR on OCT scans. AUC was 0.939 (95% CI 0.920–0.958) for RFP versus 1.00 (95% CI 0.999–1.000) for OCT. Age-related macular degeneration: Twelve studies reported diagnostic accuracy data for features of varying severity of AMD on RFP (14 cohorts) and 11 studies in OCT (21 cohorts). AUC was 0.963 (95% CI 0.948–0.979) for RFP versus 0.969 (95% CI 0.955–0.983) for OCT. Glaucoma: Seventeen studies with 30 patient cohorts reported diagnostic accuracy for features of glaucomatous optic neuropathy, optic discs or suspect glaucoma on RFP and five studies with 6 cohorts on OCT. AUC was 0.933 (95% CI 0.924–0.942) for RFP and 0.964 (95% CI 0.941–0.986) for OCT. One study[34] with six cohorts on RFP provided contingency tables. When averaging across the cohorts, the pooled sensitivity was 0.94 (95% CI 0.92–0.96) and pooled specificity was 0.95 (95% CI 0.91–0.97). The AUC of the summary receiver-operating characteristic (SROC) curve was 0.98 (95% CI 0.96–0.99)—see Supplementary Fig. 1. Retinopathy of prematurity: Three studies reported diagnostic accuracy for identifying plus diseases in ROP from RFP. Sensitivity was 0.960 (95% CI 0.913—1.008) and specificity was 0.907 (95% CI 0.907–1.066). AUC was only reported in two studies so was not pooled. Others: Eight other studies reported on diagnostic accuracy in ophthalmology either using different imaging modalities (ocular images and visual fields) or for identifying other diagnoses (pseudopapilloedema, retinal vein occlusion and retinal detachment). These studies were not included in the meta-analysis.

Respiratory imaging

One hundred and fifteen studies with 244 separate patient cohorts report on diagnostic accuracy of DL on respiratory disease (see Table 3 and Supplementary References 2). Lung nodules were largely identified on CT scans, whereas chest X-rays (CXR) were used to diagnose a wide spectrum of conditions from simply being ‘abnormal’ to more specific diagnoses, such as pneumothorax, pneumonia and tuberculosis.
Table 3

Characteristics of respiratory imaging studies.

StudyModelProspective?Test setPopulationTest datasetsType of internal validationExternal validationReference standardAI vs clinicianImaging modalityBody system/disease
Abiyev et al. 2018CNNNo380ImagesChest X-ray14Random splitNoRoutine clinical reportsNoX-rayAbnormal X-ray
Al-Shabi et al. 2019Local-GlobalNo848NodulesLIDC-IDRINRNoExpert readersNoCTNodules
Alakwaa et al. 2017U-NetNo419ScansKaggle Data Science BowlRandom splitNoExpert reader, existing labels in datasetNoCTLung cancer
Ali et al. 20183D CNNNo668NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Annarumma et al. 2019[110]CNNNo15,887ImagesKings College LondonHold-out methodNoRoutine clinical reportsNoX-ray(a) Critical radiographs; (b) normal radiographs
Ardila et al. 2019[64]Inception-v1No(a) 6716; (b) 1139Scans(a) National Lung Cancer Screening Trial; (b) Northwestern MedicineRandom splitYesHistopathology, follow upYesCTLung cancer
Baltruschat et al. 2019ResNet50No22,424X-raysChest X-ray14Random splitNoRoutine clinical reportsNoX-ray(a) Abnormal chest X-ray; (b) normal chest X-ray; (c) atelectasis; (d) cardiomegaly; (e) effusion; (f) infiltration; (g) mass; (h) nodule (i) pneumonia; (j) pneumothorax; (k) consolidation; (l) oedema; (m) emphysema; (n) fibrosis; (o) pleural thickening; (p) hernia
Bar et al. 2018CNNNo194ImagesDiagnostic Imaging Department of Sheba Medical Centre, Tel Hashomer, IsraelRandom splitNoExpert readersNoX-ray(a) Abnormal X-ray; (b) cardiomegaly
Becker et al. 2018[62]CNNYes21X-raysInfectious Diseases Institute in Kampala, UgandaRandom splitNoExpert consensusNoX-rayTuberculosis
Behzadi-Khormouji et al. 2020(a) ChestNet; (b) VGG-16; (c) DenseNet121No582X-raysGuangzhou Women and Children’s Medical CentreNRNoExpert readersNoX-rayConsolidation
Beig et al. 2019CNNNo145ScansErlangen Germany, Waukesha Wis, Cleveland Ohio, Tochigi-ken JapanRandom splitNoHistopathologyNoCTLung cancer
Causey et al. 2018CNNNo(a) 424; (b) 213NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Cha et al. 2019[76]ResNet50No(a) 1483; (b) 500X-raysSamsung Medical Centre, SeoulRandom splitNoOther imaging, expert readersYesX-ray(a) Lung cancer; (b) T1 lung cancer
Chae et al. 2019[77]Ct-LUNGNETNo60NodulesChonbuk National University HospitalRandom splitNoExpert readers, histopathology, follow upYesCTNodules
Chakravarthy et al. 2019Probabilistic neural networkNo119ScansLIDC/IDRINRNoNRNoCTLung cancer
Chen et al. 20193D CNNNo3674NodulesLIDC-IDRINRNoExpert readersNoCTNodules
Cheng et al. 2016Stacked denoising autoencoderNo1400NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Cicero et al. 2017GoogLeNetNo2443ImagesDepartment of Medical Imaging, St Michael’s Hospital, TorontoRandom splitNoExpert readers, routine clinical reportsNoX-ray(a) Effusion; (b) oedema; (c) consolidation; (d) cardiomegaly; (e) pneumothorax
Ciompi et al. 2017[78]ConvNetNo639NodulesDanish Lung Cancer Screening Trial (DLCST)Random splitNoNon-expert readersYesCT(a) Nodules—solid; (b) nodules—calcified; (c) nodules—part-solid; (d) nodules—non-solid; (e) nodules—perifissural; (f) nodules—spiculated
Correa et al. 2018CNNNo60ImagesLima, PeruNRNoExpert readersNoUltrasoundPaediatric pneumonia
da Silva et al. 2017Evolutionary CNNNo200NodulesLIDC-IDRIHold-out methodNoExpert readersNoCTNodules
da Silva et al. 2018Particle swarm optimisation algorithm within CNNNo2000NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Dai et al. 20183D DenseNet-40No211NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Dou et al. 20173D CNNNo1186NodulesLUNA16NRNoExpert readersNoCTNodules
Dunnmon et al. 2019[79]ResNet18No533ImagesStanford UniversityHold-out methodNoExpert consensusYesX-rayAbnormal X-ray
Gao et al. 2018CNNNo20ScansUniversity Hospitals of GenevaRandom splitNoNRNoCTInterstitial lung disease
Gong et al. 20193D SE-ResNetNo1186NodulesLUNA16NRNoExpert readersNoCTNodules
Gonzalez et al. 2018CNNNo1000ScansECLIPSE studyRandom splitNoNRNoCTCOPD
Gruetzemacher et al. 2018DNNNo1186NodulesLUNA16Ninefold cross validationNoNRNoCTNodules
Gu et al. 20183D CNNNo1186NodulesLUNA16Tenfold cross validationNoExpert readersNoCTNodules
Hamidian et al. 20173D CNNNo104NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Han et al. 2018Multi-CNNsNo812Regions of interestLIDC-IDRIRandom splitNoNRNoCTGround glass opacity
Heo et al. 2019VGG-19No37,677X-raysYonsei University Hospital, South KoreaHold-out methodNoExpert readersNoX-rayTuberculosis
Hua et al. 2015(a) CNN; (b) deep belief networkNo2545NodulesLIDC-IDRINRNoExpert readersNoCTNodules
Huang et al. 2019R-CNNNo176ScansLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Huang et al. 2019Amalgamated-CNNNo1795NodulesLIDC/IDRI and Ali Tianchi medicalRandom splitNoExpert readersNoCTNodules
Hussein et al. 2019VGGNo1144NodulesLIDC/IDRIRandom splitNoExpert readersNoCTLung cancer
Hwang et al. 2018[67]DCNNNo(a) 450; (b) 183; (c) 140; (d) 173; (e) 170; (f) 132; (g) 646X-rays(a) Internal validation; (b) Seoul National University Hospital; (c) Boromae Hospital; (d) Kyunghee University Hospital; (e) Daejeon Eulji Medical Centre; (f) Montgomery; (g)Random splitYesExpert readersYesX-rayTuberculosis
Hwang et al. 2019[65]Lunit INSIGHTNo1135X-raysSeoul National University HospitalNAYesExpert consensus, other imagingYesX-rayAbnormal chest X-ray
Hwang et al. 2019[66]DCNNNo(a) 1089; (b) 1015X-rays(a) Internal validation; (b) external validationRandom splitYesExpert reader, other imaging, histopathologyYesX-rayNeoplasm/TB/pneumonia/pneumothorax
Jiang et al. 2018CNNNo25,723NodulesLIDC-IDRINRNoExpert readersNoCTNodules
Jin et al. 2018ResNet 3DNo1186NodulesLUNA16NRNoExpert readersNoCTNodules
Jung et al. 20183D DCNNNo1186NodulesLUNA16NRNoExpert readersNoCTNodules
Kang et al. 20173D multi view-CNNNo776NodulesLIDC-IDRINRNoExpert readersNoCTNodules
Kermany et al. 2018Inception-v3No624X-raysGuangzhou Women and Children’s Medical CentreRandom splitNoExpert readersYesX-rayPneumonia
Kim et al. 2019MGI-CNNNo1186NodulesLIDC/IDRINRNoExpert readersNoCTNodules
Lakhani et al. 2017[90](a) AlexNet; (b) GoogLeNet; (c) Ensemble (AlexNet + GoogLeNet); (d) Radiologist augmentedNo150X-raysMontgomery County MD, Shenzhen China, Belarus TB public Health Program, Thomas Jefferson University HospitalRandom splitNoRoutine clinical reports, expert reader, histopathologyNoX-rayTuberculosis
Li et al. 2016CNNNo8937NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Li et al. 2019[81]DL-CADNo812NodulesShenzhen HospitalNRNoExpert consensusYesCTNodules
Li et al. 2019[80]CNNNo200ScansMassachusetts General HospitalRandom splitNoRoutine clinical reportsYesCTPneumothorax
Liang et al. 2020[68]CNNNo100ImagesKaohsiung Veterans General Hospital, TaiwanNAYesOther imagingNoX-rayNodules
Liang et al. 2019(a) Custom CNN; (b) VGG-16; (c) DenseNet121; (d) Inception-v3; (e) XceptionNo624X-raysGuangzhou Women and Children’s Medical CentreRandom splitNoExpert readersNoX-rayPneumonia
Liu et al. 20173D CNNNo326NodulesNational Lung Cancer Screening Trial and Early Lung Cancer Action ProgramFivefold cross validationNoHistopathology, follow upNoCTNodules
Liu et al. 2019CDP-ResNetNo539NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Liu H et al. 2019Segmentation-based deep fusion networkNo112,120X-raysChest X-ray14NRNoRoutine clinical reportsNoX-ray(a) Atelectasis; (b) cardiomegaly; (c) effusion; (d) infiltration; (e) mass; (f) nodule; (g) pneumonia; (h) pneumothorax; (i) consolidation; (j) oedema; (k) emphysema; (l) fibrosis; (m) fibrosis; (n) pleural thickening; (o) hernia
Majkowska et al. 2019[82]CNNNo(a–d) 1818; (e–h) 1962X-rays(a–d) Hospital group in India (Bangalore, Bhubaneshwar, Chennai, Hyderabad, New Delhi); (e–h) Chest X-ray14Random splitNoExpert consensusYesX-ray(a) Pneumothorax (b) nodule; (c) opacity; (d) fracture; (e) pneumothorax; (f) nodule; (g) opacity; (h) fracture
Monkam et al. 2018CNNNo2600NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Nam et al. 2018[69]CNNNo(a) 600; (b) 181; (c) 182; (d) 181; (e) 149Chest radiographs(a) Internal validation; (b) Seoul National University Hospital; (c) Boromae Hospital; (d) National Cancer Centre, Korea; (e) University of California an Francisco Medical CentreRandom splitYes(a) Routine clinical reports, histopathology; (b–e) histopathology, follow up, other imagingNoX-rayNodules
Naqi et al. 2018Two-level stacked autoencoder + softmaxNo777NodulesLIDC-IDRINRNoExpert readersNoCTNodules
Nasrullah et al. 2019Faster R-CNNNo2562NodulesLIDC/IDRINRNoExpert readersNoCTNodules
Nibali et al. 2017ResNetNo166NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Nishio et al. 2018VGG-16No123NodulesKyoto University HospitalRandom splitNoNRNoCTNodules
Onishi et al. 2019AlexNetNo60NodulesNRNRNoHistopathology, follow upNoCTNodules
Onishi et al. 2019Wasserstein generative adversarial networkNo60NodulesFujita Health University HospitalNRNoHistopathology, follow upNoCTNodules
Park et al. 2019[89]YOLONo503X-raysAsan Medical Centre and Seoul National University Bundang HospitalHold-out methodNoExpert readerNoX-rayPneumothorax
Park et al. 2019[83]CNNNo200ImagesAsan Medical Centre and Seoul National University Bundang HospitalHold-out methodNoExpert consensusYesX-ray(a) Nodules; (b) opacity; (c) effusion; (d) pneumothorax; (e) abnormal chest X-ray
Pasa et al. 2019Custom CNNNo220X-raysNIH Tuberculosis Chest X-ray dataset and Belarus Tuberculosis Portal datasetRandom splitNoNRNoX-rayTuberculosis
Patel et al. 2019[84]CheXMaxNo50X-raysStanford UniversityHold-out methodNoExpert reader, other imaging, clinical notesYesX-rayPneumonia
Paul et al. 2018VGG-s CNNNo237NodulesNational Lung Cancer Screening TrialHold-out methodNoExpert readers, follow upNoCTNodules
Pesce et al. 2019Convolution networks with attention feedback (CONAF)No7850X-raysGuy’s and St. Thomas’ NHS Foundation TrustRandom splitNoRoutine clinical reportsNoX-rayLung lesions
Pezeshk et al. 20193D CNNNo128NodulesLUNA16Random splitNoExpert readersNoCTNodules
Qin et al. 2019[70](a) Lunit; (b) qXR (Qure.ai); (c) CAD4TBNo1196X-raysNepal and CameroonNAYesExpert readersYesX-rayTuberculosis
Rajpurkar et al. 2018[85]CNNNo420X-raysChestXray-14Random splitNoRoutine clinical reportsYesX-ray(a) Atelectasis; (b) cardiomegaly; (c) consolidation; (d) oedema; (e) effusion; (f) emphysema; (g) fibrosis; (h) hernia; (i) infiltration; (j) mass; (k) nodule; (l) pleural thickening; (m) pneumonia; (n) pneumothorax
Ren et al. 2019Manifold regularized classification deep neural networkNo98NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Sahu et al. 2019Multi-section CNNNo130NodulesLIDC-IDRITenfold cross validationNoExpert readersNoCTNodules
Schwyzer et al. 2018CNNNo100PatientsNRNRNoNRNoFDG-PETLung cancer
Setio et al. 2016[71]ConvNetNo(a) 1186; (b) 50; (c) 898(a) Nodules; (b) scans; (c) nodulesLIDC-IDRIFivefold cross validationYes(a) Expert readers; (b, c) NRNoCTNodules
Shaffie et al. 2018Deep autoencoderNo727NodulesLIDC-IDRINRNoExpert readersNoCTNodules
Shen et al. 2017Multiscale CNNNo1375NodulesLIDC-IDRINRNoExpert readersNoCTNodules
Sim et al. 2019[72]ResNet50No800ImagesFreiberg University Hospital Freiburg, Massachusetts General Hospital Boston, Samsung Medical Centre Seoul, Severance Hospital SeoulNAYesOther imaging, histopathologyYesX-rayNodules
Singh et al. 2018[86]Qure-AINo724Chest radiographsChest X-ray8Random splitNoRoutine clinical reportsYesX-ray(a) Lesions; (b) effusion; (c) hilar prominence; (d) cardiomegaly
Song et al. 2017(a) CNN; (b) DNN; (c) stacked autoencoderNo5024NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Stephen et al. 2019CNNNo2134ImagesGuangzhou Women and Children’s Medical CentreRandom splitNoNRNoX-rayPneumonia
Sun et al. 2017(a) CNN; (b) deep belief network; (c) stacked denoising autoencoderNo88,948SamplesLIDC-IDRITenfold cross validationNoExpert readersNoCTNodules
Tan et al. 2019CNNNo280NodulesLIDC-IDRITenfold cross validationNoNRNoCTNodules
Taylor et al. 2018[73](a) Inception-v3; (b) VGG-19; (c) Inception-v3; (d) VGG-19No(a, b) 1990; (c, d) 112,120X-rays(a,b) Internal validation (c,d) Chest X-ray14Random splitYesExpert consensusNoX-rayPneumothorax
Teramoto et al. 2016CNNNo104ScansFujita Health University HospitalNRNoExpert readerNoPET/CTNodules
Togacar et al. 2019AlexNet + VGG-16 + VGG-19No1754X-raysFirat University, TurkeyRandom splitNoNRNoX-rayPneumonia
Togacar et al. 2020(a) LeNet; (b) AlexNet; (c) VGG-16No100ImagesCancer Imaging ArchiveNRNoExpert readersNoCTLung cancer
Tran et al. 2019LdcNetNo1186NodulesLUNA16Tenfold cross validationNoExpert readersNoCTNodules
Tu et al. 2017CNNNo20NodulesLIDC-IDRITenfold cross validationNoExpert readersNoCT(a) Nodules—non-solid; (b) nodules—part-solid; (c) nodules—solid
Uthoff et al. 2019[74]CNNNo100NodulesINHALE STUDYNAYesHistopathology, follow upNoCTNodules
Walsh et al. 2018[87]Inception-ResNet-v2No150ScansLa Fondazione Policlinico Universitario A Gemelli IRCCS, Rome, Italy, and University of Parma, Parma, ItalyRandom splitNoExpert readersYesCTInterstitial lung disease
Wang et al. 2017AlexNetNo230X-raysJapanese Society of Radiological Technology (JSRT) databaseTenfold cross validationNoOther imagingNoX-rayNodules
Wang et al. 2018[88]3D CNNNo200ScansFudan University Shanghai Cancer CentreRandom splitNoExpert readers, histopathologyYesHRCTLung cancer
Wang et al. 2018VGG-16No744X-raysJSRT, OpenI, SZCX and MCRandom splitNoOther imagingNoX-ray(a) Abnormal chest X-ray; (b) normal chest X-ray
Wang et al. 2019ChestNetNo442X-raysZhejiang University School of Medicine (ZJU-2) and Chest X-ray14Random splitNoExpert readersNoX-rayPneumothorax
Wang et al. 2019(a) AlexNet; (b) GoogLeNet; (c) ResNetNo7580NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Wang et al. 2019ResNet152No25,596X-raysChest X-ray14Random splitNoRoutine clinical reportsNoX-ray(a) Atelectasis; (b) cardiomegaly; (c) effusion; (d) infiltration; (e) mass; (f) nodule; (g) pneumonia; (h) pneumothorax; (i) consolidation; (j) oedema; (k) emphysema; (l) fibrosis; (m) pleural thickening; (n) hernia; (o) abnormal chest X-ray
Xie et al. 2018LeNet-5No1972NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Xie et al. 2019ResNet50No1945NodulesLIDC-IDRITenfold cross validationNoExpert readersNoCTNodules
Yates et al. 2018Inception-v3No5505X-raysChest X-ray14 + Indiana UniversityRandom splitNoRoutine clinical reportsNoX-rayAbnormal chest X-ray
Ye et al. 2019(a) AlexNet; (b) GoogLeNet; (c) Res-Net150No(a) 321; (b) 321; (c) 593(a) Nodules; (b) nodules; (c) regions of interest(a, b) LIDC-IDRI; (c) privateRandom splitNoExpert readersNoCT(a, b) Nodules; (c) ground glass opacity
Zech et al. 2018[75]CNNNo(a) 30,450; (b) 3807X-rays(a) Mount Sinai and Chest X-ray14; (b) Indiana University Network for Patient CareRandom splitYesExpert readersNoX-rayPneumonia
Zhang et al. 20183D DCNNNo1186NodulesLUNA16NRNoExpert readersNoCTNodules
Zhang et al. 2019Voxel-level-1D CNNNo67NodulesStony Brook University HospitalTwofold cross validationNoHistopathologyNoCTNodules
Zhang et al. 20193D deep dual path networkNo1004NodulesLIDC/IDRITenfold cross validationNoExpert readersNoCTNodules
Zhang C et al. 20193D CNNYes50ImagesGuangdong Lung Cancer InstituteRandom splitYesHistopathology, follow upYesCTNodules
Zhang et al. 2019[63]Mask R-CNNNo134SlicesShenzhen HospitalRandom splitNoExpert readersNoCT/PETLung cancer
Zhang S et al. 2019Le-Net5No762NodulesLIDC/IDRIRandom splitNoExpert readersNoCTNodules
Zhang T et al. 2017Deep Belief NetworkNo1664NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Zhao X et al. 2018Agile CNNNo743NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Zhao X et al. 2019(a) AlexNet; (b) GoogLeNet; (c) ResNet; (d) VifarNetNo2028NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Zheng et al. 2019CNNNo1186NodulesLIDC-IDRIRandom splitNoExpert readersNoCTNodules
Zhou et al. 2019Inception-v3 and ResNet50No600ImagesChest X-ray8Random splitNoRoutine clinical reportsNoX-rayCardiomegaly
Only two studies[62,63] used prospectively collected data and 13 (refs. [63-75]) studies validated algorithms on external data. No studies provided a prespecified sample size calculation. Twenty-one[54,63-67,70,72,76-88] studies compared algorithm performance against healthcare professionals. Reference standards varied greatly as did the method of internal validation used. There was high heterogeneity across all studies (see Table 3). Lung nodules: Fifty-six studies with 74 separate patient cohorts reported diagnostic accuracy for identifying lung nodules on CT scans on a per lesion basis, compared with nine studies and 14 patient cohorts on CXR. AUC was 0.937 (95% CI 0.924–0.949) for CT versus 0.884 (95% CI 0.842–0.925) for CXR. Seven studies reported on diagnostic accuracy for identifying lung nodules on CT scans on a per scan basis, these were not included in the meta-analysis. Lung cancer or mass: Six studies with nine patient cohorts reported diagnostic accuracy for identifying mass lesions or lung cancer on CT scans compared with eight studies and ten cohorts on CXR. AUC was 0.887 (95% CI 0.847–0.928) for CT versus 0.864 (95% CI 0.827–0.901) for CXR. Abnormal Chest X-ray: Twelve studies reported diagnostic accuracy for abnormal CXR with 13 different patient cohorts. AUC was 0.917 (95% CI 0.869–0.966), sensitivity was 0.873 (95% CI 0.762–0.985) and specificity was 0.894 (95% CI 0.860–0.929). Pneumothorax: Ten studies reported diagnostic accuracy for pneumothorax on CXR with 14 different patient cohorts. AUC was 0.910 (95% CI 0.863–0.957), sensitivity was 0.718 (95% CI 0.433–1.004) and specificity was 0.918 (95% CI 0.870–0.965). Five patient cohorts from two studies[73,89] provided contingency tables with raw diagnostic accuracy. When averaging across the cohorts, the pooled sensitivity was 0.70 (95% CI 0.45–0.87) and pooled specificity was 0.94 (95% CI 0.90–0.97). The AUC of the SROC curve was 0.94 (95% CI 0.92–0.96)—see Supplementary Fig. 2. Pneumonia: Ten studies reported diagnostic accuracy for pneumonia on CXR with 15 different patient cohorts. AUC was 0.845 (95% CI 0.782–0.907), sensitivity was 0.951 (95% CI 0.936–0.965) and specificity was 0.716 (95% CI 0.480–0.953). Tuberculosis: Six studies reported diagnostic accuracy for tuberculosis on CXR with 17 different patient cohorts. AUC was 0.979 (95% CI 0.978–0.981), sensitivity was 0.998 (95% CI 0.997–0.999) and specificity was 1.000 (95% CI 0.999–1.000). Four patient cohorts from one study[90] provided contingency tables with raw diagnostic accuracy. When averaging across the cohorts, the pooled sensitivity was 0.95 (95% CI 0.91–0.97) and pooled specificity was 0.97 (95% CI 0.93–0.99). The AUC of the SROC curve was 0.97 (95% CI 0.96–0.99)—see Supplementary Fig. 3. X-ray imaging was also used to identify atelectasis, pleural thickening, fibrosis, emphysema, consolidation, hiatus hernia, pulmonary oedema, infiltration, effusion, mass and cardiomegaly. CT imaging was also used to diagnose COPD, ground glass opacity and interstitial lung disease, but these were not included in the meta-analysis.

Breast imaging

Eighty-two studies with 100 separate patient cohorts report on diagnostic accuracy of DL on breast disease (see Table 4 and Supplementary References 3). The four imaging modalities of mammography (MMG), digital breast tomosynthesis (DBT), ultrasound and magnetic resonance imaging (MRI) were used to diagnose breast cancer. No studies used prospectively collected data and eight[91-98] studies validated algorithms on external data. No studies provided a prespecified sample size calculation. Sixteen studies[62,91,92,94,97-107] compared algorithm performance against healthcare professionals. Reference standards varied greatly as did the method of internal validation used. There was high heterogeneity across all studies (see Table 4). Breast cancer: Forty-eight studies with 59 separate patient cohorts reported diagnostic accuracy for identifying breast cancer on MMG (AUC 0.873 [95% CI 0.853–0.894]), 22 studies and 25 patient cohorts on ultrasound (AUC 0.909 [95% CI 0.881–0.936]), and eight studies on MRI (AUC 0.868 [95% CI 0.850–0.886]) and DBT (AUC 0.908 [95% CI 0.880–0.937]).

Other specialities

Our literature search also identified 224 studies in other medical specialities reporting on diagnostic accuracy of DL algorithms to identify disease. These included large numbers of studies in the fields of neurology/neurosurgery (78), gastroenterology/hepatology (24) and urology (25). Out of the 224 studies, only 55 compared algorithm performance against healthcare professionals, although 80% of studies in the field of dermatology did (see Supplementary References 4, Supplementary Table 1 and Supplementary Fig. 4).

Variation of reporting

A key finding of our review was the large degree of variation in methodology, reference standards, terminology and reporting among studies in all specialities. The most common variables amongst DL studies in medical imaging include issues with the quality and size of datasets, metrics used to report performance and methods used for validation (see Table 5). Only eight studies in ophthalmology imaging[14,21,32,33,43,55,108,109], ten studies in respiratory imaging[64,66,70,72,75,79,82,87,89,110] and six studies in breast imaging[62,91,97,104,106,111] mentioned adherence to the STARD-2015 guidelines or had a STARD flow diagram in the manuscript.
Table 5

Variation in DL imaging studies.

Data
Image pre-processing, augmentation and preparation

Are data augmentation techniques such as cropping, padding and flipping used?

Is there quality control of the images being used to train the algorithm? I.e., were poor quality images excluded.

Were relevant images manually selected?

Study designRetrospective or prospective data collection.
Image eligibility

How are images chosen for inclusion in the study?

Were the data from private or open-access repositories?

Training, validation, test sets

Are each of the three sets independent of each other, without overlap?

Does data from the same patient appear in multiple datasets?

Datasets

Are the datasets used single or multicentre?

Is a public or open-source dataset used?

Size of datasets

Wide variation in size of datasets for training and testing.

Is the size of the datasets justified?

Are sample size statistical considerations applied for the test set?

Use of ‘external’ test sets for final reporting

Is an independent test set used for ‘external validation’?

Is the independent test set constructed using an unenriched representative sample?

Multi-vendor images

Are images from different scanners and vendors included in the datasets to enhance generalisability?

Are imaging acquisition parameters described?

Algorithm
Index test

Was sufficient detail given on the algorithm to allow replication and independent validation?

What type of algorithm was used? E.g., CNN, Autoencoder, SVM.

Was the algorithm made publicly or commercially available?

Was the construct or architecture of the algorithm made available?

Additional AI algorithmic informationIs the algorithm a static model or is it continuously evolving?
Demonstrate how algorithm makes decisionsIs there a specific design for end-user interpretability, e.g., saliency or probability maps
Methods
Transfer learningWas transfer learning used for training and validation?
Cross validationWas k-fold cross validation used during training to reduce the effects of randomness in dataset splits?
Reference standard

Is the reference standard used of high quality and widely accepted in the field?

What was the rationale for choosing the reference standard?

Additional clinical informationWas additional clinical information given to healthcare professionals to simulate normal clinical process?
Performance benchmarking

What was performance of algorithm benchmarked to?

What is expertise level and level of consensus of healthcare professionals if used?

Results
Raw diagnostic accuracy dataAre raw diagnostic accuracy data reported in a contingency table demonstrating TP, FP, FN, TN?
Metrics for estimating diagnostic accuracy performanceWhich diagnostic accuracy metrics reported? Sensitivity, Sensitivity, PPV, NPV, Accuracy, AUROC
Unit of assessmentWhich unit of assessment reported, e.g., per patient, per scan or per lesion.

Rows in bold are part of STARD-2015 criteria.

Variation in DL imaging studies. Are data augmentation techniques such as cropping, padding and flipping used? Is there quality control of the images being used to train the algorithm? I.e., were poor quality images excluded. Were relevant images manually selected? How are images chosen for inclusion in the study? Were the data from private or open-access repositories? Are each of the three sets independent of each other, without overlap? Does data from the same patient appear in multiple datasets? Are the datasets used single or multicentre? Is a public or open-source dataset used? Wide variation in size of datasets for training and testing. Is the size of the datasets justified? Are sample size statistical considerations applied for the test set? Is an independent test set used for ‘external validation’? Is the independent test set constructed using an unenriched representative sample? Are images from different scanners and vendors included in the datasets to enhance generalisability? Are imaging acquisition parameters described? Was sufficient detail given on the algorithm to allow replication and independent validation? What type of algorithm was used? E.g., CNN, Autoencoder, SVM. Was the algorithm made publicly or commercially available? Was the construct or architecture of the algorithm made available? Is the reference standard used of high quality and widely accepted in the field? What was the rationale for choosing the reference standard? What was performance of algorithm benchmarked to? What is expertise level and level of consensus of healthcare professionals if used? Rows in bold are part of STARD-2015 criteria. Funnel plots were produced for the diagnostic accuracy outcome measure with the largest number of patient cohorts in each medical speciality, in order to detect bias in the studies included[112] (see Supplementary Figs. 5–7). These demonstrate that there is high risk of bias in studies detecting lung nodules on CT scans and detecting DR on RFP, but not for detecting breast cancer on MMG.

Assessment of the validity and applicability of the evidenc

The overall risk of bias and applicability using Quality Assessment of Diagnostic Accuracies Studies 2 (QUADAS-2) led to a majority of studies in all specialities being classified as high risk, particularly with major deficiencies in regard to patient selection, flow and timing and applicability of the reference standard (see Fig. 2). For the patient selection domain, a high or unclear risk of bias was seen in 59/82 (72%) of ophthalmic studies, 89/115 (77%) of respiratory studies and 62/82 (76%) or breast studies. These were mostly related to a case–control study design and sampling issues. For the flow and timing domain, a high or unclear risk of bias was seen in 66/82 (80%) of ophthalmic studies, 93/115 (81%) of respiratory studies and 70/82 (85%) of breast studies. This was largely due to missing information about patients not receiving the index test or whether all patients received the same reference standard. For the reference standard domain, concerns regarding applicability was seen in 60/82 (73%) of ophthalmic studies, 104/115 (90%) of respiratory studies and 78/82 (95%) of breast studies. This was mostly due to reference standard inconsistencies if the index test was validated on external datasets.
Fig. 2

QUADAS-2 summary plots.

Risk of bias and applicability concerns summary about each QUADAS-2 domain presented as percentages across the 82 included studies in ophthalmic imaging (a), 115 in respiratory imaging (b) and 82 in breast imaging (c).

QUADAS-2 summary plots.

Risk of bias and applicability concerns summary about each QUADAS-2 domain presented as percentages across the 82 included studies in ophthalmic imaging (a), 115 in respiratory imaging (b) and 82 in breast imaging (c).

Discussion

This study sought to (1) quantify the diagnostic accuracy of DL algorithms to identify specific pathology across distinct radiological modalities, and (2) appraise the variation in study reporting of DL-based radiological diagnosis. The findings of our speciality-specific meta-analysis suggest that DL algorithms generally have a high and clinically acceptable diagnostic accuracy in identifying disease. High diagnostic accuracy with analogous DL approaches was identified in all specialities despite different workflows, pathology and imaging modalities, suggesting that DL algorithms can be deployed across different areas in radiology. However, due to high heterogeneity and variance between studies, there is considerable uncertainty around estimates of diagnostic accuracy in this meta-analysis. In ophthalmology, the findings suggest features of diseases, such as DR, AMD and glaucoma can be identified with a high sensitivity, specificity and AUC, using DL on both RFP and OCT scans. In general, we found higher sensitivity, specificity, accuracy and AUC with DL on OCT scans over RFP for DR, AMD and glaucoma. Only sensitivity was higher for DR on RFP over OCT. In respiratory medicine, our findings suggest that DL has high sensitivity, specificity and AUC to identify chest pathology on CT scans and CXR. DL on CT had higher sensitivity and AUC for detecting lung nodules; however, we found a higher specificity, PPV and F1 score on CXR. For diagnosing cancer or lung mass, DL on CT had a higher sensitivity than CXR. In breast cancer imaging, our findings suggest that DL generally has a high diagnostic accuracy to identify breast cancer on mammograms, ultrasound and DBT. The performance was found to be very similar for these modalities. In MRI, however, the diagnostic accuracy was lower; this may be due to small datasets and the use of 2D images. The utilisation of larger databases and multiparametric MRI may increase the diagnostic accuracy[113]. Extensive variation in the methodology, data interpretability, terminology and outcome measures could be explained by a lack of consensus in how to conduct and report DL studies. The STARD-2015 checklist[114], designed for reporting of diagnostic accuracy studies is not fully applicable to clinical DL studies[115]. The variation in reporting makes it very difficult to formally evaluate the performance of algorithms. Furthermore, differences in reference standards, grader capabilities, disease definitions and thresholds for diagnosis make direct comparison between studies and algorithms very difficult. This can only be improved with well-designed and executed studies that explicitly address questions concerning transparency, reproducibility, ethics and effectiveness[116] and specific reporting standards for AI studies[115,117]. The QUADAS-2 (ref. [118]) assessment tool was used to systematically evaluate the risk of bias and any applicability concerns of the diagnostic accuracy studies. Although this tool was not designed for DL diagnostic accuracy studies, the evaluation allowed us to judge that a majority of studies in this field are at risk of bias or concerning for applicability. Of particular concern was the applicability of reference standards and patient selection. Despite our results demonstrating that DL algorithms have a high diagnostic accuracy in medical imaging, it is currently difficult to determine if they are clinically acceptable or applicable. This is partially due to the extensive variation and risk of bias identified in the literature to date. Furthermore, the definition of what threshold is acceptable for clinical use and tolerance for errors varies greatly across diseases and clinical scenarios[119].

Limitations in the literature

Dataset

There are broad methodological deficiencies among the included studies. Most studies were performed using retrospectively collected data, using reference standards and labels that were not intended for the purposes of DL analysis. Minimal prospective studies and only two randomised studies[109,120], evaluating the performance of DL algorithms in clinical settings were identified in the literature. Proper acquisition of test data is essential to interpret model performance in a real-world clinical setting. Poor quality reference standards may result in the decreased model performance due to suboptimal data labelling in the validation set[28], which could be a barrier to understanding the true capabilities of the model on the test set. This is symptomatic of the larger issue that there is a paucity of gold-standard, prospectively collected, representative datasets for the purposes of DL model testing. However, as there are many advantages to using retrospectively collected data, the resourceful use of retrospective or synthetic data with the use of labels of varying modality and quality represent important areas of research in DL[121].

Study methodology

Many studies did not undertake external validation of the algorithm in a separate test set and relied upon results from the internal validation data; the same dataset used to train the algorithm initially. This may lead to an overestimation of the diagnostic accuracy of the algorithm. The problem of overfitting has been well described in relation to machine learning algorithms[122]. True demonstration of the performance of these algorithms can only be assumed if they are externally validated on separate test sets with previously unseen data that are representative of the target population. Surprisingly, few studies compared the diagnostic accuracy of DL algorithms against expert human clinicians for medical imaging. This would provide a more objective standard that would enable better comparison of models across studies. Furthermore, application of the same test dataset for diagnostic performance assessment of DL algorithms versus healthcare professionals was identified in only select studies[13]. This methodological deficiency limits the ability to gauge the clinical applicability of these algorithms into clinical practice. Similarly, this issue can extend to model-versus-model comparisons. Specific methods of model training or model architecture may not be described well enough to permit emulation for comparison[123]. Thus, standards for model development and comparison against controls will be needed as DL architectures and techniques continue to develop and are applied in medical contexts.

Reporting

There was varying terminology and a lack of transparency used in DL studies with regards to the validation or test sets used. The term ‘validation’ was identified as being used interchangeably to either describe an external test set for the final algorithm or for an internal dataset that is used to fine tune the model prior to ‘testing’. Furthermore, the inconsistent terminology led to difficulties understanding whether an independent external test set was used to test diagnostic performance[13]. Crucially, we found broad variation in the metrics used as outcomes for the performance of the DL algorithms in the literature. Very few studies reported true positives, false positives, true negatives and false negatives in a contingency table as should be the minimum for diagnostic accuracy studies[114]. Moreover, some studies only reported metrics, such as dice coefficient, F1 score, competition performance metric and Top-1 accuracy that are often used in computer science, but may be unfamiliar to clinicians[13]. Metrics such as AUC, sensitivity, specificity, PPV and NPV should be reported, as these are more widely understood by healthcare professionals. However, it is noted that NPV and PPV are dependent on the underlying prevalence of disease and as many test sets are artificially constructed or balanced, then reporting the NPV or PPV may not be valid. The wide range of metrics reported also leads to difficulty in comparing the performance of algorithms on similar datasets.

Study strengths and limitations

This systematic review and meta-analysis statistically appraises pooled data collected from 279 studies. It is the largest study to date examining the diagnostic accuracy of DL on medical imaging. However, our findings must be viewed in consideration of several limitations. Firstly, as we believe that many studies have methodological deficiencies or are poorly reported, these studies may not be a reliable source for evaluating diagnostic accuracy. Consequently, the estimates of diagnostic performance provided in our meta-analysis are uncertain and may represent an over-estimation of the true accuracy. Secondly, we did not conduct a quality assessment for the transparency of reporting in this review. This was because current guidelines to assess diagnostic accuracy reporting standards (STARD-2015[114]) were not designed for DL studies and are not fully applicable to the specifics and nuances of DL research[115]. Thirdly, due to the nature of DL studies, we were not able to perform classical statistical comparison of measures of diagnostic accuracy between different imaging modalities. Fourthly, we were unable to separate each imaging modality into different subsets, to enable comparison across subsets and allow the heterogeneity and variance to be broken down. This was because our study aimed to provide an overview of the literature in each specific speciality, and it was beyond the scope of this review to examine each modality individually. The inherent differences in imaging technology, patient populations, pathologies and study designs meant that attempting to derive common lessons across the board did not always offer easy comparisons. Finally, our review concentrated on DL for speciality-specific medical imaging, and therefore it may not be appropriate to generalise our findings to other forms of medical imaging or AI studies.

Future work

For the quality of DL research to flourish in the future, we believe that the adoption of the following recommendations are required as a starting point.

Availability of large, open-source, diverse anonymised datasets with annotations

This can be achieved through governmental support and will enable greater reproducibility of DL models[124].

Collaboration with academic centres to utilise their expertise in pragmatic trial design and methodology[125]

Rather than classical trials, novel experimental and quasi-experimental methods to evaluate DL have been proposed and should be evaluated[126]. This may include ongoing evaluation of algorithms once in clinical practice, as they continue to learn and adapt to the population that they are implemented in.

Creation of AI-specific reporting standards

A major reason for the difficulties encountered in evaluating the performance of DL on medical imaging are largely due to inconsistent and haphazard reporting. Although DL is widely considered as a ‘predictive’ model (where TRIPOD may be applied) the majority of AI interventions close to translation currently published are predominantly in the field of diagnostics (with specifics on index tests, reference standards and true/false positive/negatives and summary diagnostic scores, centred directly in the domain of STARD). Existing reporting guidelines for diagnostic accuracy studies (STARD)[114], prediction models (TRIPOD)[127], randomised trials (CONSORT)[128] and interventional trial protocols (SPIRIT)[129] do not fully cover DL research due to specific considerations in methodology, data and interpretation required for these studies. As such, we applaud the recent publication of the CONSORT-AI[117] and SPIRIT-AI[130] guidelines, and await AI-specific amendments of the TRIPOD-AI[131] and STARD-AI[115] statements (which we are convening). We trust that when these are published, studies being conducted will have a framework that enables higher quality and more consistent reporting.

Development of specific tools for determining the risk of study bias and applicability

An update to the QUADAS-2 tool taking into account the nuances of DL diagnostic accuracy research should be considered.

Updated specific ethical and legal framework

Outdated policies need to be updated and key questions answered in terms of liability in cases of medical error, doctor and patient understanding, control over algorithms and protection of medical data[132]. The World Health Organisation[133] and others have started to develop guidelines and principles to regulate the use of AI. These regulations will need to be adapted by each country to fit their own political and healthcare context[134]. Furthermore, these guidelines will need to proactively and objectively evaluate technology to ensure best practices are developed and implemented in an evidence-based manner[135].

Conclusion

DL is a rapidly developing field that has great potential in all aspects of healthcare, particularly radiology. This systematic review and meta-analysis appraised the quality of the literature and provided pooled diagnostic accuracy for DL techniques in three medical specialities. While the results demonstrate that DL currently has a high diagnostic accuracy, it is important that these findings are assumed in the presence of poor design, conduct and reporting of studies, which can lead to bias and overestimating the power of these algorithms. The application of DL can only be improved with standardised guidance around study design and reporting, which could help clarify clinical utility in the future. There is an immediate need for the development of AI-specific STARD and TRIPOD statements to provide robust guidance around key issues in this field before the potential of DL in diagnostic healthcare is truly realised in clinical practice.

Methods

This systematic review was conducted in accordance with the guidelines for the ‘Preferred Reporting Items for Systematic Reviews and Meta-Analyses’ extension for diagnostic accuracy studies statement (PRISMA-DTA)[136].

Eligibility criteria

Studies that report upon the diagnostic accuracy of DL algorithms to investigate pathology or disease on medical imaging were sought. The primary outcome was various diagnostic accuracy metrics. Secondary outcomes were study design and quality of reporting.

Data sources and searches

Electronic bibliographic searches were conducted in Medline and EMBASE up to 3rd January 2020. MESH terms and all-field search terms were searched for ‘neural networks’ (DL or convolutional or cnn) and ‘imaging’ (magnetic resonance or computed tomography or OCT or ultrasound or X-ray) and ‘diagnostic accuracy metrics’ (sensitivity or specificity or AUC). For the full search strategy, please see Supplementary Methods 1. The search included all study designs. Further studies were identified through manual searches of bibliographies and citations until no further relevant studies were identified. Two investigators (R.A. and V.S.) independently screened titles and abstracts, and selected all relevant citations for full-text review. Disagreement regarding study inclusion was resolved by discussion with a third investigator (H.A.).

Inclusion criteria

Studies that comprised a diagnostic accuracy assessment of a DL algorithm on medical imaging in human populations were eligible. Only studies that stated either diagnostic accuracy raw data, or sensitivity, specificity, AUC, NPV, PPV or accuracy data were included in the meta-analysis. No limitations were placed on the date range and the last search was performed in January 2020.

Exclusion criteria

Articles were excluded if the article was not written in English. Abstracts, conference articles, pre-prints, reviews and meta-analyses were not considered because an aim of this review was to appraise the methodology, reporting standards and quality of primary research studies being published in peer-reviewed journals. Studies that investigated the accuracy of image segmentation or predicting disease rather than identification or classification were excluded.

Data extraction and quality assessment

Two investigators (R.A. and V.S.) independently extracted demographic and diagnostic accuracy data from the studies, using a predefined electronic data extraction spreadsheet. The data fields were chosen subsequent to an initial scoping review and were, in the opinion of the investigators, sufficient to fulfil the aims of this review. Data were extracted on (i) first author, (ii) year of publication, (iii) type of neural network, (iv) population, (v) dataset—split into training, validation and test sets, (vi) imaging modality, (vii) body system/disease, (viii) internal/external validation methods, (ix) reference standard, (x) diagnostic accuracy raw data—true and false positives and negatives, (xi) percentages of AUC, accuracy, sensitivity, specificity, PPV, NPV and other metrics reported. Three investigators (R.A., V.S. and GM) assessed study methodology using the QUADAS-2 checklist to evaluate the risk of bias and any applicability concerns of the studies[118].

Data synthesis and analysis

A bivariate model for diagnostic meta-analysis was used to calculate summary estimates of sensitivity, specificity and AUC data[137]. Independent proportion and their differences were calculated and pooled through DerSimonian and Laird random-effects modelling[138]. This considered both between-study and within-study variances that contributed to study weighting. Study-specific estimates and 95% CIs were computed and represented on forest plots. Heterogeneity between studies was assessed using I (25–49% was considered to be low heterogeneity, 50–74% was moderate and >75% was high heterogeneity). Where raw diagnostic accuracy data were available, the SROC model was used to evaluate the relationship between sensitivity and specificity[139]. We utilised Stata version 15 (Stata Corp LP, College Station, TX, USA) for all statistical analyses. We chose to appraise the performance of DL algorithms to identify individual disease or pathology patterns on different imaging modalities in isolation, e.g., identifying lung nodules on a thoracic CT scan. We felt that combining imaging modalities and diagnoses would add heterogeneity and variation to the analysis. Meta-analysis was only performed where there were greater than or equal to three patient cohorts, reporting for each specific pathology and imaging modality. This study is registered with PROSPERO, CRD42020167503.
  131 in total

1.  Fusing Results of Several Deep Learning Architectures for Automatic Classification of Normal and Diabetic Macular Edema in Optical Coherence Tomography.

Authors:  Genevieve C Y Chan; Ravi Kamble; Henning Muller; Syed A A Shah; T B Tang; Fabrice Meriaudeau
Journal:  Annu Int Conf IEEE Eng Med Biol Soc       Date:  2018-07

2.  Meta-analysis in clinical trials.

Authors:  R DerSimonian; N Laird
Journal:  Control Clin Trials       Date:  1986-09

3.  Development and Validation of a Deep Learning System for Diabetic Retinopathy and Related Eye Diseases Using Retinal Images From Multiethnic Populations With Diabetes.

Authors:  Daniel Shu Wei Ting; Carol Yim-Lui Cheung; Gilbert Lim; Gavin Siew Wei Tan; Nguyen D Quang; Alfred Gan; Haslina Hamzah; Renata Garcia-Franco; Ian Yew San Yeo; Shu Yen Lee; Edmund Yick Mun Wong; Charumathi Sabanayagam; Mani Baskaran; Farah Ibrahim; Ngiap Chuan Tan; Eric A Finkelstein; Ecosse L Lamoureux; Ian Y Wong; Neil M Bressler; Sobha Sivaprasad; Rohit Varma; Jost B Jonas; Ming Guang He; Ching-Yu Cheng; Gemmy Chui Ming Cheung; Tin Aung; Wynne Hsu; Mong Li Lee; Tien Yin Wong
Journal:  JAMA       Date:  2017-12-12       Impact factor: 56.272

4.  AOCT-NET: a convolutional network automated classification of multiclass retinal diseases using spectral-domain optical coherence tomography images.

Authors:  Ali Mohammad Alqudah
Journal:  Med Biol Eng Comput       Date:  2019-11-14       Impact factor: 2.602

5.  The state of artificial intelligence-based FDA-approved medical devices and algorithms: an online database.

Authors:  Stan Benjamens; Pranavsingh Dhunnoo; Bertalan Meskó
Journal:  NPJ Digit Med       Date:  2020-09-11

6.  Classification of breast cancer in ultrasound imaging using a generic deep learning analysis software: a pilot study.

Authors:  Anton S Becker; Michael Mueller; Elina Stoffel; Magda Marcon; Soleen Ghafoor; Andreas Boss
Journal:  Br J Radiol       Date:  2018-01-10       Impact factor: 3.039

7.  Grader Variability and the Importance of Reference Standards for Evaluating Machine Learning Models for Diabetic Retinopathy.

Authors:  Jonathan Krause; Varun Gulshan; Ehsan Rahimy; Peter Karth; Kasumi Widner; Greg S Corrado; Lily Peng; Dale R Webster
Journal:  Ophthalmology       Date:  2018-03-13       Impact factor: 12.079

8.  Assessment of Convolutional Neural Networks for Automated Classification of Chest Radiographs.

Authors:  Jared A Dunnmon; Darvin Yi; Curtis P Langlotz; Christopher Ré; Daniel L Rubin; Matthew P Lungren
Journal:  Radiology       Date:  2018-11-13       Impact factor: 29.146

9.  STARD 2015: an updated list of essential items for reporting diagnostic accuracy studies.

Authors:  Patrick M Bossuyt; Johannes B Reitsma; David E Bruns; Constantine A Gatsonis; Paul P Glasziou; Les Irwig; Jeroen G Lijmer; David Moher; Drummond Rennie; Henrica C W de Vet; Herbert Y Kressel; Nader Rifai; Robert M Golub; Douglas G Altman; Lotty Hooft; Daniël A Korevaar; Jérémie F Cohen
Journal:  BMJ       Date:  2015-10-28

10.  Using artificial intelligence to read chest radiographs for tuberculosis detection: A multi-site evaluation of the diagnostic accuracy of three deep learning systems.

Authors:  Zhi Zhen Qin; Melissa S Sander; Bishwa Rai; Collins N Titahong; Santat Sudrungrot; Sylvain N Laah; Lal Mani Adhikari; E Jane Carter; Lekha Puri; Andrew J Codlin; Jacob Creswell
Journal:  Sci Rep       Date:  2019-10-18       Impact factor: 4.379

View more
  32 in total

1.  Impact of deep learning architectures on accelerated cardiac T1 mapping using MyoMapNet.

Authors:  Amine Amyar; Rui Guo; Xiaoying Cai; Salah Assana; Kelvin Chow; Jennifer Rodriguez; Tuyen Yankama; Julia Cirillo; Patrick Pierce; Beth Goddu; Long Ngo; Reza Nezafat
Journal:  NMR Biomed       Date:  2022-07-14       Impact factor: 4.478

2.  Development of deep learning-assisted overscan decision algorithm in low-dose chest CT: Application to lung cancer screening in Korean National CT accreditation program.

Authors:  Sihwan Kim; Woo Kyoung Jeong; Jin Hwa Choi; Jong Hyo Kim; Minsoo Chun
Journal:  PLoS One       Date:  2022-09-29       Impact factor: 3.752

3.  Protocol for a systematic review and meta-analysis of the diagnostic accuracy of artificial intelligence for grading of ophthalmology imaging modalities.

Authors:  Jessica Cao; Brittany Chang-Kit; Glen Katsnelson; Parsa Merhraban Far; Elizabeth Uleryk; Adeteju Ogunbameru; Rafael N Miranda; Tina Felfeli
Journal:  Diagn Progn Res       Date:  2022-07-14

4.  Evaluation of AI-Based Segmentation Tools for COVID-19 Lung Lesions on Conventional and Ultra-low Dose CT Scans.

Authors:  Marco Aiello; Dario Baldi; Giuseppina Esposito; Marika Valentino; Marco Randon; Marco Salvatore; Carlo Cavaliere
Journal:  Dose Response       Date:  2022-04-06       Impact factor: 2.658

5.  Artificial intelligence in orthopaedics: A scoping review.

Authors:  Simon J Federer; Gareth G Jones
Journal:  PLoS One       Date:  2021-11-23       Impact factor: 3.240

Review 6.  A Comprehensive Survey on Deep-Learning-Based Breast Cancer Diagnosis.

Authors:  Muhammad Firoz Mridha; Md Abdul Hamid; Muhammad Mostafa Monowar; Ashfia Jannat Keya; Abu Quwsar Ohi; Md Rashedul Islam; Jong-Myon Kim
Journal:  Cancers (Basel)       Date:  2021-12-04       Impact factor: 6.639

7.  A High-Performance Deep Neural Network Model for BI-RADS Classification of Screening Mammography.

Authors:  Kuen-Jang Tsai; Mei-Chun Chou; Hao-Ming Li; Shin-Tso Liu; Jung-Hsiu Hsu; Wei-Cheng Yeh; Chao-Ming Hung; Cheng-Yu Yeh; Shaw-Hwa Hwang
Journal:  Sensors (Basel)       Date:  2022-02-03       Impact factor: 3.576

8.  Artificial intelligence in orthopedic implant model classification: a systematic review.

Authors:  Mark Ren; Paul H Yi
Journal:  Skeletal Radiol       Date:  2021-08-05       Impact factor: 2.199

9.  Do comprehensive deep learning algorithms suffer from hidden stratification? A retrospective study on pneumothorax detection in chest radiography.

Authors:  Jarrel Seah; Cyril Tang; Quinlan D Buchlak; Michael Robert Milne; Xavier Holt; Hassan Ahmad; John Lambert; Nazanin Esmaili; Luke Oakden-Rayner; Peter Brotchie; Catherine M Jones
Journal:  BMJ Open       Date:  2021-12-07       Impact factor: 2.692

10.  Deep semi-supervised learning for automatic segmentation of inferior alveolar nerve using a convolutional neural network.

Authors:  Ho-Kyung Lim; Seok-Ki Jung; Seung-Hyun Kim; Yongwon Cho; In-Seok Song
Journal:  BMC Oral Health       Date:  2021-12-07       Impact factor: 2.757

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.