Literature DB >> 36147628

Artificial intelligence performance in image-based ovarian cancer identification: A systematic review and meta-analysis.

He-Li Xu1,2,3, Ting-Ting Gong4, Fang-Hua Liu1,2,3, Hong-Yu Chen1,2,3, Qian Xiao1, Yang Hou5, Ying Huang6, Hong-Zan Sun5, Yu Shi5, Song Gao4, Yan Lou7, Qing Chang1,2,3, Yu-Hong Zhao1,2,3, Qing-Lei Gao8, Qi-Jun Wu1,2,3,4.   

Abstract

Background: Accurate identification of ovarian cancer (OC) is of paramount importance in clinical treatment success. Artificial intelligence (AI) is a potentially reliable assistant for the medical imaging recognition. We systematically review articles on the diagnostic performance of AI in OC from medical imaging for the first time.
Methods: The Medline, Embase, IEEE, PubMed, Web of Science, and the Cochrane library databases were searched for related studies published until August 1, 2022. Inclusion criteria were studies that developed or used AI algorithms in the diagnosis of OC from medical images. The binary diagnostic accuracy data were extracted to derive the outcomes of interest: sensitivity (SE), specificity (SP), and Area Under the Curve (AUC). The study was registered with the PROSPERO, CRD42022324611. Findings: Thirty-four eligible studies were identified, of which twenty-eight studies were included in the meta-analysis with a pooled SE of 88% (95%CI: 85-90%), SP of 85% (82-88%), and AUC of 0.93 (0.91-0.95). Analysis for different algorithms revealed a pooled SE of 89% (85-92%) and SP of 88% (82-92%) for machine learning; and a pooled SE of 88% (84-91%) and SP of 84% (80-87%) for deep learning. Acceptable diagnostic performance was demonstrated in subgroup analyses stratified by imaging modalities (Ultrasound, Magnetic Resonance Imaging, or Computed Tomography), sample size (≤300 or >300), AI algorithms versus clinicians, year of publication (before or after 2020), geographical distribution (Asia or non Asia), and the different risk of bias levels (≥3 domain low risk or < 3 domain low risk). Interpretation: AI algorithms exhibited favorable performance for the diagnosis of OC through medical imaging. More rigorous reporting standards that address specific challenges of AI research could improve future studies. Funding: This work was supported by the Natural Science Foundation of China (No. 82073647 to Q-JW and No. 82103914 to T-TG), LiaoNing Revitalization Talents Program (No. XLYC1907102 to Q-JW), and 345 Talent Project of Shengjing Hospital of China Medical University (No. M0268 to Q-JW and No. M0952 to T-TG).
© 2022 The Author(s).

Entities:  

Keywords:  AI, Artificial intelligence; AUC, Area Under the Curve; Artificial intelligence; CT, Computed Tomography; DL, Deep learning; ML, Machine learning; MRI, Magnetic Resonance Imaging; Medical imaging; Meta-analysis; OC, Ovarian cancer; Ovarian cancer; SE, Sensitivity; SP, Specificity; US, Ultrasound; XAI, Explainable artificial intelligence

Year:  2022        PMID: 36147628      PMCID: PMC9486055          DOI: 10.1016/j.eclinm.2022.101662

Source DB:  PubMed          Journal:  EClinicalMedicine        ISSN: 2589-5370


Evidence before this study

The accurate preoperative differentiation between benign and malignant masses of the ovary is crucial for determining the appropriate treatment strategies and improving the postoperative quality of life. Imaging is an useful tool in medical science and is invoked in clinical practice to facilitate decision making for the diagnosis, staging, and treatment. The advances of artificial intelligence (AI) might help to bridge the gap between the intense demand for diagnostic from imaging and relatively limited healthcare resources. Up to date, there is a lack of quantitative synthesis to comprehensively summarize the available evidence of the AI-based methods on ovarian cancer (OC) detection. The Medline, Embase, IEEE, Pubmed, Web of Science, and the Cochrane library were systematically searched for studies that developed an AI algorithm for the diagnostic performance of OC from medical imaging, published until August 1, 2022. Only English language articles were considered. We performed a systematic review and meta-analysis of published data on diagnostic performance of AI algorithms and radiomics models for OC detection.

Added value of this study

To our best knowledge, this is the first systematic review and meta-analysis specifically dedicated to AI system performance in the diagnosis of OC. We are strictly in line with the guidelines for diagnostic reviews, and conducted a comprehensive literature search in both medical databases and engineering and technology databases to ensure the rigor of the study. After a careful selection of research on relevant topics, we found that AI algorithms excelled in the identification of OC using medical radiography imaging.

Implications of all the available evidence

AI algorithms exhibited favorable performance for the diagnosis of OC through medical imaging. More rigorous reporting standards that address specific challenges of AI research could improve future studies. Alt-text: Unlabelled box

Introduction

Ovarian tumors comprise a remarkably heterogeneous group of benign, borderline, and malignant lesions and exhibit extensive morphological characteristics., Among these, ovarian cancer (OC) is the most lethal gynecological malignancy. While malignant ovarian neoplasms may need a more aggressive surgical approach, benign masses can either be safely monitored or undergo simple resection allowing for a fertility- and ovary-sparing approach. Therefore, accurate preoperative differentiation between benign and malignant masses of the ovary is crucial for determining the appropriate treatment strategies and improving the postoperative quality of life. Imaging is a useful tool in medical science and is invoked in clinical practice to facilitate decision making for the diagnosis, staging, and treatment., The ultrasound (US) is commonly used to recognize the presence of an ovarian mass and to determine between benign and malignant lesions. Magnetic resonance imaging (MRI) plays a significant role in characterizing ovarian tumors due to its high soft-tissue resolution, and it is recommended in assessing the need for surgery for an adnexal mass. Computed tomography (CT) may be helpful for judging the gross extent of hematogenous, peritoneal, and lymphatic spread of OC: because of its ability to evaluate the liver, paraaortic region, omentum, and mesentery. While studies have reported the utility of PET CT in diagnosing ovarian tumors, its cost-effectiveness for this purpose remains unproven. Currently, US and MRI are the most commonly used imaging modalities for the diagnosis and characterization of ovarian tumors. Of note, the diagnosis of OC has been conventionally dependent on the subjective assessment of radiologists or gynecologists who use their clinical practice experience to scrutinize imaging features and examine ovarian tumors with high heterogeneity., Owing to the intricacy generated by inadequate or absent radiology in resource-poor health regions and the influence of wide disparity in the human rater expertise, making a proper and immediate diagnosis from medical imaging is challenging., The advances of artificial intelligence (AI) might help to bridge the gap between the intense demand for diagnostic from imaging and relatively limited healthcare resources. Meanwhile, as an interesting research hotspot, radiomics is described as a new 'data-driven' approach for extracting large sets of quantitative signatures from radiological images. These data can be subsequently analyzed using conventional biostatistics or AI methods. With sophisticated image processing methods, all medical images are transferred to mineable high-throughput image features, which thereafter can be used to correlate these processed feature signatures with pathology diagnoses or treatment responses. Radiomics models and AI algorithms have shown promising results in integrating medical images for the detection of OC. For example, aramedia-vidaurreta et al. emphasized that a machine learning (ML) algorithm based on US images achieved a diagnostic accuracy of 0.98 in one hundred and forty-five patients. Additionally, a deep learning (DL) model was used to automatically discriminate between benign and malignant ovarian tumor images, with high accuracy of 87.6%. Even though, researchers have still tried different ways, including but not limited to improving image quality, expanding sample sizes, and optimizing algorithms, to raise diagnostic accuracy. Up to date, there is a lack of quantitative synthesis to comprehensively summarize the available evidence of the AI-based methods on OC detection. Therefore, the purpose of this study is to first perform a systematic review and meta-analysis of published data on the diagnostic performance of AI algorithms and radiomics models for OC detection.

Methods

Protocol registration and study design

The study was registered in the PROSPERO (CRD42022324611). The meta-analysis was conducted following the PRISMA, MOOSE, and CHARMS reporting guidelines.

Search strategy and eligibility criteria

The Medline, Embase, IEEE, Pubmed, Web of Science, and the Cochrane library were systematically searched for studies that developed an AI algorithm for the diagnostic performance of OC from medical imaging, published until August 1, 2022. Only English-language articles were considered. Supplementary Note 1 summarizes the search strategy used in each database. Eligible studies reported the AI technologies for the diagnosis of OC from medical radiology images with diagnostic outcomes, such as sensitivity (SE), and specificity (SP), or detailed information on 2×2 contingency tables. The following studies were excluded: duplicate publications; reviews; editorials; non-human samples; histopathology images; combining non-image information; no classification task; and no AI model. Two reviewers (H-LX and F-HL) independently screened the titles and abstracts according to these eligibility criteria, and relevant articles for full text were downloaded and reviewed. Disagreement was discussed with a third author (Q-JW) and subsequently resolved via consensus.

Data extraction

Two reviewers (H-LX and H-YC) extracted study characteristics and diagnostic performance independently using a standardized data extraction sheet. Disagreements were resolved by discussion or a third investigator (F-HL) was consulted. The diagnostic accuracy data including true-positive (TP), false-positive (FP), true-negative (TN), and false-negative (FN) were extracted directly into contingency tables, and were used to calculate SE and SP. If a study provided multiple contingency tables for the same or different AI algorithms, we assumed that they were independent of each other. Supplementary Table 1 summarizes the contingency tables extracted from included studies.

Study quality assessment

All selected studies were assessed for quality with the use of quality assessment of diagnostic accuracy studies-AI (QUADAS-AI) criteria by two independent reviewers (H-LX and T-TG). The details are listed in Supplementary Table 2. This guideline includes four domains (patient selection, index test, reference standard, flow, and timing) in the risk of bias and three domains (patient selection, index test, reference standard) in applicability concerns. This new tool is an AI-specific extension to QUADAS-2 and QUADAS-C, providing researchers with a specific framework to evaluate the risk of bias and applicability when conducting reviews that evaluate AI-centered diagnostic test accuracy. Conflicts were discussed with a third collaborator (F-HL).

Meta-analysis

A hierarchical summary receiver-operating characteristic curve (SROC) was fitted to evaluate the accuracy of the AI model. We plotted the combined curve with corresponding 95% confidence region and 95% prediction region around averaged SE, SP, and Area Under the Curve (AUC) estimates in SROC figures. When same or different AI models were tested within the same paper, the proposed model with the best accuracy was used for further meta-analysis. Heterogeneity was assessed using the I2 statistic. Subgroup and regression analyses were performed to explore potential sources of heterogeneity. The random effects model was conducted because of the assumed differences between studies. The risk of publication bias was evaluated using funnel plot and regression test. Seven sub-analysis were performed: (1) according to sample size (≤300 or >300); (2) according to AI algorithms (ML or DL); (3) according to imaging modalities (CT, US, or MRI); (4) according to the pooled performance using the same dataset (AI algorithms or human clinicians); (5) according to the year of publication (before or after 2020); (6) according to the geographical distribution (Asia or non Asia); (7) according to different risk of bias levels (≥3 domain low risk or < 3 domain low risk) The methodological quality of included studies was evaluated using the QUADAS-AI by RevMan (Version 5.4). A cross-hairs plot was also produced (R V.4.2.1) to better display the variability between sensitivity/specificity estimates. All other statistical analyses were conducted in Stata software (Version 15.0) with two-tailed probability of type I error of 0.05 (α = 0.05).

Role of the funding source

Our study was funded by the Natural Science Foundation of China, the LiaoNing Revitalization Talents Program, and the 345 Talent Project of Shengjing Hospital of China Medical University. The funder of the study had no role in study design, data collection, data analysis, data interpretation, or writing of the report. The corresponding authors had full access to all the data in the study and had final responsibility for the decision to submit for publication.

Results

Study selection and characteristics of eligible studies

A total of 1212 records were retrieved on initial search and 513 duplicates were removed, and of these 642 studies were excluded based on screening of titles and abstracts, resulting in 57 studies for full-text review. Finally, 34 articles were included in the present systematic review and 28 had sufficient data for meta-analysis (Figure 1).
Figure 1

PRISMA flowchart of study selection.

PRISMA flowchart of study selection. Majority of the studies (n = 31) were based on retrospective patient data except four studies. Only two studies using prospective data. One studies used images from public databases. Eight studies excluded low quality images, while twenty-six studies did not mention this process. Only three studies using out-of-sample dataset to perform external validation, of which two studies did not provide data of our concern for integrated analysis. Eight studies compared AI model with clinicians in the same dataset. Moreover, imaging modalities were classified as US (n  =  19), MRI (n  =  10), CT (n  =  3), MRI and US (n  =  1), and MRI and CT (n  =  1). Furthermore, the distribution of the number of studies on AI algorithms in the present study is as follows: DL (11 studies) and ML (23 studies). Table 1, Table 2, Table 3, Table 4 show the detailed characteristics of these including studies.
Table 1

Participant demographics for the 35 included studies.

Author [ref], yearParticipants
NMean or median age (SD; range)
Inclusion criteriaExclusion criteria
Liu et al,31 2022aPatients with no previous pelvic surgery; patients with no previous gynecological disease history; patients who had MRI examinations performed at our institution before pelvic or laparoscopic surgery.Patients with previous pelvic surgical history or radiation history; patients whose MRI data were unavailable either due to the examination being performed at another institution or due to claustrophobia; patients whose data lacked histological results.19646.3
Gao et al,22 2022aConsecutive adult patients (aged ≥18 years) who presented with adnexal lesions in ultrasound in ten hospitals between September 2003, and May 2019.Duplicated cases; postoperative patients who were deprived of adnexa; patients without histological diagnosis.1,07,624NR
Saida et al,32 2022aAged above 20 years for ethical reasons; pelvic MRI scan obtained as per the protocol followed at our hospital between January 2015 and December 2020; pathologically proven malignant epithelial tumors (i.e., carcinomas) or borderline tumors of the ovary for the malignant group; pathologically proven or clinically apparent benign lesions in the non-malignant group.Malignant tumors in the pelvis other than the ovary; history of surgery of the uterus or ovaries other than caesarean section, chemotherapy, or radiation therapy of the pelvis; malignant ovarian epithelial tumors mixed with non-epithelial components.46550 (20–90)
Guo et al,33 2022aDefinite pathological diagnosis after operation; MRI and ultrasound were performed and the data were complete; the images could be used for diagnostic analysis; patient informed consent.Incomplete ultrasound, MRI, or pathological data; combined with severe organic diseases, such as coagulation dysfunction, renal insufficiency, heart failure, and other surgical contraindications; history of ovarian surgery; combined with other pelvic diseases, such as endometrial cancer and rectal cancer.207NR
Li et al,34 2022aPatients with ovarian tumor confirmed by histopathology; no history of malignant tumors other than ovarian tumor; patients who were undergoing pelvic CT examination within half a month before surgery.Those who had received radiotherapy, chemotherapy, or radiotherapy–chemotherapy before CT examination; patients diagnosed with inflammatory diseases; patients with low image quality.140NR
Wang et al,35 2021aA histologic diagnosis of benign, borderline, or malignant SOTs between March 2013 and December 2016; availability of diagnostic-quality preoperative US images; US scanning before neoadjuvant therapy or surgical resection.No ultrasound results or the ovarian mass was not completely in the images; mucinous, clear cell, endometrioid, or metastatic cancer.26551 (15–79)
Chiappa et al,36 2021aDiagnosis of OM; execution of a preoperative ultrasonographic examination within 2 weeks before surgery; surgery performed.Age<18 years; absence of ultrasonographic images stored; consent withdrawn.24155 (18–84)
Jian et al,37 2021All patients were histopathologically proven to have either BEOT (n = 165) or MEOT (n = 336).NR501NR
Wang et al,38 2021aBenign or malignant ovarian lesions confirmed by either pathology or imaging follow-up; available preoperative MRI examination including T1C and T2WI; the quality of images was clear without motion or artifacts and were fit for analysis.Lack pre-operative MRI; lack clear ovarian lesion; lack T1C images.45145.7
Hu et al,39 2021aNRPatients with poor image quality; patients without enhanced scanning; patients with unclear boundary and unable to outline110NR
Yu et al,40 2021aSBOTs and SMOTs were diagnosed by postoperative pathology; SBOTs and SMOTs were in an early stage (I and II) according to the guideline of the FIGO; the images were of sufficient quality for radiomics analysis.SBOTs and SMOTs which were in a late stage (III and IV) according to the FIGO guideline; patients who received any treatment before CT examination or were on treatment at the time of CT examination were also excluded to eliminate the effect of treatment on imaging features.18247.7
Ștefan et al,41 2021aA lesion with a minimum diameter of at least 20 mm; the availability of conventional B-mode images; lack of imaging artifacts; and the existence of a patient's serial number.No medical data corresponding to the PSN; the absence of a final pathological diagnosis to indicate the benign or malignant nature of the lesions; the pathological analysis performed at more than 30 days after the image acquisition; and no gynecological follow-up.12038.2
Christiansen et al,42 2021aSurgery within 120 days after the ultrasound examination or ultrasound follow-up for a minimum of 3 years or until resolution of the lesion.NR758NR
Akazawa et al,43 2020Patients were ovarian tumors which had been diagnosed pathologically after surgical resection.Lack of sufficient preoperative clinical data, such as tumor markers or the records of imaging tests.20251 (14–84)
Martínez et al,44 2019aNRNR384NR
Zhang et al,20 2019aNo previous pelvic surgery; no previous gynecological disease history; MRI examinations before pelvic or laparoscopic surgery were performed at our institution.Previous pelvic surgical history or radiation history; MRI data were unavailable either for the examination performed at another institution or due to claustrophobia; no histological results.43852.7
Mol et al,45 2001aWomen who had surgery for an adnexal mass between January 1991 and December 1998 were included.NR17046 (20–89)
Liu D et al,46 2017aPatients with histologically proven diagnosis of EOCs; patients complete CT or MRI examination before operation in two weeks.Surgery was performed outside our institution without definite histological diagnosis, incomplete clinical or CT and MRI records preoperatively.6556.4
Kazerooni et al,47 2017aPatients were scheduled for surgical removal of suspicious ovarian masses and postoperative histopathological assessment within 2 weeks of MRI exam.NR5538.4
Acharya et al,48 2014aNRWomen with no anatomopathological evaluation.2049.5
Acharya et al,49 2013aNRPatients with no anatomopathological evaluation.2049.5
Acharya et al,50 2012aNRNR2049.5
Umar et al,51 2012NRNR24NR
Acharya et al,52 2012aNRPatients with no anatomopathological evaluation.2049.5
Al-Karawi et al,53 2021aAll ovarian tumors were given a histological diagnosis label.NR232NR
Jian et al,54 2021Histologically proven EOC; MRI performed within 1 month prior to gynecological operation; all four axial MRI sequences obtained: fast spin-echo T2-weighted imaging with fat saturation(T2WI FS), echo-planar DWI with gradient b factors of 0 and 600, 800, or 1000 s/mm2, ADC map, and 2D volumetric interpolated breath hold examination (VIBE) contrast enhanced T1-weighted imaging with FS (CE-T1WI) in the late phase (150–190 s after the intravenous administration of contrast agent); absence of prior gynecological operation or chemotherapy prior to MRI scanning.Patients without definitive histopathology or with poor MRI image quality (image has artifacts that cannot outline the tumor).294(51.2–57.2)
Li et al,55 2020Histologically proven BEOT or MEOT from January 2010 to June 2018; MRI performed within 2 weeks prior to gynecological operation.Lacking any one of these four axial MRI sequences; prior gynecological operation and/or chemotherapy before MRI scanning; poor MRI image quality with artifacts that affected the delineation of the tumor.501(47.2–51.6)
Acharya et al,56 2014aNRNR20NR
Pathak et al,57 2015aNRNR120NR
Ameye et al,58 2009aNRExclusion criteria were pregnancy, inability to tolerate transvaginalsonography, and surgery performed more than 120 days after sonographic assessment.157346 (9–94)
Jian et al,59 2022aInclusion criteria were as follows: patients with 1) BEOT or MEOT that was proven by surgery and histopathology from January 2010 to June 2018; 2) an MRI performed within 2 weeks before gynecological operation which included the following three axial MRI sequences: fast spin echo T2-weighted imaging with fat saturation (T2WI FS), echo planar diffusion-weighted imaging (DWI) with apparent diffusion coefficient (ADC) maps generated from maximum b-value imaging if images with multiple b-values available, and 2D volumetric interpolated breath-hold examination of contrast-enhanced T1-weighted imaging (CE-T1WI) with FS in the late phase (150– 190 seconds after the intravenous administration of contrast agent); and 3) no history of gynecological operations or chemotherapy prior to the MRI scan.Patients with poor quality images were excluded (based on the evaluation of the radiologist with 10 years’ experience in gynecological imaging) because artifacts could affect the observation of the tumor.50158.92 (14.05)
Alqasemi et al,51 2012aNRNR24NR
Chen et al,60 2012aInclusion criteria were as follows: patients with at least one persisting ovarian tumor detected at US (except for physiologic cysts) from January 2019 to November 2019, patients who underwent a surgical procedure with histopathologic results, an interval of 30 days between US examination and surgery, and patients who had no previous history of ovarian cancer.Exclusion criteria were histopathologic analysis–confirmed uterine sarcomas or nongynecologic tumors, inconclusive histopathologic results, or poor US image quality.42246.4 (14.8)
Zheng et al,61 2022Patients with either SBOTs or SMOTs, who underwent preoperativeMRI scans and confirmed by postoperative pathology.Exclusion criteria were as follows: (1) solid tissue <80% in lesion (25); (2) the tumor had significant metastases; (3) significant image artifacts.126061 (20–79)

Abbreviation: BEOT: borderline epithelial ovarian tumor; CT: computed tomography; EOC: epithelial ovarian cancer; FIGO: International Federation of Gynecology and Obstetrics; MEOT: malignant epithelial ovarian tumors; NR=not reported; MRI: magnetic resonance imaging; OM: ovarian mass; SBOT; serous borderline ovarian tumors; SMOT: serous malignant ovarian tumors; SOT: serous ovarian tumors; T1C: T1-weighted contrast-enhanced sequence; T2WI: T2-weighted sequence; US: ultrasound.

Studies (n = 28) included in the meta-analysis.

Table 2

Model training and validation for the 35 included studies.

Author [ref], yearReference standardType of internal validationExternal validationAI versus clinicians
Liu et al,31 2022aHistopathologyNRNoNo
Gao et al,22 2022aHistopathologyRandom split sample validationYesYes
Saida et al,32 2022aHistopathologyNRNoYes
Guo et al,33 2022aHistopathologyK-fold cross validationNoNo
Li et al,34 2022aHistopathologyTen-fold cross-validationNoNo
Wang et al,35 2021aHistopathologyThree-fold cross validationNoNo
Chiappa et al,36 2021aHistopathologyTen-fold cross validationNoNo
Jian et al,37 2021HistopathologyRandom split sample validationNoNo
Wang et al,38 2021aHistopathologyCross validationNoYes
Hu et al,39 2021aNRTen-fold cross-validationNoNo
Yu et al,40 2021aHistopathologyNRNoNo
Ștefan et al,41 2021aHistopathologyNRNoNo
Christiansen et al,42 2021aHistopathologyNRNoYes
Akazawa et al,43 2020HistopathologyK-fold cross validationNoNo
Zhang et al, 2019aHistopathologyTen-fold cross validationNoNo
Martínez et al,44 2019aHistopathologyCross validationNoNo
Zhang et al,20 2019aHistopathologyLeave-one-out cross-validationNoYes
Mol et al,45 2001aHistopathologyCross validationNoNo
Liu D et al,46 2017aHistopathologyCross validationNoNo
Kazerooni et al,47 2017aHistopathologyLeave-one-out cross-validationNoNo
Acharya et al,48 2014aHistopathologyTen-fold cross validationNoNo
Acharya et al,49 2013aHistopathologyTen-fold cross validationNoNo
Acharya et al,50 2012aNRK-fold cross validationNoNo
Umar et al,51 2012HistopathologyNRNoNo
Acharya et al,52 2012aHistopathologyTen-fold cross validationNoNo
Al-Karawi et al,53 2021aHistopathologyRandom split sample validationNoNo
Jian et al,54 2021HistopathologyNRYesYes
Li et al,55 2020HistopathologyNRYesYes
Acharya et al,56 2014aNRTen-fold cross validationNoNo
Pathak et al,57 2015aNRCross validationNoNo
Ameye et al,58 2009aHistopathologyNRNoYes
Jian et al,59 2022aHistopathologyNRNoNo
Alqasemi et al,51 2012aHistopathologyNRNoNo
Chen et al,60 2012aHistopathologyNRNoYes
Zheng et al,61 2022HistopathologyTen-fold cross validationNoNo

Abbreviation: AI: artificial intelligence; NR=not reported.

Studies (n = 28) included in the meta-analysis.

Table 3

Indicator, algorithm, and data source for the 35 included studies.

Author [ref], yearIndicator definition
Algorithm
DeviceExclusion of poor-quality imagingHeatmap providedAlgorithm architectureML/DLTransfer learning applied
Liu et al,31 2022aMRINRNoLASSOMLNo
Gao et al,22 2022aUSYesNoDCNNDLNo
Saida et al,32 2022aMRINRYesCNNDLNo
Guo et al,33 2022aMRI, USNRNoLRMLNo
Li et al,34 2022aCTYesNoLRMLNo
Wang et al,35 2021aUSNRYesDCNNDLNo
Chiappa et al,36 2021aUSNRNoSVMMLNo
Jian et al,37 2021MRINRNoMAC-NetDLNo
Wang et al,38 2021aMRIYesNoCNNDLNo
Hu et al,39 2021aCTYesNoLRMLNo
Yu et al,40 2021aCTYesYesSVMMLNo
Ștefan et al,41 2021aUSNRNoKNNMLNo
Christiansen et al,42 2021aUSNRNoDNNDLNo
Akazawa et al,43 2020USNRNoSVM, KNN, RF, NB, XGBoostMLNo
Martínez et al,44 2019aUSNRNoKNN, LD, SVM, ELMMLNo
Zhang et al,20 2019aMRINRNoLASSOMLNo
Mol et al,45 2001aUSNRNoLR, NNMLNo
Liu D et al,46 2017aCT, MRINRNoRFMLNo
Kazerooni et al,47 2017aMRINRNoSVM, LDADLNo
Acharya et al,48 2014aUSNRNoPNNMLNo
Acharya et al,49 2013aUSNRNoDTMLNo
Acharya et al,50 2012aUSNRNoSVMMLNo
Umar et al,51 2012USNRNoSVMMLNo
Acharya et al,52 2012aUSNRNoDTMLNo
Al-Karawi et al,53 2021aUSNRNoSVMMLNo
Jian et al,54 2021MRIYesNoLASSOMLNo
Li et al,55 2020MRINRNoLRMLNo
Acharya et al,56 2014aUSNRNoPNNMLNo
Pathak et al,57 2015aUSNRNoSVMMLNo
Ameye et al,58 2009aUSNRNoLRMLNo
Jian et al,59 2022aMRIYesNoMICNNDLNo
Alqasemi et al,51 2012aUSNRNoSVMMLNo
Chen et al,60 2012aUSYesNoResNetDLNo
Zheng et al,61 2022MRINRNoLASSOMLNo

Abbreviation: AI: artificial intelligence; CNN: convolutional neural network; CT: computed tomography; DCNN: deep convolutional neural network; DL: deep learning; DT: decision tree; DNN: deep neural network; ELM: extreme learning machine; KNN: k-nearest neighbor; LASSO: least absolute shrinkage and selection operator method; LD: linear discriminant; LR: logistic regression; ML: machine learning; MRI: magnetic resonance imaging; NB: naïve bayes; NR=not reported; PNN: probabilistic neural networks; RF: random forest; SVM: support vector machine; US: ultrasound.

Studies (n = 28) included in the meta-analysis.

Table 4

Data source for the 35 included studies.

Author [ref], yearSource of dataNumber of images for training/ /testingData rangeOpen access data
Liu et al,31 2022aRetrospective study, data from Gynecological and Obstetric Hospital, School of Medicine, Fudan University, Shanghai, China.99/972014.01–2017.12No
Gao et al,22 2022aRetrospective study, data from Tongji Hospital, Tongji Medical College, Huazhong University of Science and Technology, and seven other hospitals, Jingzhou First People's Hospital and Xiangyang Central Hospital.575930/8416/79292003.09–2019.05No
Saida et al,32 2022aRetrospective study, data from Faculty of Medicine, University of Tsukuba.3663/1002015.01–2020.12No
Guo et al,33 2022aRetrospective study, data from Qilu Hospital.138/692018.04–2021.04No
Li et al,34 2022aRetrospective study, data from the First Affiliated Hospital of Nanchang Medical College.99/412017–2020No
Wang et al,35 2021aRetrospective study, data from Tianjin Medical University Cancer Institute and Hospital.195/842013.03–2016.12No
Chiappa et al,36 2021aRetrospective study, data from Fondazione IRCCS Istituto Nazionale dei Tumori di Milano.NR2017.01–2019.12No
Jian et al,37 2021Retrospective, data from eight clinical centers in china.282/119NRNo
Wang et al,38 2021aRetrospective study, data from one large academic center in the United States.384/161NRNo
Hu et al,39 2021aRetrospective study, data from Lishui Hospital of Zhejiang University76/342010.01–2018.12No
Yu et al,40 2021aRetrospective study, data from the Affiliated Hospital of Qingdao University.127/552017.12–2020.06No
Ștefan et al,41 2021aRetrospective study, data from University of Medicine and PharmacyNR2017.10–2019.02No
Christiansen et al,42 2021aRetrospective study, data from the Karolinska University Hospital(tertiary referral center)and Sodersjukhuset (secondary/tertiary referral center) in Stockholm, Sweden.508/2502010–2019No
Akazawa et al,43 2020Prospective study, date from Tokyo Women's Medical University Medical Center East.141/612013.12–2019.01No
Martínez et al,44 2019aRetrospective study, data from the University Hospital of the Catholic University of Leuven.NRNRNo
Zhang et al,20 2019aRetrospective study, data from Gynecological and Obstetric Hospital, School of Medicine, Fudan University, Shanghai, China.NR2014.01–2017.12No
Mol et al,45 2001aProspective study, data from in the Saint Joseph Hospital in Veldhoven.NR1991.01–1998.12No
Liu D et al,46 2017aRetrospective study, date from Department of Radiology, Shanghai Tenth People's hospital of Tongji University.NR2009.01–2015.10No
Kazerooni et al,47 2017aProspectively study, NR.NRNRNo
Acharya et al,48 2014aRetrospective study, NR.2340/260NRNo
Acharya et al,49 2013aRetrospective study, NR.1800/200NRNo
Acharya et al,50 2012aRetrospective study, NR.1800/200NRNo
Umar et al,51 2012Retrospective study, NR.NRNRNo
Acharya et al,52 2012aRetrospective study, NR.1800/200NRNo
Al-Karawi et al,53 2021aRetrospective study, data from the IOTA research.150/14874/762005.11–2013.11No
Jian et al,54 2021Retrospective study, eight centers.144/75/752010.01–2019.02No
Li et al,55 2020Retrospective study, NR.250/92/1592010.01–2018.06No
Acharya et al,56 2014aRetrospective study, NR.2340/260NRNo
Pathak et al,57 2015aRetrospective study, NR.70/50NRNo
Ameye et al,58 2009aRetrospective study, data from the IOTA research.754/5071999–2006No
Jian et al,59 2022aRetrospective study, NR.342/15920102018No
Alqasemi et al,51 2012aRetrospective study, NR.400/95NRYes
Chen et al,60 2012 aRetrospective study, data from the Ruijin Hospital affiliated with Shanghai Jiaotong university School of Medicine.296/41/852019.01–2019.11No
Zheng et al,61 2022Retrospective study, data from the Tianjin Medical University General Hospital from November 2010 to May 2020.125/312010–2020No

Studies (n = 28) included in the meta-analysis.

Participant demographics for the 35 included studies. Abbreviation: BEOT: borderline epithelial ovarian tumor; CT: computed tomography; EOC: epithelial ovarian cancer; FIGO: International Federation of Gynecology and Obstetrics; MEOT: malignant epithelial ovarian tumors; NR=not reported; MRI: magnetic resonance imaging; OM: ovarian mass; SBOT; serous borderline ovarian tumors; SMOT: serous malignant ovarian tumors; SOT: serous ovarian tumors; T1C: T1-weighted contrast-enhanced sequence; T2WI: T2-weighted sequence; US: ultrasound. Studies (n = 28) included in the meta-analysis. Model training and validation for the 35 included studies. Abbreviation: AI: artificial intelligence; NR=not reported. Studies (n = 28) included in the meta-analysis. Indicator, algorithm, and data source for the 35 included studies. Abbreviation: AI: artificial intelligence; CNN: convolutional neural network; CT: computed tomography; DCNN: deep convolutional neural network; DL: deep learning; DT: decision tree; DNN: deep neural network; ELM: extreme learning machine; KNN: k-nearest neighbor; LASSO: least absolute shrinkage and selection operator method; LD: linear discriminant; LR: logistic regression; ML: machine learning; MRI: magnetic resonance imaging; NB: naïve bayes; NR=not reported; PNN: probabilistic neural networks; RF: random forest; SVM: support vector machine; US: ultrasound. Studies (n = 28) included in the meta-analysis. Data source for the 35 included studies. Studies (n = 28) included in the meta-analysis.

Pooled performance of AI algorithms

The SROC curves for 28 included studies with 160 contingency tables are provided in Figure 2a, the combined SE and SP were 88% (95%CI: 85–90%) and 85% (82–88%), respectively, with an AUC of 0.93 (0.91–0.95) for all AI algorithms. When the highest accuracy contingency table was selected from these 28 studies, the pooled SE and SP were the same as 91% (84–95%) and 94% (89–97%), respectively (Figure 2b). A cross hairs plot shows reported point estimates and confidence intervals in Figure 3.
Figure 2

(a, b). SROC curves of all studies included in the meta-analysis (28 studies). a: SROC curves of ll studies included in the meta-analysis (28 studies with 160 tables). b: SROC curves of studies when selecting contingency tables reporting the highest accuracy (28 studies with 28 tables).

Abbreviations: AI: artificial intelligence; SROC = summary receiver operating characteristic; SENS = summary sensitivity; SPEC = summary specificity.

Figure 3

Cross-hair Plot of all studies included in the meta-analysis (28 studies with 160 tables).

(a, b). SROC curves of all studies included in the meta-analysis (28 studies). a: SROC curves of ll studies included in the meta-analysis (28 studies with 160 tables). b: SROC curves of studies when selecting contingency tables reporting the highest accuracy (28 studies with 28 tables). Abbreviations: AI: artificial intelligence; SROC = summary receiver operating characteristic; SENS = summary sensitivity; SPEC = summary specificity. Cross-hair Plot of all studies included in the meta-analysis (28 studies with 160 tables).

Quality assessment

The quality of included studies was determined by the QUADAS-AI (Supplementary figure 1). The detailed assessment results are presented with a diagram in Supplementary figure 2. Over half of the studies showed a high risk or an unclear risk of bias respectively for patient selections (n = 23) and index test (n = 31) because these studies did not clarify description of included patients detailing previous testing, presentation, setting, the intended use of the index test and lack of adequate external evaluation.

Subgroup meta-analyses

Considering the stage of development of the algorithm and the difference in nature, we categorized them into ML and DL algorithms and did a sub-analysis. The results demonstrated a pooled SE of 89% (95%CI: 85–92%) for ML and 88% (95%CI: 84–91%) for DL, and a pooled SP of 88% (95%CI: 82–92%) for ML and 84% (95%CI: 80–87%) for DL (Supplementary figure 3a, b). Seventeen US studies had a pooled SE of 91% (87–93%), a pooled SP of 87% (82–91%), and with an AUC of 0.95 (0.93–0.97). Six MRI studies with a pooled SE of 83% (77–88%), pooled SP of 84% (80–87%), and an AUC of 0.90 (0.87–0.92). Three CT studies that had a pooled SE of 75% (68–81%), pooled SP of 75% (67–82%), and an AUC of 0.82 (0.78–0.85) (Supplementary figure 4a, b, c). Eight studies presented the diagnostic accuracy between AI algorithms and human clinicians in the same dataset. The pooled SE was 82% (77–87%) for AI algorithms, and human clinicians had 77% (73–80%). The pooled SP was 86% (83–89%) for AI algorithms, and 80% (75–84%) in human clinicians. The AUC was 0.91 (0.88–0.93) and 0.85 (0.81–0.88) for AI algorithms and human clinicians, respectively (Supplementary figure 5a, b). Fifteen studies had sample sizes ≤ 300 and thirteen studies had sample sizes > 300. The pooled SE was 85% (81–88%) for sample size ≤ 300, and 93% (89–95%) for sample size > 300. The SP for ≤ 300 was 82% (80–85%) and 91% (84–96%) for > 300. The AUC was 0.90 (0.87–0.92) for ≤ 300 and 0.97 (0.95–0.98) for > 300 (Supplementary figure 6a, b). Thirteen studies were published before 2020. Fifteen studies were published after 2020. The pooled SE was 89% (84–93%) for published before 2020, and 88 (85–90%) for published after 2020. The SP was 89% (83–93%) and 83% (80–85%), respectively. The AUC was 0.95 (0.93–0.97) and 0.92 (0.89–0.94), respectively ((Supplementary figure 7a, b). Fifteen studies were geographically distributed in Asia and thirteen studies were geographically distributed outside Asia. The pooled SE was 87% (84–90%) and 90 (85–93%), respectively. The SP was 83% (80–86%) and 89% (82–93%), respectively. The AUC was 0.92 (0.89–0.94) and 0.95 (0.93–0.97), respectively (Supplementary figure 8a, b). There were ten studies with low risk in more than three evaluation domains and eighteen studies with high risk. The pooled SE was 86% (78–91%) and 89% (87–91%), respectively. The SP was 92% (88–95%) and 81% (76–85%), respectively. The AUC was 0.93 (0.90–0.95) and 0.93 (0.91–0.95), respectively (Supplementary figure 9a, b).

Heterogeneity analysis

The meta-analysis results of 28 studies suggested that AI algorithms were beneficial for the diagnosis of OC from medical imaging from random-effects model. However, there was substantial heterogeneity among the included studies, SE had an I2  =  94.68%, while SP had I2  =  97.50% (p < 0.01). The detailed results of subgroup and meta-regression analyses exploring the potential source of between-study heterogeneity are shown in Table. 5 and Supplementary figure 10-23. The results highlighted a statistically significant difference. Visual inspection of funnel plots suggested there was no publication bias (p  =  0.83) (Supplementary figure 24).
Table 5

Summary estimate of pooled performance of artificial intelligence in image-based ovarian cancer detection.

No. of studiesSensitivity
P valuebSpecificity
P valueb
SensitivityP value aI2 (95%CI)SpecificityP value aI2 (95%CI)
Overall280.88 (0.85–0.90)< 0.0594.68 (94.16–95.19)0.85 (0.82–0.88)< 0.0597.50 (97.31–97.69)
Algorithm< 0.05< 0.05
Machine learning190.89 (0.85–0.92)< 0.0595.11 (94.49–95.72)0.88 (0.82–0.92)< 0.0597.69 (97.46–97.92)
Deep learning90.88 (0.84–0.91)< 0.0595.48 (94.84–96.11)0.84 (0.80–0.87)< 0.0595.84 (95.28–96.41)
Imaging modality< 0.05< 0.05
Ultrasound170.91 (0.87–0.93)< 0.0596.58 (96.22–96.94)0.87(0.82–0.91)< 0.0598.55 (98.43–98.66)
Magnetic resonance imaging60.83 (0.77–0.88)< 0.0585.72 (82.32–89.12)0.84(0.80–0.87)< 0.0583.47 (79.37–87.58)
Computed tomography30.75 (0.68–0.81)0.430.00 (0.00–100.00)0.75 (0.67–0.82)0.830.00 (0.00–100.00)
Sample size< 0.05< 0.05
≤ 300150.85 (0.81–0.88)< 0.0591.75 (90.61–92.90)0.82 (0.80–0.85)< 0.0583.00 (80.08–84.93)
> 300130.93 (0.89–0.95)< 0.0597.96 (97.72–98.20)0.91 (0.84–0.96)< 0.0599.42 (99.38–99.47)
Risk of bias< 0.05< 0.05
Low100.86 (0.78–0.91)< 0.0597.49 (97.14–97.84)0.92 (0.88–0.95)< 0.0597.31 (96.92–97.69)
High180.89 (0.87–0.91)< 0.0591.78 (90.70–92.87)0.81 (0.76–0.85)< 0.0595.94 (95.51–96.37)
Geographical distribution< 0.05< 0.05
Asia130.87 (0.84–0.90)< 0.0594.48 (93.74–95.22)0.83 (0.80–0.86)< 0.0595.00 (94.35–95.65)
Non Asia150.90 (0.85–0.93)< 0.0596.36 (95.91–96.82)0.89 (0.82–0.93)< 0.0598.17 (97.99–98.36)
Year of publication< 0.05< 0.05
Before 2020150.89 (0.84–0.93)< 0.0596.26 (95.81–96.71)0.89 (0.83–0.93)< 0.0597.89 (97.68–98.10)
After 2020130.88 (0.85–0.90)< 0.0594.63 (93.87–95.39)0.83 (0.80–0.85)< 0.0595.12 (94.45–95.79)

P-Value for heterogeneity within each subgroup.

P-Value for heterogeneity between subgroups with meta-regression analysis.

Summary estimate of pooled performance of artificial intelligence in image-based ovarian cancer detection. P-Value for heterogeneity within each subgroup. P-Value for heterogeneity between subgroups with meta-regression analysis.

Discussion

With the widespread application of AI in medical imaging during recent years, radiomics and AI models are now being actively evaluated for diagnostic accuracy in a variety of malignancy types. To our best knowledge, this is the first systematic review and meta-analysis specifically dedicated to AI system performance in the diagnosis of OC. We are strictly in line with the guidelines for diagnostic reviews, and conducted a comprehensive literature search in both medical databases and engineering and technology databases to ensure the rigor of the study. After a careful selection of research on relevant topics, we found that AI algorithms excelled in the identification of OC using medical radiography imaging, which manifested an equivalent or even better performance than independent detection by human clinicians. This study also described the performance of the different imaging modalities, sample size, the year of publication, geographical distribution, and the different risk of bias levels. Potential sources of inter-study heterogeneity were identified based on the above subgroup and meta-regression analyses. More importantly, we rigorously rated study quality and risk of bias using an adapted QUADAS-AI assessment tool, which is a strength of this systematic review and will also better guide future related studies. Advances in ML techniques may facilitate processing of large amounts of medical image data. Notwithstanding their utility, ML methods are known to have limitations related to: manual extraction and selection of features, this is a fundamental task in order to find a group of significant variables to predict and correlate with outcome; Poor performance when dealing with imbalanced datasets. DL is the newest class of ML and has been found to be advantageous to other forms of ML. DL employs multiple layers of neural networks, leading to expanded ‘neuronal’ complexity, to significantly enhance computational power. However, with DL methods being more prone to overfitting and hence often requiring more data. Considering the stage of development of the algorithm and the difference in nature,, we also carried out a sub-analysis by the different algorithms, where no significant difference was observed. This may be attributed to the small dataset of included studies, most of which collected a few hundred data, limiting the advantages of DL. Although great promise has been shown with AI algorithms in a variety of tasks across radiology and medicine as a whole, these systems are far from perfect, we should also critically consider some methodological issues: First, data continues to be the most central and crucial constituent for learning AI systems. Exploiting radiology report databases by using modern information processing technologies may improve report search and retrieval and help radiologists in diagnosis. We need to call for advocacy for creating interconnected networks of identifying patient data from around the world and training AI on a large scale according to different patient demographics, geographic areas, diseases, etc. In addition, we emphasize that rare cancers, including OC, require more diverse image databases. In fact, maximization of the power of AI will require the deposition of medical data with sufficient annotation in large‐scale databases. However, such data are rarely curated, and this represents a major bottleneck in attempting to learn any AI model. International collaborative projects (such as The Cancer Imaging Archive [http://www.cancerimagingarchive.net]) that build large, labeled datasets should make a substantial contribution to meeting this challenge. Curation can refer to patient cohort selection relevant for a specific AI task but can also refer to segmenting objects within images., Curation ensures that training data adheres to a defined set of quality criteria and is clear of compromising artefacts. It can also help avoid unwanted variance in data owing to differences in data-acquisition standards and imaging protocols, especially across institutions, such as the time between contrast agent administration and actual imaging.71, 72, 73 Only in this way can we create an AI that is socially responsible and benefits more people. Second, as the advent of AI-based diagnostic test studies, there has been a parallel increase in the number of systematic reviews summarizing such findings.,, Noteworthy, 94% studies have been performed in the absence of an AI-specific quality assessment criteria in those published systematic reviews. During the past decade, the most frequently utilized tool is the QUADAS-2. However, QUADAS-2 does not address the particular terminology that arises from AI diagnostic test studies, nor does it consider other issues that appear in AI research, such as the setting of the data set, sources of bias, etc. Therefore, Sounderajah, V et al. proposed an AI-specific risk of bias tool, termed QUADAS-AI in 2021. This tool provided us with a specific instruction to assess the risk of bias and applicability of the present study. Not surprisingly, most of the relevant studies were more often designed or conducted prior this guideline. We therefore accepted the low quality of some of the studies and the heterogeneity between the included studies. It also makes sense that we assume that patient selection, index test and flow and timing of studies used to evaluate the diagnostic performance of AI models will be optimized soon. Third, although no publication bias was noticed in the present study, we must be honest about the fact that the available AI research is often a publication of positive results. We venture to guess that this phenomenon stems from reporting bias by researchers, which may have skewed the dataset and not conducive to the comparison between AI models and clinicians., One more point, the extraordinary applications of AI technology in medicine will require healthcare workers to enhance their clinical workflow combination. Of the included studies, only two evaluated the performance of integrating AI with clinicians. It has been suggested that scientific research should shift from an AI-physician dichotomy to a combination of AI and clinicians, which would be more in line with realistic medical workflows. Fourth, 28 of the 34 studies that met the inclusion criteria for this systematic review provided information of our concern for the development of contingency tables. There is a broad range of indicators employed in AI research to report diagnostic abilities. Metrics such as SE, SP, and accuracy are the most applied in numerous studies. If the number of subjects with/without disease is shown in the study, we can combine SE and SP to derive TP, TN, FP, and FN for the construction of the contingency table. Other metrics like precision, dice ratio, F1 score and recall, which are frequently used in computer science, also present as the default standard of measurement in some studies. However, these metrics are not all-encompassing and alone we do not receive sufficient information to build a contingency table. Well-defined metrics at the intersection of health care systems and computer science are also prudent to consider for future research. Additionally, for AI based models, the obtained heatmaps show what aspects of the images are important for a given classification, whereas few included studies provide such information. To reduce bias, we emphasize reporting information about segmentation properties or heat maps in AI model-based studies to draw conclusions about the elements of interest in AI models. Fifth, there is a disagreement around the critical terminology applied in AI research. Different papers have defined the same terminology in different ways. For example, for an AI-based model, the sample set is generally grouped into several separate sections, including a training set and a test set for evaluating the effectiveness of the model. Although the term 'validation' is used in a causal sense, some researchers used this phrase to denote the dataset used to assess the diagnostic performance of the ultimate model. Other investigations have described it as a dataset with a tweaking function in the exploitation process. The inconsistency of naming renders it challenging to determine whether the set is independent. It is vital that the validation set comprises data isolated from training data and is exclusively dedicated to assess the eventual model. It has been proposed to classify the sample data set into a training set, a tuning set and a validation set, whose functions are to be applied for training the model, for tuning the parameters and for assessing the performance of the final model, respectively. Considering the different sorts of validation sets, Altman et al. designated the datasets used for in-sample validation as internal validation set and those for out-of-sample validation as external validation set, suggestions which are very realistic and contribute to the quality of the study. Researchers concerned with the application of AI in healthcare should be careful about the phenomena and optimize it for future research. Sixth, within a purely image-based setting, AI can achieve on par or superior performance to physicians, thereby highlighting its potential as a decision support system with immediate clinical implications. Although a fairly good evaluation can be made in this way, it does not take into account all the information that radiologists rely on when evaluating a difficult examination. Nonimaging-based patient characteristics, such as demographic information, history of cancer, and genetic information, may be integrated into the model. Given a sufficiently large data set, AI could use these pieces of information in conjunction with the image data to identify women at high risk of cancer. Seventh, the high performance of AI model comes at the cost of high complexity and vast number of parameters. We may be unable to understand and explain why an AI model has made certain classifications in image analysis. This type of algorithm is often referred to as a “black box”. Compared with AI techniques, explainable artificial intelligence (XAI) can provide both decision-making and explanations of the model. Some research have been conducted into XAI to overcome the limitation of the black-box nature of AI methods. For example, Laios et al. have pioneered the implementation of XAI models in the field of gynecological oncology. They presented an ensemble AI-based model that predicted the outcomes following cytoreductive surgery for OC with high accuracy, and an XAI strategy that explained the patient and surgery-specific factors that led to that risk. The team also made a pioneering attempt to implement XAI models to explain the prediction of surgical effort at OC cytoreduction, by feeding the models with features that also include human factors. However, most of radiomics extraction and imaging biomarkers analyses included in this review are used as “black box”, and their application in clinical practice still lacks reliability and interpretability. This phenomenon is understandable given that the use of XAI in oncology is still in its infancy. Understanding the principles and applications of AI in medical imaging will facilitate assimilation and expedite advantages to practice. We encourage future researches to consider the interpretability of AI models in modeling, to address challenges, and to find clinical approaches for the development of AI in the field of radiomics. Eighth, most studies were carried out in a single center with limited data availability. Only three of the included studies have external validation, which refers to validating the performance of the model with out-of-sample datasets from other institutions. However, among the three included studies with external validation, only one was included in the meta-analysis. This precluded a subgroup meta-analysis in the present study, but emphasized the necessity for rigorous and reliable evaluation of AI performance in external datasets. The included studies are more likely to group an institution's dataset into a training set, a test set or an internal validation set. The performance was judged by the test set or internal validation. As the intention of the validation was to examine the performance of the model applied to patients from different populations, it is preferable to obtain a new dataset from a different organization. The lack of an external validation set may potentially lead to overestimation of the results, which could compromise the generalizability of the model. Several reviews, in the AI field reported that studies with internally validated AI models outperform externally validated models in the detection of cervical cancer, breast cancer and tumor metastases. However, this is not surprising as the samples in the same dataset are often homogeneous and the diagnostic performance of the algorithm can easily be misjudged. Rigorous external validation is warranted in the design of AI-related diagnostic studies. Multicentric studies will have an significant role in this research field. The use of interoperable standards and uniform protocols will also be needed prior to conducting such a study. AI methods can provide valuable models for quality assurance, personalized and predictive medicine. For this purpose, the contribution of clinicians and researchers in the interpretation of models and their application has a crucial role in the daily clinical practice. Additionally, limited prospective studies were carried out in real clinical environments. Most of the included studies were based on retrospective data and whose patients chosen from hospital medical records. It is well known that prospective studies would provide more favorable evidence, and we anticipate more prospective AI research to emerge in the future. And only considering the inclusion of English articles may omit important information from other language studies. Another limitation is that we did not contact the authors because most of the studies included in full-text screening (93%) provided the necessary data. The present study represents a summary of the enormous potential of AI algorithms that are useful for detecting OC using medical radiology imaging. However, it is also acknowledged that this finding is derived from relatively low methodological quality research, which inevitably overestimates the accuracy of the algorithm. The research of AI-based systems in diagnosing OC needs to be further improved in terms of study design.

Contributors

H-LX, T-TG, H-ZS, YS, Y-HZ, and Q-JW contributed to the conception and design of the study. H-LX, F-HL, and H-YC contributed to the literature search and data extraction. H-LX, F-HL, and T-TG contributed to risk of bias evaluation. H-LX, T-TG, F-HL, H-YC, and Q-JW contributed to data analysis and interpretation. H-LX, F-HL, H-YC, QX, H-ZS, YS, SG, T-TG, and Q-JW wrote the first draft of the manuscript and edited the manuscript. All authors contributed to critical revision of the manuscript. All authors approved the manuscript. H-LX and T-TG contributed equally to this work.

Data sharing statement

The search strategy was shown in Supplementary Note 1, and the contingency tables of 28 studies included in the meta-analysis were shown in Supplementary Table 1. The results of risk of bias and publication bias were separately provided in the Supplementary Figure 8 and 9. Additional data are available on request.

Declaration of interests

All authors declare no competing interests.
  84 in total

1.  Ovarian tumor characterization using 3D ultrasound.

Authors:  U Rajendra Acharya; S Vinitha Sree; M Muthu Rama Krishnan; Luca Saba; Filippo Molinari; Stefano Guerriero; Jasjit S Suri
Journal:  Technol Cancer Res Treat       Date:  2012-07-10

2.  Magnetic resonance imaging radiomics in categorizing ovarian masses and predicting clinical outcome: a preliminary study.

Authors:  He Zhang; Yunfei Mao; Xiaojun Chen; Guoqing Wu; Xuefen Liu; Peng Zhang; Yu Bai; Pengcong Lu; Weigen Yao; Yuanyuan Wang; Jinhua Yu; Guofu Zhang
Journal:  Eur Radiol       Date:  2019-04-08       Impact factor: 5.315

3.  A fast learning algorithm for deep belief nets.

Authors:  Geoffrey E Hinton; Simon Osindero; Yee-Whye Teh
Journal:  Neural Comput       Date:  2006-07       Impact factor: 2.026

4.  Bilateral analysis based false positive reduction for computer-aided mass detection.

Authors:  Yi-Ta Wu; Jun Wei; Lubomir M Hadjiiski; Berkman Sahiner; Chuan Zhou; Jun Ge; Jiazheng Shi; Yiheng Zhang; Heang-Ping Chan
Journal:  Med Phys       Date:  2007-08       Impact factor: 4.071

5.  Ultrasonography in the Diagnosis of Adnexal Lesions: The Role of Texture Analysis.

Authors:  Paul-Andrei Ștefan; Roxana-Adelina Lupean; Carmen Mihaela Mihu; Andrei Lebovici; Mihaela Daniela Oancea; Liviu Hîțu; Daniel Duma; Csaba Csutak
Journal:  Diagnostics (Basel)       Date:  2021-04-29

6.  QUADAS-C: A Tool for Assessing Risk of Bias in Comparative Diagnostic Accuracy Studies.

Authors:  Bada Yang; Sue Mallett; Yemisi Takwoingi; Clare F Davenport; Christopher J Hyde; Penny F Whiting; Jonathan J Deeks; Mariska M G Leeflang; Patrick M M Bossuyt; Miriam G Brazzelli; Jacqueline Dinnes; Kurinchi S Gurusamy; Hayley E Jones; Stefan Lange; Miranda W Langendam; Petra Macaskill; Matthew D F McInnes; Johannes B Reitsma; Anne W S Rutjes; Alison Sinclair; Henrica C W de Vet; Gianni Virgili; Ros Wade; Marie E Westwood
Journal:  Ann Intern Med       Date:  2021-10-26       Impact factor: 25.391

Review 7.  Artificial intelligence in medical imaging of the liver.

Authors:  Li-Qiang Zhou; Jia-Yu Wang; Song-Yuan Yu; Ge-Ge Wu; Qi Wei; You-Bin Deng; Xing-Long Wu; Xin-Wu Cui; Christoph F Dietrich
Journal:  World J Gastroenterol       Date:  2019-02-14       Impact factor: 5.742

Review 8.  Radiomics in liver diseases: Current progress and future opportunities.

Authors:  Jingwei Wei; Hanyu Jiang; Dongsheng Gu; Meng Niu; Fangfang Fu; Yuqi Han; Bin Song; Jie Tian
Journal:  Liver Int       Date:  2020-07-02       Impact factor: 5.828

9.  Ultrasound image analysis using deep neural networks for discriminating between benign and malignant ovarian tumors: comparison with expert subjective assessment.

Authors:  F Christiansen; E L Epstein; E Smedberg; M Åkerlund; K Smith; E Epstein
Journal:  Ultrasound Obstet Gynecol       Date:  2021-01       Impact factor: 7.299

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.