Literature DB >> 35253378

Artificial intelligence to classify ear disease from otoscopy: A systematic review and meta-analysis.

Al-Rahim Habib1,2,3, Majid Kajbafzadeh1, Zubair Hasan3, Eugene Wong3, Hasantha Gunasekera1,4, Chris Perry2,5, Raymond Sacks1, Ashnil Kumar6, Narinder Singh1,3.   

Abstract

OBJECTIVES: To summarise the accuracy of artificial intelligence (AI) computer vision algorithms to classify ear disease from otoscopy.
DESIGN: Systematic review and meta-analysis.
METHODS: Using the PRISMA guidelines, nine online databases were searched for articles that used AI computer vision algorithms developed from various methods (convolutional neural networks, artificial neural networks, support vector machines, decision trees and k-nearest neighbours) to classify otoscopic images. Diagnostic classes of interest: normal tympanic membrane, acute otitis media (AOM), otitis media with effusion (OME), chronic otitis media (COM) with or without perforation, cholesteatoma and canal obstruction. MAIN OUTCOME MEASURES: Accuracy to correctly classify otoscopic images compared to otolaryngologists (ground truth). The Quality Assessment of Diagnostic Accuracy Studies Version 2 tool was used to assess the quality of methodology and risk of bias.
RESULTS: Thirty-nine articles were included. Algorithms achieved 90.7% (95%CI: 90.1-91.3%) accuracy to difference between normal or abnormal otoscopy images in 14 studies. The most common multiclassification algorithm (3 or more diagnostic classes) achieved 97.6% (95%CI: 97.3-97.9%) accuracy to differentiate between normal, AOM and OME in three studies. AI algorithms outperformed human assessors to classify otoscopy images achieving 93.4% (95%CI: 90.5-96.4%) versus 73.2% (95%CI: 67.9-78.5%) accuracy in three studies. Convolutional neural networks achieved the highest accuracy compared to other classification methods.
CONCLUSION: AI can classify ear disease from otoscopy. A concerted effort is required to establish a comprehensive and reliable otoscopy database for algorithm training. An AI-supported otoscopy system may assist health care workers, trainees and primary care practitioners with less otology experience identify ear disease.
© 2022 The Authors. Clinical Otolaryngology published by John Wiley & Sons Ltd.

Entities:  

Keywords:  artificial intelligence; computer vision; diagnosis; machine learning; otoscopy

Mesh:

Year:  2022        PMID: 35253378      PMCID: PMC9310803          DOI: 10.1111/coa.13925

Source DB:  PubMed          Journal:  Clin Otolaryngol        ISSN: 1749-4478            Impact factor:   2.729


AI‐based computer vision algorithms can differentiate between binary (2 diagnosis options) and multiple ear disease diagnoses (3 or more diagnosis options). AI‐based computer vision algorithms have been shown to classify otoscopy images more accurately than human, nonexpert assessors. AI‐based computer vision algorithms have been developed using various machine learning techniques of which, models using CNNs achieve the greatest classification accuracy. Substantial heterogeneity was found between studies reflecting, in part, diverse sources of images and collection practices, machine learning methods, diagnostic classes and ground‐truth definitions. Future efforts are needed to establish a standardised, comprehensive and validated database to develop clinically relevant AI‐based computer vision algorithms for otoscopy.

INTRODUCTION

Otoscopy is a routine component of the ear assessment that visualises the external auditory canal and tympanic membrane (TM). It is used to identify common conditions, including infection (otitis externa, otitis media—acute and chronic), otitis media with effusion (OME), perforation, cholesteatoma, tympanosclerosis, foreign body, tympanostomy tube presence/position and cerumen impaction. Ear examinations are often conducted in primary care settings by local community health workers, nurses, medical students, general practitioners and emergency physicians. Concerning findings typically then lead to intervention or to specialist referral, where an otolaryngologist will examine the ear. Otoscopy accuracy in primary care settings varies based on user training and experience. However, the reported literature describes diagnostic accuracy estimates between 30% and 67.5% when compared to otolaryngologists (ground truth) for specific diagnoses. Previous efforts to improve performance have focussed on educational techniques, including online teaching modules comparing normal and abnormal otoscopic TM images, practical tutorials conducted by otolaryngologists and simulation with artificial ear models. , Comparisons made before and after education sessions demonstrate short‐term improvements in otoscopy performance, although benefits are not sustained long‐term and can decrease from initial assessments. The frequency of otoscopy in primary care settings and the observed performance inconsistencies may provide an opportunity for artificial intelligence (AI) to assist in the accurate identification of ear disease. AI can replicate the ability of human cognition to recognise patterns, identify anomalies and construct rational solutions to potential obstacles. Popularised applications of AI in health care include the use of computer vision to differentiate benign versus malignant skin lesions, identify diabetic retinopathy from fundoscopic images, assist radiologists to interpret chest x‐rays and predict infectious disease outbreaks, as in the case of the novel coronavirus (COVID‐19) pandemic. , , , AI‐based computer vision algorithms are an emerging technology that can be used to classify ear disease using otoscopic images. The use of AI‐based computer vision algorithms as an adjunct to otoscopy performed in primary care settings may be most relevant in rural and remote areas where access to otolaryngologists is limited. In these scenarios, ear examinations are often performed by nurses and community health workers with less clinical experience than otolaryngologists in accurately recognising ear disease. In rural and remote areas, telemedicine initiatives, such as tele‐otoscopy, are feasible strategies to capture otoscopic images for subsequent metropolitan specialist review but are often disadvantaged by delays in clinical decision making and implementation of interventions. The aim of this review was to evaluate the performance of AI‐based computer vision algorithms to classify ear disease from otoscopy. Our objectives were to synthesise existing literature related to the use of AI‐based computer vision algorithms for otoscopy, assess the performance of existing models and propose a guide for future algorithm development.

METHODS

The present systematic review was conducted in accordance with the PRISMA guidelines and registered with the International prospective register of systematic reviews (PROSPERO) on 18 February 2021 (ID number: CRD42021202594). Ethics approval and patient consent were not required for this review.

Literature search

A systematic search of online databases (including Google Scholar, MEDLINE, Embase, PubMed, Scopus, ProQuest, ACP Journal Club, Health Technology Assessment and the Cochrane Library) for articles, abstracts or conference proceedings published in the past 10 years through 31 October 2021 that used AI‐based computer vision approaches to classify otoscopic TM images was conducted. Searches were limited to those involving human subjects and those published in the English Language. Medical subject headings (MeSH) terms and non‐MeSH terms related to AI approaches included: ‘artificial intelligence’, ‘machine learning’, ‘deep learning’, ‘convolutional neural networks’, ‘support vector machines’, ‘image recognition’, ‘image classification’, ‘object detection’ and ‘computer‐assisted diagnosis’. MeSH terms and keywords related to otoscopic TM images included: ‘otoscopy’, ‘ear’, ‘eardrum’, ‘tympanic membrane’, ‘ear disease’, ‘acute otitis media’, ‘otitis media’, ‘chronic otitis media’ and ‘chronic suppurative otitis media’.

Selection criteria

Titles and abstracts were reviewed for eligibility by two independent investigators (ARH and MK). Discrepancies between the two investigators were resolved by the senior author (NS), a board‐certified otolaryngologist. Reference lists of available full‐text articles were also manually screened for further studies eligible for inclusion in this review. Diagnostic observational studies describing the development of an autonomous, supervised algorithm to classify otoscopic TM images using the AI approaches described above were included. Articles that were excluded consisted of those that did not use AI approaches of interest, utilised imaging modalities other than otoscopy or were review articles or editorials. Study inclusion/exclusion is summarised in a PRISMA flow diagram (Figure 1).
FIGURE 1

PRISMA study flow diagram [Colour figure can be viewed at wileyonlinelibrary.com]

PRISMA study flow diagram [Colour figure can be viewed at wileyonlinelibrary.com]

Data extraction

Two investigators (ARH and MK) independently extracted data from included studies for analysis. The following characteristics were extracted from included studies: primary author, year, study objective, AI technique to achieve study objective, data source, otoscope type and manufacturer, image labelling method, source for ground‐truth classification, diagnostic categories of interest, image size and quality, number and distribution of training images, number and distribution of test images, number and distribution of validation images, ratio of training/test/validation images by diagnostic categories and performance characteristics. Additional characteristics extracted included the type of deep learning models, batch size, learning rate, method used to standardise input images and use of image augmentation methods, segmentation, hyperparameter tuning and cross‐validation techniques.

Critical appraisal and risk of bias assessment

The quality of included studies was assessed using the Quality Assessment and Diagnostic Accuracy Studies‐2 (QUADAS‐2) tool, as per the Cochrane Collaboration for critical appraisal of diagnostic test accuracy evaluations. The QUADAS‐2 assessment tool is composed of patient selection, index test, reference standard and flow and timing. Study quality was assessed by two independent investigators (ARH and MK). Uncertainties or discrepancies were discussed with the senior author (NS) to achieve consensus. As described by Whiting et al., applicability concerns were determined for patient selection, the index test and the reference standard. For patient selection, reviewers considered whether the subjects recruited for the study and used to train the algorithm were applicable to the target population where algorithms would be implemented. For the index test, reviewers considered whether the classification categories of the algorithm were applicable to its intended use. For the reference standard, reviewers considered whether the method used reflected clinical practice for future applications.

Outcomes

A systematic review and meta‐analysis were performed with the primary outcome assessed being algorithm diagnostic accuracy in classifying ear disease from otoscopic images. Secondary outcomes included sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), F1 score and area under the curve (AUC).

Statistical analysis

The statistical analysis was performed using ReviewManager (RevMan 5.3, Copenhagen: The Nordic Cochrane Centre, The Cochrane Collaboration, 2014) in accordance with the Cochrane Handbook and Stata 17 (StataCorp LLC, 2021). Heterogeneity of model performance was summarised with the I 2 statistic, H 2 statistic, Cochrane's Q statistic and chi‐square test. Heterogeneity values >75% were considered as substantial heterogeneity. Inverse variance weighting was used for pooling model performance, and fixed effects were applied.

RESULTS

Study selection

The search strategy yielded a total of 1485 relevant articles and abstracts. Following full‐text review, 39 articles met the inclusion criteria (Figure 1).

Characteristics of included studies

Table 1 provides a summary of characteristics for included studies. Thirty studies collected images from outpatient and inpatient primary care settings, six studies combined images from primary care settings and online sources (Google Images), one study collected images from Google Images alone and two studies did not report the source of images.
TABLE 1

Summary of included studies

Author/yearImage sourceGround truthClassesTotal imagesAccuracy (%)Sensitivity (%)Specificity (%)PPV (%)NPV (%)F1 scoreAUC
Wang 2021 14 H2 otolaryngologists210081.783.380.080.6NRNR0.88
Sundgaard 2021 15 H1 otolaryngologist3133686.0NRNRNRNRNRNR
Tsutsumi 2021 16 H1, OO: NR, H1: 2 otolaryngologists, 1 paediatrician240077.070.084.481.473.7NR0.90
Tsutsumi 2021 16 H1, OO: NR, H1: 2 otolaryngologists, 1 paediatrician540066.055.4NR79.8NRNR0.88
Byun 2021 17 H3 otolaryngologists4237297.2NRNRNRNRNRNR
Zeng 2021 18 H6 otologists82054295.5NRNRNRNRNR0.99
Crowson 2021 19 H1 of 5 paediatric otolaryngologists233883.8NRNRNRNR80.00.93
Alhudhaif 2021 20 H1 2 otolaryngologists, 1 paediatrician485798.297.799.3NRNR96.9NR
Cai 2021 21 Hotolaryngologists (number not reported).4606693.4NRNRNRNR96.80.98
Camalan 2021 22 H1 1 otologist, 1 paediatric otolaryngologist330085.8NRNRNRNRNRNR
Ucar 2021 23 H1 1 otolaryngologist488098.198.199.498.2NRNR0.99
Camalan 2020 24 H1 NR345488.1NRNRNRNRNRNR
Wu 2020 25 H2 otologists31220397.896.898.096.998.4NR0.99
Basaran 2020 26 H1 2 otolaryngologists, 1 paediatrician228290.486.893.5NRNR87.30.95
Goshtasbi 2020 27 H1, OO: NR, H1: 2 otolaryngologists, 1 paediatrician240077.070.084.081.074.0NR0.89
Goshtasbi 2020 27 H1, OO: NR, H1: 2 otolaryngologists, 1 paediatrician540071.0NRNRNRNRNR0.91
Habib 2020 28 O2 otolaryngologists223376.076.076.076.076.0NR0.87
Khan 2020 29 H2 otolaryngologists3248487.095.0NR95.2NR95.10.99
Simon 2020 30 H1 2 otolaryngologists, 1 paediatrician295681.483.683.8NRNRNR0.89
Viscaino 2020 31 H1 1 otolaryngologist472088.187.895.987.7NRNR1.00
Cömert 2020 32 H1 2 otolaryngologists, 1 paediatrician485799.499.499.8NRNR99.3NR
Basaran 2019 33 H1 2 otolaryngologists, 1 paediatrician259897.999.198.5NRNRNRNR
Basaran 2019 34 H1 2 otolaryngologists, 1 paediatrician222376.170.880.1NRNRNR0.81
Cha 2019 35 H1 otologist, 1 physician61054494.293.796.8NRNRNRNR
Lee 2019 36 H2 otologists2133891.090.592.998.072.3NR0.92
Livingstone 2019 37 H, O1 otologist, 1 otolaryngology resident14136688.786.1NR90.9NRNRNR
Livingstone 2019 38 H2 otolaryngologists352984.4NRNRNRNRNRNR
Seok 2019 39 H2 otolaryngologists292092.9NRNR92.9NRNRNR
Huang 2018 40 NRNR32070.0NRNRNRNRNRNR
Kasher 2018 41 H, ONR210882.1NRNRNRNRNRNR
Myburgh 2018 42 NR2 otologists538986.286.896.487.496.4NRNR
Senaras 2018 43 H, ONR2108284.386.881.9NRNRNRNR
Tran 2018 44 HNR221491.489.593.391.9NRNR0.92
Senaras 2017 45 Hotolaryngologists (number not reported)224784.687.381.4NRNRNRNR
Myburgh 2016 46 H2 otologists548680.680.694.481.094.6NRNR
Wang 2015 47 H, ONR221590.085.092.0NRNRNRNR
Shie 2014 48 Hotolaryngologists (number not reported)486588.091.679.9NRNRNR0.91
Kuruvilla 2013 49 H3 otologists318189.9NRNRNRNRNRNR
Kuruvilla 2012 50 H3 otologists313584.0NRNRNRNRNRNR
Mironica 2011 51 H1 otolaryngologist218673.1NRNRNRNRNRNR
Vertan 2011 52 H1 otolaryngologist310059.9NRNRNRNRNRNR

Classes refers to the number of diagnostic categories included in the algorithm; (1) online data set.

Abbreviations: AUC—area under the curve; H—primary care hospital database; NPV—negative predictive value; NR—not reported; O—online repository; PPV—positive predictive value.

Summary of included studies Classes refers to the number of diagnostic categories included in the algorithm; (1) online data set. Abbreviations: AUC—area under the curve; H—primary care hospital database; NPV—negative predictive value; NR—not reported; O—online repository; PPV—positive predictive value.

Risk of bias of included studies

Figure 2 and Figure S1 illustrate the risk of bias and applicability concerns of included studies using the QUADAS‐2 tool. High risk of bias (37 of 39 studies) was identified in patient selection criteria due to failure to utilise consecutive or random sampling. High risk of bias (15 of 39 studies) was also observed in use of the reference standard (ground truth), as these articles did not utilise more than 1 otolaryngologist to review otoscopic images to confirm class labels.
FIGURE 2

Summary of risk of bias and applicability of concerns graph using the Quality Assessment of Diagnostic Accuracy Studies tool, version 2 (QUADAS‐2) [Colour figure can be viewed at wileyonlinelibrary.com]

Summary of risk of bias and applicability of concerns graph using the Quality Assessment of Diagnostic Accuracy Studies tool, version 2 (QUADAS‐2) [Colour figure can be viewed at wileyonlinelibrary.com]

Binary classification algorithms

Nine unique binary classification algorithms to classify otoscopy images were reported (Table S1).

Normal versus abnormal

AI algorithms achieved a pooled accuracy of 90.7% (95%CI: 90.1–91.3%) to difference between normal or abnormal otoscopy images with substantial heterogeneity between studies (n = 14 studies, I 2 = 96.9%, p = .001, Figure 3).
FIGURE 3

Forest plot comparing accuracy of AI algorithms to classify normal versus abnormal otoscopy images [Colour figure can be viewed at wileyonlinelibrary.com]

Forest plot comparing accuracy of AI algorithms to classify normal versus abnormal otoscopy images [Colour figure can be viewed at wileyonlinelibrary.com] Four studies , , , used the Özel Van Akdamar Hospital otoscopic image database to train binary algorithms using various classification techniques. Simon et al. demonstrated that pretrained convolutional neural networks (CNNs) and support vector machines (SVMs) could achieve greater classification accuracy than k‐nearest neighbours (k‐NNs), artificial neural networks (ANNs), decision trees (DTs) or the Naïve Bayes technique. Basaran et al. utilised multiple pretrained CNNs to evaluate the effect of segmentation, distribution between training and test data and cross‐validation on algorithm performance. In this study, pretrained CNNs achieved enhanced accuracy by applying basic image augmentation techniques and using region of interest patches, rather than full otoscopic images. The pretrained CNNs VGGNet‐16 and VGGNet‐19 achieved the greatest classification accuracy to differentiate normal from abnormal otoscopic images (90.5% and 90.1% respectively). Mironica et al. demonstrated that the greatest classification accuracy was achieved by CNNs and SVMs in 186 images collected from outpatient primary care settings. Enhanced performance was achieved by adding a colour coherence vector (CCV) to the algorithms (models with CCV vs without: CNN – 73.1% vs. 68.8%, SVM – 72.0% vs. 64.5%). Using intraoperative assessment of children taken to the operating room with the intent of myringotomy to determine the ground truth, Crowson et al. differentiated between normal and OME with 83.8% accuracy using ResNet‐34 and Monte Carlo cross‐validation resampling with 5 repetitions.

Multiclassification algorithms

Seventeen unique multiclassification algorithms to classify otoscopy images were identified (Table S2). Overall, multiclassification algorithms achieved a pooled accuracy of 96.2% (95%CI: 96.1–96.4%) with substantial heterogeneity between studies (n = 18 studies, I 2 = 98.8%, p = .001, Figure 4).
FIGURE 4

Forest plot comparing accuracy of multiclassification AI algorithms to classify ear disease from otoscopy, stratified by number diagnostic classes [Colour figure can be viewed at wileyonlinelibrary.com]

Forest plot comparing accuracy of multiclassification AI algorithms to classify ear disease from otoscopy, stratified by number diagnostic classes [Colour figure can be viewed at wileyonlinelibrary.com]

Normal, AOM and OME

The most common multiclassification model differentiated between normal, AOM and OME in five studies , , , , Overall, the algorithms achieved an accuracy of 97.6% (95%CI: 97.3–97.9%) with substantial heterogeneity between studies (n = 3 studies, I 2 = 98.7%, p = .001, Figure S2). , , Two studies were excluded from the meta‐analysis because the number of test images was not reported. , Wu et al. used 12 203 otoscopic images and applied the Xception and MobileNets‐V2 pretrained CNNs for classification. Image augmentation was used during the preprocessing phase consisting of rotation, width shift, height shift, shearing, zooming and horizontal flip. Hyperparametric tuning was used to establish the optimal classification algorithm (epochs: 100, initial learning rate: 0.001, batch size: 14 [Xception] and 64 [MobileNets‐V2]). Between diagnostic categories, OME was classified less accurately (96.7%) than normal (98.3%) or AOM (98.5%).

AI versus human classification

Five studies , , , , compared the performance of image classification algorithms to human assessors. Of these, three studies compared the performance of the AI algorithm to human assessors using the same test set. , , Restricting the meta‐analysis to these three studies demonstrated that AI algorithms outperformed human assessors having achieved an accuracy of 93.4% (95%CI: 90.5–96.4%) versus 73.2% (95%CI: 67.9–78.5%) with substantial heterogeneity between studies (AI: n = 3 studies, I 2 = 78.6%; human assessors: n = 3 studies, I 2 = 83.7%, p = .001, Figure 5, Table S3).
FIGURE 5

Forest plot comparing ear disease classification accuracy from otoscopy between AI algorithms and human assessors [Colour figure can be viewed at wileyonlinelibrary.com]

Forest plot comparing ear disease classification accuracy from otoscopy between AI algorithms and human assessors [Colour figure can be viewed at wileyonlinelibrary.com] Byun et al. found that a 4‐class image classification algorithm using ResNet‐18 and a shuffle attention model outperformed 10 nonexperts (first‐ and second‐year resident physicians) (97.1% vs. 82.9%) to classify normal, OME, COM or cholesteatoma. Otolaryngology experience of the resident physicians was not reported. Livingstone et al. recruited 10 nonexperts (general practitioners and trainees from paediatrics, emergency medicine and otolaryngology) to review a test set of 89 images for 14 diagnostic classes. Human, nonexpert assessors were summarised together, and stratification by specialty was not provided. Overall, the algorithm outperformed all the human assessors in 12 out of 14 categories (algorithm: 88.7% vs. human assessors: 58.9%). The algorithm classified cholesteatoma and otomycosis less accurately than human assessors (cholesteatoma – 50.0% vs. 55.0%, otomycosis – 0.0% vs. 40.0% respectively). Khan et al. recruited 17 human expert and nonexpert assessors (7 specialist otolaryngologists and 10 nonexperts) to review 100 test images for three diagnostic classes (normal, OME, COM). Overall, the algorithm achieved a classification accuracy of 87.0% compared to 74.0% for the pooled human expert and nonexpert assessors. Stratification of accuracy results by diagnostic class or assessor seniority was not provided.

DISCUSSION

Performance in diagnosing ear disease from otoscopy varies by user training and experience. This study summarises the performance of AI‐based computer vision algorithms to diagnosis of ear disease from otoscopy. Thirty‐nine studies were included in this review, evaluating the performance of 9 binary and 17 multiclassification algorithms. AI‐based computer vision algorithms achieved 90.7% (14 studies) accuracy to differentiate between normal or abnormal otoscopy images and 97.6% (3 studies) to differentiate between normal, AOM or OME. Compared to manual classification, AI‐based computer vision algorithms outperformed human assessors to classify otoscopy images (93.4% vs. 73.2% accuracy, respectively) in three studies. Substantial heterogeneity in performance was identified between studies. AI‐based computer vision algorithms with CNNs achieved greater accuracy in binary and multiclassification categories. Harnessing transfer learning by freezing the final layers of a CNN’s architecture is advantageous by using established computer vision performance to evaluate new tasks of interest. CNNs may have the potential to achieve greater accuracy than ANNs, SVMs or DTs due in part to their architecture, weighting system and features extracted in hidden layers. , However, CNNs vary in relation to the amount of computational power, speed and size of models produced. For example, Inception‐V3 yielded greater accuracy than MobileNet‐V2 in multiclassification but produced a larger size model. This is relevant whether considering deployment of the model into an edge device (e.g. otoscope or smartphone) versus workstations with graphics processing units (GPUs) or cloud‐based platforms. Focus on the TM and localising anatomical or pathological characteristics with segmentation models has achieved substantial improvements in accuracy while considering various middle ear conditions. This review has several strengths. Firstly, broad inclusion criteria were used to explore articles clinically relevant to the study question. This identified algorithms with various combinations of diagnostic classes and classification approaches. Secondly, the breadth of diagnostic classes is important for clinicians to appreciate the strengths and weaknesses of computer vision algorithms and highlight potential for future exploration. Thirdly, the QUADAS‐2 assessment tool was used to identify risk of biases and applicability concerns. Before image classification algorithms are introduced into routine clinical practice, rigorous efforts are required to evaluate performance and reliability. This systematic review has limitations. Firstly, this review is limited by the variability between eligible studies in terms of data sources, image capture devices and machine learning methods. Studies included used various patient selection criteria, classification categories and ground‐truth definitions. As a result, this could plausibly introduce selection and measurement bias in the outcomes reported and overall conclusions drawn from the results. The heterogeneity between classification categories suggests a lack of standardisation or quality control in terms of image acquisition. It is important to acknowledge the clinical need for an AI‐based computer vision algorithm and apply this technology accordingly. AI‐based computer vision algorithms depend on high‐quality training data with accurate and reliable ground‐truth labels. Ideally ground‐truth labels should be determined by consensus of multiple independent experts and/ or by additional evidence (e.g. clinical history, histopathology or independent investigation results). , For otoscopy, ground truth may be based on expert review of images along with clinical history, tympanometry, audiometry or myringotomy results. In this review, most studies used single otolaryngologists to classify otoscopic images during routine outpatient examinations. High risk of bias was identified in patient and image selection as most sampling methods for training images were not randomised or did not use consecutive recruitment. Furthermore, ground‐truth assessment was typically not validated by multiple independent experts or via other means. Crowson et al. uniquely used intraoperative otoscopic images of children undergoing myringotomy for recurrent AOM or OME. The authors suggested that this approach may represent the gold standard of detecting OME. Despite this, the authors did not report whether myringotomy was performed in all children, even those with low pretest probability of disease. Furthermore, positive pressure ventilation may displace middle ear fluid immediately prior to myringotomy in up to 15% of children with OME.  While myringotomy results may add useful additional evidence, this approach significantly reduces the number of test images available. Only those patients proceeding to intervention can be included, effectively excluding all patients with suspected normal ears (that do not proceed to intervention) from the training dataset. Accordingly, myringotomy results cannot be used as the basis for ground‐truth labelling in large training datasets. To create large datasets for training AI‐based computer vision algorithms, we recommend that ground‐truth labelling for otoscopic images be based primarily on the consensus of multiple independent experts along with clinical history, and tympanometry and audiometry results. This may reflect real‐world practices, where clinical suspicion from experts (e.g. otolaryngologists with subspeciality interest in otology) is used to determine which patient is appropriate for invasive interventions. Tele‐otoscopy databases may be a suitable source for this research as images have been collected by clinical staff with experience in performing otoscopy and have undergone quality control, vetting by an otolaryngologist, and often review in conjunction with clinical history, tympanometry and audiometry to establish a diagnosis and treatment plan. As identified in this review, the optimal approach to develop an AI‐based computer vision algorithm can use data augmentation, CNNs (DenseNet, Xception or Inception ResNet‐V2), ensemble models combining multiple classifying features, sequential multistep segmentation to localise the TM, hyperparameter tuning and cross‐validation resampling to yield the greatest performance. It is an important methodological consideration to partition data into training, validation and test groups to minimise the risk of overfitting, limit training bias and consider the generalisability of model predictions to independent or heterogenous data. Training a model and then testing, it on the same data would overestimate prediction performance and misrepresent its function on unseen data. However, partitioning data into discrete groups for training, validation and testing can reduce the number of samples used for model learning. To address this, cross‐validation techniques can be applied to divide the training group into sets (i.e. folds) and hold‐out sets to validate the model by averaging performance on a predetermined number of loops. , Most studies in this review applied cross‐validation techniques. Basaran et al. (2020) demonstrated that 10‐fold cross‐validation yielded higher accuracy, sensitivity and specificity results than utilising 50% of the data source for training and reserving 50% for testing. The collection of otoscopic images requires time, equipment, adequate storage, staff training and consideration of privacy, patient confidentiality and security. In this scenario, efforts to include a greater number of images for training and limiting bias may be important to maximise generalisability and clinical value. Attention models can be applied to visualise important areas used to establish predictions. Future algorithms may incorporate specific segmentation models using the U‐Net architecture. For example, Pham et al. (2021) proposed EAR‐UNet, an automatic segmentation model for TMs from video‐otoscopic images integrating pretrained CNNs (EfficientNet‐B4 and ResNet), achieving 96% accuracy to localise the TM in normal, AOM, OME and COM otoscopic images. Transitioning the AI‐based computer vision algorithm for otoscopy from a virtual environment to the clinical frontline will depend on real‐world test performance and applicability in daily clinical practice. This technology has the potential to inform judgement, improve triage and save time and resources for batch screening. Despite this, performance of the algorithm depends on training data. Bias in patient selection, ground‐truth labelling and diagnostic classes may impact generalisability. Implementation in real‐world settings may depend on desirably, feasibility and viability of this technology as an adjunct to existing clinical practices. Further efforts to progress the application of AI for otoscopy may be directed at establishing a comprehensive, publicly available, open access otoscopy database with associated symptoms, tympanometry, pneumatic otoscopy and audiometry findings, validated by multiple otologists. Future studies may build on previous studies evaluating the concordance between expert and nonexpert assessments, by exploring the use of AI as an adjunct to clinical decision making. Initial efforts for real‐world applications may be targeted at identifying poor quality otoscopic images that limit accurate assessment (e.g. blurriness, over saturation, lack of white balance, moisture, TM not visualised and wax obstruction). Delineating the strengths and limitations of AI to autonomously classify otoscopic images is necessary for real‐world applications.

CONCLUSION

In this review, 39 articles explore the role of AI‐based computer vision algorithms for otoscopy. AI algorithms achieved 90.7% accuracy to differentiate between normal or abnormal otoscopy images and 97.6% to differentiate between normal, AOM or OME. Compared to manual classification, AI algorithms outperformed human assessors to classify otoscopy images (93.4% versus 73.2% accuracy respectively). However, substantial heterogeneity in performance was identified between studies. A concerted effort is warranted to establish a standardised, robust, comprehensive and reliable database to develop clinically relevant computer vision algorithms for otoscopy. An AI‐based computer vision algorithm for otoscopy has the potential to support health care workers and primary care practitioners with less otology experience to identify and manage ear conditions early to minimise the risk of sequelae from untreated disease.

CONFLICT OF INTEREST

Dr Al‐Rahim Habib receives grant funding from the Garnett Passe and Rodney Williams Memorial Foundation, Microsoft AI for Humanitarian Action Grant and the Avant Foundation Doctor‐in‐Training Research Grant. Professor Raymond Sacks is a consultant for Medtronic. Associate Professor Narinder Singh receives grant funding from the Garnett Passe and Rodney Williams Memorial Foundation and Microsoft AI for Humanitarian Action Grant. He is a consultant for ENT Technologies, Optinose, Nasus Medical and ResMed, and a principal investigator for drug/ device trials for GSK, Lyra and Covance.

ETHICS STATEMENT

Ethics approval and patient consent were not in need for this study.

AUTHOR CONTRIBUTION

ARH, AK and NS designed the study; ARH and MK extracted and analysed the data; ARH prepared the manuscript; ARH, MK, ZH, EW, HG, CP, RS, AK, and NS involved in synthesis, revision and approval of manuscript. All authors agreed to be accountable for all aspects of the work. [Correction added on March 21, 2022, after first online publication: Peer review history is not available for this article, so the peer review history statement has been removed.] Supplementary Material Click here for additional data file.
  36 in total

1.  Detecting tympanostomy tubes from otoscopic images via offline and online training.

Authors:  Xin Wang; Tulio A Valdez; Jinbo Bi
Journal:  Comput Biol Med       Date:  2015-04-04       Impact factor: 4.589

2.  EAR-UNet: A deep learning-based approach for segmentation of tympanic membranes from otoscopic images.

Authors:  Van-Truong Pham; Thi-Thao Tran; Pa-Chun Wang; Po-Yu Chen; Men-Tzung Lo
Journal:  Artif Intell Med       Date:  2021-04-08       Impact factor: 5.326

3.  Deep metric learning for otitis media classification.

Authors:  Josefine Vilsbøll Sundgaard; James Harte; Peter Bray; Søren Laugesen; Yosuke Kamide; Chiemi Tanaka; Rasmus R Paulsen; Anders Nymark Christensen
Journal:  Med Image Anal       Date:  2021-03-14       Impact factor: 8.545

4.  Development and Validation of a Deep Learning Algorithm for Detection of Diabetic Retinopathy in Retinal Fundus Photographs.

Authors:  Varun Gulshan; Lily Peng; Marc Coram; Martin C Stumpe; Derek Wu; Arunachalam Narayanaswamy; Subhashini Venugopalan; Kasumi Widner; Tom Madams; Jorge Cuadros; Ramasamy Kim; Rajiv Raman; Philip C Nelson; Jessica L Mega; Dale R Webster
Journal:  JAMA       Date:  2016-12-13       Impact factor: 56.272

5.  Is otoscopy reliable? A structured teaching method to improve otoscopic accuracy in trainees.

Authors:  P J Wormald; G G Browning; K Robinson
Journal:  Clin Otolaryngol Allied Sci       Date:  1995-02

6.  Automatic detection of tympanic membrane and middle ear infection from oto-endoscopic images via convolutional neural networks.

Authors:  Mohammad Azam Khan; Soonwook Kwon; Jaegul Choo; Seok Min Hong; Sung Hun Kang; Il-Ho Park; Sung Kyun Kim; Seok Jin Hong
Journal:  Neural Netw       Date:  2020-04-01

7.  Machine Learning for Accurate Intraoperative Pediatric Middle Ear Effusion Diagnosis.

Authors:  Matthew G Crowson; Christopher J Hartnick; Gillian R Diercks; Thomas Q Gallagher; Mary S Fracchia; Jennifer Setlur; Michael S Cohen
Journal:  Pediatrics       Date:  2021-03-17       Impact factor: 7.124

Review 8.  A mobile telemedicine-enabled ear screening service for Indigenous children in Queensland: activity and outcomes in the first three years.

Authors:  Anthony C Smith; Nigel R Armfield; Wei-I Wu; Cecil A Brown; Chris Perry
Journal:  J Telemed Telecare       Date:  2012-12-03       Impact factor: 6.184

9.  Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement.

Authors:  David Moher; Alessandro Liberati; Jennifer Tetzlaff; Douglas G Altman
Journal:  PLoS Med       Date:  2009-07-21       Impact factor: 11.069

10.  Pneumonia of unknown aetiology in Wuhan, China: potential for international spread via commercial air travel.

Authors:  Isaac I Bogoch; Alexander Watts; Andrea Thomas-Bachli; Carmen Huber; Moritz U G Kraemer; Kamran Khan
Journal:  J Travel Med       Date:  2020-03-13       Impact factor: 8.490

View more
  2 in total

1.  A Machine Learning Approach to Screen for Otitis Media Using Digital Otoscope Images Labelled by an Expert Panel.

Authors:  Josefin Sandström; Hermanus Myburgh; Claude Laurent; De Wet Swanepoel; Thorbjörn Lundberg
Journal:  Diagnostics (Basel)       Date:  2022-05-25

2.  Artificial intelligence to classify ear disease from otoscopy: A systematic review and meta-analysis.

Authors:  Al-Rahim Habib; Majid Kajbafzadeh; Zubair Hasan; Eugene Wong; Hasantha Gunasekera; Chris Perry; Raymond Sacks; Ashnil Kumar; Narinder Singh
Journal:  Clin Otolaryngol       Date:  2022-03-15       Impact factor: 2.729

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.