Literature DB >> 31695786

Current status and future trends of clinical diagnoses via image-based deep learning.

Jie Xu1, Kanmin Xue2, Kang Zhang3.   

Abstract

With the recent developments in deep learning technologies, artificial intelligence (AI) has gradually been transformed from cutting-edge technology into practical applications. AI plays an important role in disease diagnosis and treatment, health management, drug research and development, and precision medicine. Interdisciplinary collaborations will be crucial to develop new AI algorithms for medical applications. In this paper, we review the basic workflow for building an AI model, identify publicly available databases of ocular fundus images, and summarize over 60 papers contributing to the field of AI development. © The author(s).

Entities:  

Keywords:  artificial intelligence; deep learning; machine learning; ophthalmology

Year:  2019        PMID: 31695786      PMCID: PMC6831476          DOI: 10.7150/thno.38065

Source DB:  PubMed          Journal:  Theranostics        ISSN: 1838-7640            Impact factor:   11.556


Introduction

Artificial intelligence (AI) has recently experienced an era of explosive growth across many industries, and healthcare is no exception 1. AI will have particular utility in healthcare and will dramatically change the diagnostic and treatment pathways for many, if not most, diseases. Regardless of the specific technique, the general aim of these technologies in medicine is to use computer algorithms to uncover relevant information from data and to assist clinical decision making 2. In many developed countries and China, the application of AI technology in healthcare has developed quickly, at least in part because it enhances human resources and abilities and improves the accuracy of medical treatment. As many countries that support the development of advanced technologies welcome the incoming era of AI, they will begin to develop the necessary specifications of governance by law, regulation, technology, and standards to fully optimize this developing field of technology. Ophthalmology is a discipline that is highly dependent on technological development. Modern ophthalmology currently makes full use of mechanical, electrical, magnetic, acoustic, optical, and other imaging technologies, and it will lead in fully implementing and adapting new technological developments such as AI. Ophthalmologists should enthusiastically embrace the development of AI technology and use it to promote advances in ocular medicine as much as possible.

Workflow of deep learning

AI is broadly used in both the technical and popular lexicons to encompass a spectrum of learning, including but not limited to machine learning, representation learning, deep learning, and natural language processing 1. Deep learning is making major advances in solving problems that have resisted the best attempts of the AI community for many years. It is very good at discovering intricate structures in high-dimensional data and is therefore applicable to multiple medical domains 3. Deep learning discovers intricate structure in large data sets by using multiple intermediate layers positioned between the input and output layers, allowing each level to learn to transform its input signal into the following layer (Fig 1). The application of deep learning, particularly in images of the retina include classification, e.g., detection of diabetic retinopathy (DR) and diabetic macular edema (DME) in fundus photographs 4; segmentation, e.g., segmentation of the lungs 5, brain 6, cell mitosis 7; and prediction, e.g., prediction of myopia development and progression 8. The workflow of deep learning can be defined in three stages: (1) pre-processing of the image data; (2) training of the model, validation, and model testing; and (3) evaluation (Fig. 2). Data pre-processing is a critical step that is necessary to build accurate machine learning models. The pre-processed work includes noise reduction, data normalization, feature selection, and extraction 9. For training a model, we initially split the model into three sections: data training, validation, and testing. The training set enables the model to learn to fit the data parameters of the classifier. The validation set is used to prevent overfitting, and the test set is used to evaluate the performance of the trained model. Evaluation is an integral part of the development process. It helps to determine if the model will do a good job of predicting the target on new and future data.
Figure 1

A typical deep learning neural network with multiple deep layers between input and output layers

Figure 2

Workflow diagram of developing a deep learning-based medical diagnostic algorithm.

Common open database of retina images

Many public databases have been published, and most include instructions for use by researchers in analysis and testing. For diseases of the retina, the databases include basic pathological features that usually provide information about the seven layers of the retina and about the choroid and sclera. This type of information is compiled by a process referred to as “segmentation”, which historically has been achieved by manual image processing, but increasingly it is done by computer algorithms. This information allows comparison of the performance of different algorithms analyzing the same fundus image, with reference to the reliable implementation of a gold-standard procedure 10. Commonly used fundus databases includes DRIVE 11, STARE 12, Image-Ret 13,14, e-ophtha 15, HEI-MED 16, Retinopathy Online Challenge 17, Messidor 18, RIM-ONE 19, and DRION-DB 20. Among them, DRIVE, STARE, Image-Ret, and Messidor are used mostly to diagnose DR, while DRION-DB and RIM-ONE are used mostly for segmentation of the optic nerve head in the diagnosis of glaucoma (Table 1).
Table 1

Summary of publicly available databases of ocular retinal images

DatabaseNumber of imagesCamera ModelImage Resolution (pixels) Field of ViewApplication
DRIVE 1140Canon CR5768×58445°Blood vessel segmentation
STARE 12400Topcon trv-50605×70035°Blood vessel segmentation; Optic disk detection
Image-Ret 13,14
DIARETDB0130unknown1500×115250°Diabetic retinopathy detection
DIARETDB189unknown1500×115250°Diabetic retinopathy detection
e-ophtha 15
e-ophtha_EX82OPHDIAT Tele-medical network2048×1360; 1440×960-Diabetic retinopathy detection
e-ophtha_MA3811440×960; 2544×1696Diabetic retinopathy detection
HEI-MED 16169Zeiss Visucam PRO fundus camera2196×195845°Hard exudate detection; Diabetic macular edema assessment
Retinopathy Online Challenge 17100Canon CR5-45-NM768×576; 1058×061; 1389×138345°Microaneurysms detection
Messidor 181200TopCon TRC NW61440×960; 2240×1488; 2304×153645°Diabetic retinopathy detection
RIM-ONE 19169Nidek AFC-210 with a body of a Canon EOS 5D2144 × 1424-Optic nerve head segmentation
DRIONS-DB 20110Color analogue fundus camera digitized by HP-PhotoSmart-S20 scanner600×400-Optic nerve head segmentation

Important research studies applying artificial intelligence to ophthalmic conditions

Image classification is a long-term research topic in the field of computer vision and pattern recognition. Recent advances of deep learning techniques have greatly facilitated the research of image classification. Many deep learning models have demonstrated performances comparable with well-trained human experts in the classification of natural images, and some have outperformed the experts. The flourish of deep learning-based image classification started with the work of AlexNet 21, where an eight-layered convolutional neural network (CNN) was designed for the classification task in ImageNet Large Scale Visual Recognition Challenge (ILSVRC). Subsequently, a series of deeper neural networks continuously refreshed the record of ILSVRC, including GoogLeNet (22 layers) 22, VGGNet (16 or 19 layers) 23, and Deep Residual Net (18, 34, 50, 101, or 152 layers) 24. These neural networks are the most widely used architectures that can achieve accurate classification for natural images by training deep models with millions of annotated images.

Diabetic retinopathy

Diabetic retinopathy is the most common organ complication and can manifest as the earliest sign of complication of diabetes mellitus. Early detection and continuous monitoring of DR is essential to control the disease in the early stage to prevent blindness. The automatic detection of DR has attracted a lot of attention. Most automated methods use fundus images as an input. These photographs are examined for the presence of lesions indicative of DR, including microaneurysms, hemorrhages, exudates, and cotton wool spots. The application of automated image analysis to fundus images may reduce the workload and costs by minimizing the number of photographs that need to be manually graded 25. Gulshan et al 4 were the first to present a deep learning algorithm for the detection of DR in retinal fundus photographs. In 2 validation sets of 9963 images and 1748 images, at the operating point selected for high specificity, the algorithm had 90.3%and 87.0%sensitivity and 98.1%and 98.5%specificity for detecting referable diabetic retinopathy, defined as moderate or worse diabetic retinopathy or referable macular edema by the majority decision of a panel of at least 7 US board-certified ophthalmologists. Subsequently, Ting et al26 developed a deep learning system to detect multiple related eye diseases, including DR, possible glaucoma, and age-related macular degeneration(AMD), the Area under the receiver operating characteristic curve (AUC) of 0.936 for referable DR ,sensitivity and specificity were 90.5% and 91.6%, For vision-threatening diabetic retinopathy, AUC was 0.958,sensitivity and specificity were 100% and 91.1%. More recently, deep learning was applied to automated segmentation of optical coherence tomography (OCT) images. Kermany et al 27 developed an OCT imaging diagnostic tool based on a deep learning framework for screening patients with AMD, DME, and drusen. The classifier distinguishing DME images from normal images achieved an accuracy of 98.2%, with a sensitivity of 96.8% and specificity of 99.6%. In April 2018, the first AI diagnostic system to receive US Food and Drug Administration (FDA) approval for marketing was IDx-DR, the case of IDx-DR highlights one of the earliest successes of an AI-based technology completing the regulatory process in the United States.

Glaucoma

Glaucoma is a group of eye diseases that damage the optic nerve and can result in irreversible vision loss and blindness and is the second leading cause of blindness worldwide. It is estimated that the disease affected 60.5 million people in 2010, and this figure is expected to reach 79.6 million by 2020 28. Currently, there is no cure for glaucoma, and vision loss, once it has occurred, is permanent. However, early detection and treatment are helpful to slow or stop the disease progression and can protect against serious vision loss. Many researchers have studied how to diagnose glaucoma automatically based on retinal images. These studies can be separated into two types. The first type outputs the glaucoma diagnosis results directly through deep learning models. Li et al 29 trained a CNN on LabelMe datasets for glaucoma diagnosis, In the validation dataset, this DL system achieved an AUC of 0.986 with sensitivity of 95.6% and specificity of 92.0%. The most common reasons for false-negative grading were glaucomatous optic neuropathy with coexisting eye conditions, including pathologic or high myopia, DR, and AMD. The leading reason for false-positive results was having other eye conditions, mainly including physiologic cupping. The second type of studies uses deep learning models to segment the glaucoma related tissues such as optic disc and optic cup, and then calculates medical measures (e.g., cup-to-disc ratio) for diagnosis. Previous studies have used various special forms of perimetry to discriminate preperimetric glaucoma from healthy eyes30. Asaoka et al 31 applied a DL method to differentiate the visual fields of preperimetric open-angle glaucoma patients from the healthy eyes, the AUC was 0.926.

Age-Related Macular Degeneration

AMD is a leading cause of irreversible visual loss in the aging population; the meta-analysis conducted by Wong et al 32 suggested that AMD, was responsible for 8.7% of all global blindness. Fortunately, the anti-vascular endothelial growth factor (anti-VEGF) medications have revolutionized the treatment of exudative retinal diseases, OCT is critical to guiding the administration of anti-VEGF therapy by providing a clear cross-sectional representation of the retinal pathology in these conditions. Kermany et al 27 developed an OCT imaging diagnostic tool based on a deep learning framework for screening patients with AMD, DME, and drusen. This AI system categorized images with choroidal neovascularization and images with diabetic macular edema as “urgent referrals”, drusen as “routine referrals”, normal images were labeled for “observation”. They achieved an accuracy of 96.6%, with a sensitivity of 97.8%, a specificity of 97.4%, and a weighted error of 6.6%. The classifier distinguishing choroidal neovascularization images from normal images achieved an accuracy of 100.0%, with a sensitivity of 100.0% and specificity of 100.0%. Recently, DeepMind and the Moorfields Eye Hospital 33 developed an AI system was trained on 14 884 OCT scans to detect 9 different OCT pathologies (choroidal neovascularization, macular edema, drusen, geographic atrophy, epiretinal membrane, vitreomacular traction, full-thickness macular hole, partial thickness macular hole, and central serous retinopathy). The system was then able to recommend a referral decision based on the most urgent conditions detected, the correct referral decision with 94% accuracy, matching world-leading eye experts. In addition to detecting and monitoring common blinding eye diseases, deep learning is also being expanded to the field of rare diseases, such as congenital cataracts and retinopathy of prematurity (ROP) in newborns. Long et al 34 constructed a CNN-based computer-aided diagnosis framework (CC-Cruiser) to classify and grade congenital cataract. In the clinical trial, CC-Cruiser achieved 98.25% accuracy with the identification networks; 100%, 92.86% and 100% accuracy for opacity areas, densities and locations, respectively, with the evaluation networks; and 92.86% accuracy with the strategist networks. Brown et al 35 reported the results of a fully automated DL system that could diagnose plus disease, the most important feature of severe ROP, for diagnosis of plus disease, the algorithm achieved a sensitivity of 93% with 94% specificity. For detection of pre-plus disease or worse, the sensitivity and specificity were 100% and 94%, respectively. In addition, retinal microvascular changes and retinopathy provide important clinical indicators for predicting the occurrence, development, therapeutic effect and prognosis of cardiovascular and cerebrovascular diseases. Poplin et al36 using deep-learning models trained on data from 284,335 patients and validated on two independent datasets of 12,026 and 999 patients, they predicted cardiovascular risk factors not previously thought to be present or quantifiable in retinal images, such as age (mean absolute error within 3.26 years), gender(AUC) = 0.97, smoking status (AUC = 0.71), systolic blood pressure (mean absolute error within 11.23 mmHg) and major adverse cardiac events (AUC = 0.70).Current AI studies using deep learning techniques for DR, AMD, glaucoma, cataract, and anterior ocular segment diseases are summarized in Table 2 4,26,27,29,31,34,37-63.
Table 2

Summary of influential papers on ophthalmic image analysis

DiseaseProcedures / examinationsData setsDeep learning techniquesPerformanceReference
KeratoconusPentacam194 normal, 454 keratoconus, 67 forme fruste, 28 astigmatic, 117 after refractive surgerySVMAcc: 98.9%; Sen: 99.1%; Spe: 98.5%; AUC:0.99Hidalgo et al. 37
Dual Scheimpflug Analyzer177 normal, 148 keratoconusDecision TreeSen: 100%; Spe: 99.5%Smadja et al. 38
Pentacam HR30 normal, 15 unilateral keratoconus, 30 bilateral keratoconusFNNBilateral keratoconus versus normal; AUC: 0.99; Sen: 100%; Spe: 95%Kovacs et al. 39
PterygiumAnterior segment photographed images2,692 non-pterygium, 325 pterygiumShape features + SVM/ANNAcc: 91.27%; AUC: 0.956; Sen: 88.7%; Spe: 88.3%Zaki et al. 40
CataractSlit-lamp image5,378 images with decimal grading scores ranging from 0.1 to 5.0.CRNN70.7% exact integral agreement ratio (R0); 88.4% decimal grading error ≤ 0.5 (Re0.5); 99.0% decimal grading error ≤ 1.0 (Re1.0 ).Gao et al. 41
Slit-lamp image476 normal, 410 cataractDCNNCataract vs Normal:Acc:98.87%;Sen:98.78%Spe:98.95%Long et al. 34
Fundus image767 normal, 472 cataract (246 mild cataract,128 moderate cataract, and 98 severe cataract)SVM, BPNNAcc: 93.2% for detection,84.5% for grading; Sen:94.2% for detection,74.6-89.3% for grading;Spe:91.5% for detection,90.4-98.9% for gradingYang et al. 42
Slit-lamp image476 normal, 410 pediatric cataractCNN, SVMAcc, Sen, and Spe: classification (97.07%, 97.28%, and 96.83%,) three-degree grading area (89.02%, 86.63%, and 90.75%) density (92.68%, 91.05%, and 93.94%) location (89.28%, 82.70%, and 93.08%)Liu et al. 43
POAGFundus imageTraining set:125,189; Validation set: 71,896DLSAUC: 0.942; Sen: 96.4%; Spe: 87.2%Ting et al. 26
Fundus imageTraining set: 31,745; Validation set: 8,000DCNNAUC: 0.98; Acc: 92.9%; Sen: 95.6%; Spe: 92.0%; AUC: 0.986Li et al. 29
Fundus image589 normal, 837 glaucomaCNNAcc: 98.13%; Sen: 98%; Spe: 98.3%Raghavendra et al. 44
Fundus image30 normal, 30 open-angle glaucomaSVWAcc:91.67; Sen:90%;Spe:93.33%Krishnan et al. 45
Visual fieldTraining set:257; Test set: 129ANNAUC: 0.890; Sen: 78.3%; Spe: 89.5%Oh et al. 46
Fundus image266 normal, 72 mild, 86 moderate, 86 severe glaucomaSVMAcc: 93.1%; Sen: 89.75%; Spe: 96.2%Acharya et al. 47
Fundus image and SLO imageNormal/glaucoma Fundus images:85/39; Normal/glaucoma SLO images: 46/19RIFMAcc for Fundus images: 94.4%,SLO images: 93.9%;Sen for Fundus images: 92.3%,SLO images: 89.5%;Spe for Fundus images: 95.3%,SLO images: 93.5%Haleem et al. 48
Visual fields53 glaucoma eyes, 108 normal eyesFNNAUC: 92.6%, The sensitivity was 77.8%,54.6%, and 50.0%, respectively, at the specificity of 90%, 95%,and 99%Asaoka et al. 31
DRFundus imageTraining set:76,370; Validation set: 112,648DLSFor referable DR: AUC: 0.936; Sen: 90.5; Spe: 91.6%; For vision-threatening DR: AUC: 0.958; Sen: 100%; Spe: 91.1%Ting et al. 26
Fundus imageDevelopment Data Set (EyePACS in the United States and 3 eye hospitals in India): 128,175 Validation Data Set (EyePACS-1: 9,963; Messidor-2: 1,748)DCNNAUC: 0.991 for EyePACS,0.990 for Messidor;Sen: 90.3% for EyePACS,87% for Messidor;Spe: 98.1% for EyePACS,98.5% for MessidorGulshan et al. 4
Fundus image170 DR, 170 normalPNN-GA, SVM quadratic kernelsPNN-GA: Acc:99.41%,Sen:99.41%Spe:99.41%; SVM: Acc:99.12 % Sen:98.82%,Spe:99.41%Ganesan et al. 49
Fundus imageEyePACS: 75,137 DR images; External validation: MESSIDOR 2 and E-OphthaDCNNAUC 0.94 for Messidor 2, 0.95 for E-Ophtha;Sen 93% for Messidor 2,87% for E-Ophtha;Spe 90% for Messidor 2,94% for E-OphthaGargeya et al. 50
Fundus imageTraining set: 327 diabetic patients; Validation set: 725 diabetic patientsLASSOAcc: 89.2%; AUC: 0.90; Sen: 75%; Spe: 89.6%Oh et al. 51
Fundus imageTraining set: 400; Testing set: 9,954Ensemble of classifiers with hidden Markov chain for context information, trained by genetic algorithmSen: 92.2%; Spe: 90.4%Tang et al. 52
Fundus imageMessidor-2 dataset: 1,748CNNReferable DR: AUC: 0.980; Sen: 96.8%; Spe: 87%; Vision threatening DR: AUC: 0.989; Sen: 100%; Spe: 90.8%Abramoff et al. 53
Fundus image4,445 DR; 5,494 normalDCNNAcc: 0.81Takahashi et al. 54
Fundus imageDIARETDB1, FAZ, MESSIDOR, Private dataset: 750 (Normal: 150, mild NPDR: 150, moderate NPDR: 150, severe NPDR: 150, PDR: 150)DNNAUC: 0.924; Sen: 92.18%; Spe: 94.50%Abbas et al. 55
DMESD-OCTTraining set: 11,349; DEM; 51,140 normal; Validation set: 250 DME, 250 normalCNNAcc: 98.2%; Sen: 96.8%; Spe: 99.6%Kermany et al. 27
DMEFundus image283 DR; 1,086 normalEnsemble of Gaussian mixture model and SVM with RBF kernelAcc: 96.8%; Sen: 97.3%; Spe: 95.9%Akram et al. 56
AMDFundus imageTraining set 72,610; Validation set: 35,948DLSAUC: 0.931; Sen: 93.2%; Spe: 88.7%Ting et al. 26
Fundus imageAREDS dataset: >130,000DCNNAUC: 0.94∼0.96 Acc: 88.4%∼91.6% Sen: 71%∼88.4% Spe: 91.4%∼94.1%Burlina et al. 57
Fundus imageAREDS dataset: 5,664DCNNAcc 79.4% (4-class) 81.5% (3-class); 93.4% (2-class)Burlina et al. 58
SD-OCTTraining and validation sets: 1,012 (AMD: 701; normal: 311); Test:100 (AMD: 50, normal: 50)DCNNAcc: 96%; Sen: 100%; Spe: 92%Treder et al. 59
Fundus image135 AMD subjects, 135 normal subjectsFeature extracted by Discrete wavelet transform and others for feature selection and classificationAverage Acc: 93.7%; Sen: 91.11%; Spe: 96.3%Mookiah et al. 60
OCT48,312 AMD; 52,690 normalDCNNAUC: 0.975; Sen: 92.6%; Spe: 93.7%Lee et al. 61
SD-OCT1,289CNNThe mean Dice coefficient for human interrater reliability and deep learning were 0.750 and 0.729, respectively.Lee et al. 62
CNVSD-OCTTraining set: 37,206 CNV, 51,140 normal; Validation set: 250 CNV, 250 normalCNNAcc: 100%; Sen:100%; Spe:100%Kermany et al. 27
CNVFluorescein angiography33AdaBoostAccuracy: 83.26%Tsai et al. 63

RBFNN, radial basis function neural network; SVM, support vector machine; MLP, multi-layer perceptron; CRNN, convolutional-recursive neural networks; DCNN, Deep-learning convolutional neural network; BPNN, Back propagation neural network; DLS, deep learning system; CNN-FE, convolutional neural networks feature-exaggerated; MLP-BP, Multilayer Perceptron with Back Propagation; RIFM, Regional Image Features Model; FNN, feed-forward neural network; PNN-GA, probabilistic neural network-genetic algorithms; LASSO, least absolute shrinkage and selection operator; NB, naive Bayes; PNN, probabilistic neural network; RBF, Radial basis function; SD-OCT, spectral domain optical coherence tomography; SLO, Scanning Laser Ophthalmoscopy. Acc, accuracy; Sen, sensitivity; Spe, specificity; Vs, versus; AUC, area under the curve; POAG, primary open-angle glaucoma; AMD, age-related macular degeneration; OCT, optical coherence tomography; DR, diabetic retinopathy.

Status of AI applications in clinical diagnoses

On April 2, 2019, the FDA issued a discussion paper that proposed a regulatory framework for modifications to artificial intelligence/machine learning (AI/ML)-based software as medical devices (SaMD) 64. This document explains the principles for developing artificial intelligence software, the new framework for equipment, the principles of the total product lifecycle (TPLC) certification methodology, and examples of potential real-world AI software modifications that may or may not be allowed under the proposed framework. The idea of the proposal was that with appropriately tailored regulatory oversight, AI/ML-based SaMD will deliver safe and effective software functionality that improves the quality of care that patients receive. To date, two AI algorithms have been fully approved by the FDA and used clinically. One is IDx-DR for detecting DR, and the other is Viz.AI for analyzing images for indicators associated with a stroke. These two devices are described as “locked” algorithms, meaning that they can only be modified by the manufacturer at intervals for the purpose of “training” with new data, followed by manual verification and validation of the updated algorithm. However, there is much promise beyond locked algorithms that is ripe for health care applications. These opportunities require careful oversight to ensure that the benefits of these advanced technologies outweigh the risks to patients. These machine learning algorithms can continually evolve and are often referred to as “adaptive” or “continuously learning” algorithms. Adaptive algorithms can learn from new user data presented to the algorithm through real-world use. The FDA is exploring a framework that would allow for modifications to algorithms to be made from real-world learning and adaptation, while ensuring that the safety and effectiveness of the software as a medical device is maintained 65.

Main challenges in the application of AI

At present, there are at least four limitations in AI technology that is based on machine learning, First, most machine learning methods have too few training sets and verification sets. More image data training is needed to further improve accuracy, sensitivity, and specificity. Transfer learning is an approach is more suitable when limited training data is available for the problem under consideration. In transfer learning, one can learn a complex model using data from a source domain where large-scale annotated images are available (e.g. natural images). Then, the model is further fine-tuned with data of the target domain where only a small number of annotated images are available (e.g. medical images) 66 (Fig 3).
Figure 3

Illustrations of transfer learning: a neural network is pretrained on ImageNet and subsequently trained on retinal, OCT, X-ray images, B-scans for different disease classifications

The second limitation is that the examination/detection equipment used in different countries, regions, and medical institutions is not uniform; therefore, the acquired images have differences in quality and resolution, which will inevitably affect the accuracy of image analysis and thus affect the accuracy of the diagnosis. These differences will present certain obstacles in the wide-scale applications of AI technology. One solution to this problem is to unify and standardize the examination equipment. This will be difficult to achieve. Another method is to further improve AI machine learning methods at the framework and algorithm level so that they can be flexibly applied to images of different qualities while simultaneously ensuring the accuracy of intelligent diagnosis. This will increase the applicability of AI in different regions and medical institutions. However, a lot of research support is still needed in this area. The third limitation is that the current machine learning methods for disease diagnosis lack “explanation capacity”. They do not have the ability to provide the clinician or other users with the reason for the diagnosis. The output result is based only on training and intensive learning. Thus, it is only a simple statement based upon the differences in the patient and normal images. There is no explanation for why the differences exist or the pathological basis of the differences. This, along with other issues, will affect to some extent the acceptance of these devices by doctors in clinical applications and could even provoke confusion among clinicians. Finally, the fourth limitation in machine learning-based AI technology is that the diagnosis of some rare diseases is still unreliable. Because of the scarcity of these diseases, there are not enough cases to meet the requirements of the training and verification sets. It is difficult or impossible to ensure the accuracy of the learning model in diagnosing rare diseases. To improve the diagnosis of rare diseases, it will be necessary to optimize machine learning algorithms. This can be done by transitioning from reliance on the number of learning samples for accuracy training to utilizing combinations of various training modes and types.

Suggestions for the application and development of AI in medicine

1. Consolidate the data foundation of AI applications

AI must first collect a certain amount of data in the medical industry. The structure of current medical data is very complex, in part, because there is no uniformity in the standards for compilation, and this leads to widespread information islands. It is necessary to establish a mechanism for circulating and sharing medical data. Development of data desensitization methods will promote data standardization and normalization. These methods can be used to establish standard test data sets to consolidate the data foundations of AI applications.

2. Define the positioning of AI development

In medicine, AI aims to help doctors (rather than substitute for doctors) to reduce the morbidity and mortality rates of patients waiting for professionals. Because doctors will not be replaced by AI, the diagnostic result of AI is only a reference for a clinical diagnosis, and doctors will always be responsible for the result. Currently, AI products in medicine play only an auxiliary role in the clinic, such as the diagnosis of DR, cancer screening, medical image recognition, disease rehabilitation, and in other fields.

3. Formulate regulations and laws for the application of AI in medicine

To achieve the uniform standards necessary for effective medical AI applications, laws, regulations, and other levels of governance must be established at the national level. The implementation of the national standards in industry and in routine and research clinical settings will ensure that the technology can be made widely and quickly available in the safest and most rational way. This will prudently promote the application of AI in the medical field.

4. Strengthen data security of AI data applications

As with the collection of any personal and medical data, the risk of inadvertent or pirated disclosure is a major concern. To reduce these risks, it is necessary to strengthen the construction of privacy protection, desensitize the data, and collect the data according to different levels and different granularities to reduce the risk of privacy disclosure. The United States has extensive experience in privacy protection of medical data. While enhancing privacy protection, it encourages the rational access and meaningful use of data and makes a selected portion of it available for research in real time, open access databases.

5. Promote the cultivation of interdisciplinary talents

Future development of AI in medicine can be enhanced by focusing on the integration of disciplines such as medicine, information science, and engineering. This goal can be achieved by encouraging universities, research institutes, and enterprises to cooperate with each other, set up appropriate scholarships, and establish training bases and local pilot programs. Finally, an innovative talent introduction system and mechanism should be developed to attract highly talented students to carry out innovation and entrepreneurship in the field of medical AI, thus driving the further development of the field.

Future of AI application in clinic

Machine learning has shown its great potential in ophthalmology. Most of the current studies regarding intelligent diagnosis of eye diseases focus on dual classification problems, whereas many patients suffer from multiple categorical retinal diseases in the clinical setting. It is therefore necessary to have a model for detecting and distinguishing DR, AMD, glaucoma, and other retinal disorders simultaneously 67. With a new generation of AI developed as a broad strategy, the applications of AI in the medical field will increase and improve. AI plays an important role in disease diagnosis and treatment, health management, drug research and development, precision medicine, etc. It can contribute significantly to solving problems of the uneven distribution of medical resources, reducing costs, and improving treatment efficiency. Applying AI helps to make up for the shortcomings of insufficient medical resources, enhance the fairness of medical services, and improve the construction of hierarchical diagnosis and treatment. In the future, AI will also offer important support for establishing an integrated medical service system. A qualified and efficient integrated medical service system can be built with the help of information-based systems. Laws and regulations to define the legal status, responsibility sharing mechanisms, and supervision of automated systems are not yet enacted in China and most other countries. Given the complex ethical boundaries of medical AI application, the excessive control of medical AI will hinder innovation and development. On the other hand, the lack of management brings the risk of unclear subject responsibility in AI applications. Therefore, it is necessary to reasonably define AI in the medical field. The current laws on medical AI regulation are either non-existent or are in only the most primitive stages of development. There are no sound laws that regulate medical big data, the basis for medical AI. Further, there are no clear legal instructions regarding the ownership of AI data, the right to use it, privacy standards, data security, accountability norms, and whether laws can protect researchers, clinicians, and engineers from errors of innovation.

Summary

Deep learning has the ability to probe more deeply into and discern more discriminative features from extremely large datasets. It has been applied in many research and clinical fields that rely on medical image analysis, making breakthrough progress in those disciplines. Due to unique features in ophthalmology, the diagnosis of eye diseases in clinical practice requires interpretation of many imaging studies for auxiliary diagnosis. However, detection resolution of the human eye is limited and so is human attention span. Proficiency levels of ophthalmologists also differ and it's inevitable for human errors to occur. As documented in the existing literature review, most of the current deep learning methods representing the leading level are the use of supervised learning, especially the CNN-based framework. Preliminary researches mainly focused on pre-training CNN and taking CNN as feature extractor. These pre-training networks could be downloaded directly, and conveniently applied to the analysis of any medical images. In recent two years, end-to-end training CNN has become a prioritized approach for the analysis of medical images. However, obtaining well-annotated data used for supervised learning is another major challenge for the application of deep learning to the analysis of medical images. As annotated data were usually limited at present, how to utilize unannotated images to achieve a high diagnostic accuracy using a combination of unsupervised and supervised learning will be another important development direction. In addition, electronic medical records (EMRs) contain a wealth of clinical diagnostic and treatment information that can be extracted and used to form diagnoses using natural language processing and deep learning. This information can be used to supplement the image data to formulate a complete diagnosis mimicking a human physician 68. Thus, in the near foreseeable future, AI relying on deep learning will combine image analysis with EMRs, further advancing the diagnostic power and ability to monitor disease progression and response to treatment in ways never before anticipated.
  49 in total

1.  Detecting Preperimetric Glaucoma with Standard Automated Perimetry Using a Deep Learning Classifier.

Authors:  Ryo Asaoka; Hiroshi Murata; Aiko Iwase; Makoto Araie
Journal:  Ophthalmology       Date:  2016-07-07       Impact factor: 12.079

2.  Interpretation of the Humphrey Matrix 24-2 test in the diagnosis of preperimetric glaucoma.

Authors:  Jin A Choi; Na Young Lee; Chan Kee Park
Journal:  Jpn J Ophthalmol       Date:  2009-01-30       Impact factor: 2.447

Review 3.  Automated pterygium detection method of anterior segment photographed images.

Authors:  Wan Mimi Diyana Wan Zaki; Marizuana Mat Daud; Siti Raihanah Abdani; Aini Hussain; Haliza Abdul Mutalib
Journal:  Comput Methods Programs Biomed       Date:  2017-10-31       Impact factor: 5.428

Review 4.  Promising Artificial Intelligence-Machine Learning-Deep Learning Algorithms in Ophthalmology.

Authors:  Lokman Balyen; Tunde Peto
Journal:  Asia Pac J Ophthalmol (Phila)       Date:  2019-05-31

5.  Artificial Neural Network Approach for Differentiating Open-Angle Glaucoma From Glaucoma Suspect Without a Visual Field Test.

Authors:  Ein Oh; Tae Keun Yoo; Samin Hong
Journal:  Invest Ophthalmol Vis Sci       Date:  2015-06       Impact factor: 4.799

6.  Costs and consequences of automated algorithms versus manual grading for the detection of referable diabetic retinopathy.

Authors:  G S Scotland; P McNamee; A D Fleming; K A Goatman; S Philip; G J Prescott; P F Sharp; G J Williams; W Wykes; G P Leese; J A Olson
Journal:  Br J Ophthalmol       Date:  2009-12-03       Impact factor: 4.638

Review 7.  Global prevalence of age-related macular degeneration and disease burden projection for 2020 and 2040: a systematic review and meta-analysis.

Authors:  Wan Ling Wong; Xinyi Su; Xiang Li; Chui Ming G Cheung; Ronald Klein; Ching-Yu Cheng; Tien Yin Wong
Journal:  Lancet Glob Health       Date:  2014-01-03       Impact factor: 26.763

8.  Identifying Medical Diagnoses and Treatable Diseases by Image-Based Deep Learning.

Authors:  Daniel S Kermany; Michael Goldbaum; Wenjia Cai; Carolina C S Valentim; Huiying Liang; Sally L Baxter; Alex McKeown; Ge Yang; Xiaokang Wu; Fangbing Yan; Justin Dong; Made K Prasadha; Jacqueline Pei; Magdalene Y L Ting; Jie Zhu; Christina Li; Sierra Hewett; Jason Dong; Ian Ziyar; Alexander Shi; Runze Zhang; Lianghong Zheng; Rui Hou; William Shi; Xin Fu; Yaou Duan; Viet A N Huu; Cindy Wen; Edward D Zhang; Charlotte L Zhang; Oulan Li; Xiaobo Wang; Michael A Singer; Xiaodong Sun; Jie Xu; Ali Tafreshi; M Anthony Lewis; Huimin Xia; Kang Zhang
Journal:  Cell       Date:  2018-02-22       Impact factor: 41.582

9.  Clinically applicable deep learning for diagnosis and referral in retinal disease.

Authors:  Jeffrey De Fauw; Joseph R Ledsam; Bernardino Romera-Paredes; Stanislav Nikolov; Nenad Tomasev; Sam Blackwell; Harry Askham; Xavier Glorot; Brendan O'Donoghue; Daniel Visentin; George van den Driessche; Balaji Lakshminarayanan; Clemens Meyer; Faith Mackinder; Simon Bouton; Kareem Ayoub; Reena Chopra; Dominic King; Alan Karthikesalingam; Cían O Hughes; Rosalind Raine; Julian Hughes; Dawn A Sim; Catherine Egan; Adnan Tufail; Hugh Montgomery; Demis Hassabis; Geraint Rees; Trevor Back; Peng T Khaw; Mustafa Suleyman; Julien Cornebise; Pearse A Keane; Olaf Ronneberger
Journal:  Nat Med       Date:  2018-08-13       Impact factor: 53.440

10.  Diabetic retinopathy risk prediction for fundus examination using sparse learning: a cross-sectional study.

Authors:  Ein Oh; Tae Keun Yoo; Eun-Cheol Park
Journal:  BMC Med Inform Decis Mak       Date:  2013-09-13       Impact factor: 2.796

View more
  19 in total

1.  Editorial: The Future of Geriatrics.

Authors:  J E Morley
Journal:  J Nutr Health Aging       Date:  2020       Impact factor: 4.075

2.  Retinal Glaucoma Public Datasets: What Do We Have and What Is Missing?

Authors:  José Camara; Roberto Rezende; Ivan Miguel Pires; António Cunha
Journal:  J Clin Med       Date:  2022-07-02       Impact factor: 4.964

3.  Tailoring Therapy for Children With Neuroblastoma on the Basis of Risk Group Classification: Past, Present, and Future.

Authors:  Wayne H Liang; Sara M Federico; Wendy B London; Arlene Naranjo; Meredith S Irwin; Samuel L Volchenboum; Susan L Cohn
Journal:  JCO Clin Cancer Inform       Date:  2020-10

4.  Capsules for biomedical image segmentation.

Authors:  Rodney LaLonde; Ziyue Xu; Ismail Irmakci; Sanjay Jain; Ulas Bagci
Journal:  Med Image Anal       Date:  2020-11-05       Impact factor: 8.545

5.  Development and validation of MRI-based deep learning models for prediction of microsatellite instability in rectal cancer.

Authors:  Wei Zhang; Hongkun Yin; Zixing Huang; Jian Zhao; Haoyu Zheng; Du He; Mou Li; Weixiong Tan; Song Tian; Bin Song
Journal:  Cancer Med       Date:  2021-05-08       Impact factor: 4.452

6.  Deep Learning of Ultrasound Imaging for Evaluating Ambulatory Function of Individuals with Duchenne Muscular Dystrophy.

Authors:  Ai-Ho Liao; Jheng-Ru Chen; Shi-Hong Liu; Chun-Hao Lu; Chia-Wei Lin; Jeng-Yi Shieh; Wen-Chin Weng; Po-Hsiang Tsui
Journal:  Diagnostics (Basel)       Date:  2021-05-27

7.  Identification of a Tumor Microenvironment-relevant Gene set-based Prognostic Signature and Related Therapy Targets in Gastric Cancer.

Authors:  Wang-Yu Cai; Zi-Nan Dong; Xiao-Teng Fu; Ling-Yun Lin; Lin Wang; Guo-Dong Ye; Qi-Cong Luo; Yu-Chao Chen
Journal:  Theranostics       Date:  2020-07-09       Impact factor: 11.556

8.  Prediction of clinically relevant Pancreatico-enteric Anastomotic Fistulas after Pancreatoduodenectomy using deep learning of Preoperative Computed Tomography.

Authors:  Wei Mu; Chang Liu; Feng Gao; Yafei Qi; Hong Lu; Zaiyi Liu; Xianyi Zhang; Xiaoli Cai; Ruo Yun Ji; Yang Hou; Jie Tian; Yu Shi
Journal:  Theranostics       Date:  2020-08-01       Impact factor: 11.556

9.  Usefulness of Denoising Process to Depict Myopic Choroidal Neovascularisation Using a Single Optical Coherence Tomography Angiography Image.

Authors:  Yuka Sawai; Manabu Miyata; Akihito Uji; Sotaro Ooto; Hiroshi Tamura; Naoko Ueda-Arakawa; Yuki Muraoka; Masahiro Miyake; Ayako Takahashi; Yu Kawashima; Shin Kadomoto; Yasuyuki Oritani; Kentaro Kawai; Kenji Yamashiro; Akitaka Tsujikawa
Journal:  Sci Rep       Date:  2020-04-10       Impact factor: 4.996

10.  ALICE: a hybrid AI paradigm with enhanced connectivity and cybersecurity for a serendipitous encounter with circulating hybrid cells.

Authors:  Kok Suen Cheng; Rongbin Pan; Huaping Pan; Binglin Li; Stephene Shadrack Meena; Huan Xing; Ying Jing Ng; Kaili Qin; Xuan Liao; Benson Kiprono Kosgei; Zhipeng Wang; Ray P S Han
Journal:  Theranostics       Date:  2020-09-02       Impact factor: 11.556

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.