Literature DB >> 31692958

Application of deep learning to the classification of uterine cervical squamous epithelial lesion from colposcopy images.

Yasunari Miyagi1,2,3, Kazuhiro Takehara4, Takahito Miyake5.   

Abstract

The aim of the present study was to explore the feasibility of using deep learning as artificial intelligence (AI) to classify cervical squamous epithelial lesions (SIL) from colposcopy images. A total of 330 patients who underwent colposcopy and biopsy by gynecologic oncologists were enrolled in the current study. A total of 97 patients received a pathological diagnosis of low-grade SIL (LSIL) and 213 of high-grade SIL (HSIL). An original AI-classifier with 11 layers of the convolutional neural network was developed and trained. The accuracy, sensitivity, specificity and Youden's J index of the AI-classifier and oncologists for diagnosing HSIL were 0.823 and 0.797, 0.800 and 0.831, 0.882 and 0.773, and 0.682 and 0.604, respectively. The area under the receiver-operating characteristic curve was 0.826±0.052 (mean ± standard error), and the 95% confidence interval 0.721-0.928. The optimal cut-off point was 0.692. Cohen's Kappa coefficient for AI and colposcopy was 0.437 (P<0.0005). The AI-classifier performed better than oncologists, although not significantly. Although further study is required, the clinical use of AI for the classification of HSIL/LSIL from by colposcopy may be feasible. Copyright: © Miyagi et al.

Entities:  

Keywords:  artificial intelligence; cervical cancer; cervical intraepithelial neoplasia; colposcopy; deep learning

Year:  2019        PMID: 31692958      PMCID: PMC6826263          DOI: 10.3892/mco.2019.1932

Source DB:  PubMed          Journal:  Mol Clin Oncol        ISSN: 2049-9450


Introduction

With current advancements in computer science, artificial intelligence (AI) has made remarkable progress recently. The hypothetical moment in time when AI becomes so advanced that humanity undergoes a dramatic and irreversible change (1) is likely to arrive in this century. AI has already exceeded human experts in the field of games with perfect information, such as Go (2), showing us novel tactics. Therefore, since AI can recognize some information that conventional procedures cannot detect, it may provide a more precise diagnosis in practical medicine. AI may also be able to support clinicians in practical medicine, reducing the time and effort necessary. The aim of the present study was to investigate the feasibility of applying deep learning, a type of AI, for gynecological clinical practice. Uterine cervical cancer continues to be a major public health problem. Cervical cancer is the third most common female cancer and the leading cause of cancer-related mortality among women in Eastern, Western and Middle Africa, Central America, South-Central Asia and Melanesia. New methodologies of cervical cancer prevention should be made available and accessible to women of all countries (3). Colposcopy is a well-established tool for examining the cervix under magnification (4–6). When lesions are treated with acetic acid diluted to 3–5%, colposcopy can detect and recognize cervical intraepithelial neoplasia (CIN). Classification systems, such as the Bethesda System in 2002 are used to categorize lesions as high-grade squamous intraepithelial lesions (HSIL) or low-grade SIL (LSIL) (7,8). HSIL and LSIL were previously referred to as CIN2/CIN3 and CIN1, respectively. In clinical practice, it is important for clinicians to distinguish HSIL from LSIL in biopsy specimens, since further examination or treatment, such as conization, may be required for HSIL. Expert gynecologic oncologists spend much time and effort to provide precise colposcopy findings. For these reasons, we explored whether AI can evaluate colposcopy findings as well as a gynecologic oncologist. In the present study, we applied deep learning with a convolutional neural network to the realm of AI to develop an original classifier for predicting HSIL or LSIL from colposcopy images. Deep learning is becoming very popular among machine learning methods, such as logistic regression (9), naive Bayes (10), nearest neighbor (11), random forest (12) and neural network (13). The classifier program was developed using supervised deep learning with a convolutional neural network (14) that tried to mimic the visual cortex of the mammal brain (15–23), in order to categorize colposcopy images as either HSIL or LSIL. The present study demonstrated the effective use of the AI colposcopy image classifier in predicting HSIL or LSIL by comparing the colposcopic diagnosis to that of gynecologic oncologists.

Patients and methods

Patients

This retrospective study used fully deidentified patient data and was approved by the Institutional Review Board of Shikoku Cancer Center (approval no. 2017-81). This study was explained to the patients, who underwent cervical biopsy by gynecologic oncologists at Shikoku Cancer Center from January 1, 2012 to December 31, 2017. Patients were also directed to a website with additional information, including an opt-out option for the study. The Institutional Review Board of Shikoku Cancer Center approved the opt-out option for patients to choose to withdraw from this study. Gynecologic oncologists at the Shikoku Cancer Center determined the biopsy in routine conventional practice when necessary. A total of 330 patients were enrolled in this study.

Images

Colposcopy images of lesions processed with acetic acid prior to biopsy were captured, cropped to a square and saved in JPEG format. The images were used retrospectively as the input data for deep learning. Gynecologists biopsied the most advanced lesion, the pathological results of which were revealed later.

Preparation for AI

All deidentified images stored offline were transferred to our AI-based system. Each image was cropped to a square and then saved. Twenty percent of the images were randomly selected as the test dataset, and the rest were used as the training dataset. Next, 20% of the training dataset was used as the validation dataset, and the rest was used to train the AI-classifier. Thus, the training, validation and test datasets did not overlap. That way, the AI classifier was trained by a training dataset and simultaneously validated and then tested for the test dataset. The number of training datasets was augmented, as is often done in computer science, in a process known as data augmentation. The training dataset was augmented in this study because the image processing of the arbitrary degrees of rotation can lead to images being included in the same category of different vector data.

AI-classifier

We developed classifier programs using supervised deep learning with a convolutional neural network (14,19). We tested a lot of convolutional neural networks by varying L2 regularization (24,25) and the architectures consisted of a combination of convolution layers with kernels (26–28), pooling layers (29–32), flattened layers (33), linear layers (34,35), rectified linear unit layers (36,37) and a softmax layer (38,39) that demonstrated the probability of LSIL or HSIL from an image (Table I). We also tested the ResNet-50 network (40), which performed very well in the ImageNet Large Scale Visual Recognition Challenge (41) and Microsoft common objects in context MS-COCO (42) competition. We modified the ResNet-50, the first layer of which was replaced with the convolutional layer with a kernel size of 4, stride size of 2, padding size of 2, and input image size of 111×111 pixels, which is the minimum size for the ResNet-50. The last layer of the ResNet-50 was also replaced with a linear layer, followed by a softmax layer with an output with a vector size of 2.
Table I.

Architectures of the top classifier that exhibited the highest accuracy.

LayersSupplementations
Convolution layerOutput channels; 64, Kernel size; 3×3
ReLUN/A
Pooling layerKernel size; 2×2
Convolution layerOutput channels; 64, Kernel size; 3×3
ReLUN/A
Pooling layerKernel size; 2×2
Flatten layerN/A
Linear layerSize; 29
ReLUN/A
Linear layer2
Softmax layerN/A

The convolutional neural network structures, which consisted of 11 layers of convolutional deep learning, were obtained. ReLU, rectified linear units.

Cross-validation (43–45), a powerful method for model selection, was applied to identify the optimal method of machine learning. The suitable number of images for the training data was investigated by evaluating the accuracy and variances using the 5-fold cross-validation method. This calculation procedure reveals the optimal number of training data and can be used to avoid overfitting (46–51), which is a modeling error that occurs when a classifier is too closely fit to a limited set of data points. After the optimal number of training data was obtained, the classifier that showed the best accuracy was selected, as is standard practice in computer science. Conventional colposcopy diagnosis and AI colposcopy diagnosis for test dataset were compared.

Development environment

The following development environment was used in the present study: A Mac running OS X 10.14.3 (Apple, Inc.) and Mathematica 11.3.0.0 (Wolfram Research).

Statistical analysis

The laboratory and AI-classifier data were compared. The two proportion between gynecologic oncologists and the classifier using deep learning was compared by two-proportion z-test. The agreements among the conventional colposcopy, the AI classifier and pathological results were evaluated by Cohen's Kappa (52) coefficients. The formula to calculate Cohen's kappa for two raters is as follows: (Aobserved - Aexpected by chance)/(1- Aexpected by chance) where: Aobserved = the relative observed agreement among raters, Aexpected by chance = the hypothetical probability of chance agreement. Mathematica 11.3.0.0 (Wolfram Research) was used for all statistical analyses.

Results

Profiles of pathological diagnosis and colposcopy

The pathological diagnoses and corresponding number of patients were as follows: LSIL, 97; HSIL, 213; squamous cell carcinoma, 12; microinvasive squamous cell carcinoma, 1; adenocarcinoma, 5; adenocarcinoma in situ, 2. A total of 310 images of both pathological LSIL and HSIL were used, due of the limited number of available images. Among the 213 pathological HSIL cases, 177, 32, 3 and 1 received a conventional colposcopy diagnosis by gynecologists of HSIL, LSIL, invasive cancer and cervicitis, respectively. Among the 97 pathological LSIL cases, 22, 70 and 5 received a conventional colposcopy diagnosis by gynecologists of HSIL, LSIL and cervicitis, respectively (Table II) The accuracy, sensitivity, specificity, positive predictive value, negative predictive value and Youden's J index (53) for HSIL, as determined by gynecologists were 0.797 (247/310), 0.831 (177/213), 0.773 (75/97), 0.889 (117/199), 0.686 (70/102) and 0.604, respectively (Table III).
Table II.

Charactersitics of the 330 patients that underwent colposcopy and biopsy by gynecologic oncologists.

Patient characteristicsPathological HSIL (n=213)Pathological LSIL (n=97)
Age (years)
  Mean ± SD31.66±5.0133.75±8.94
  Median3233
  Range19-4619-62
HPV
  Type 16 positive752
  Type 18 positive52
  Type 16 and 18 positive10
  Positive, but not type 16 or 1812333
  Negative66
  Not available354
Colposcopic diagnosis
  HSIL17722
  LSIL3270
  Cervicitis15
  Invasive cancer30

HSIL, high-grade squamous intraepithelial lesions; LSIL, low-grade squamous intraepithelial lesions; SD, standard deviation.

Table III.

Comparison between gynecologic oncologists and the top classifier using deep learning.

VariableGynecologic oncologistsAI
Accuracy0.797 (247/310)0.823 (51/62)
Sensitivity0.831 (177/213)0.800 (36/45)
Specificity0.773 (75/97)0.882 (15/17)
Positive predictive value0.889 (177/199)0.947 (36/38)
Negative predictive value0.686 (70/102)0.625 (15/24)
Youden's J index0.6040.682

Bracketed data indicates the number of corresponding selected cases/the number of relevant cases. AI, artificial intelligence.

AI-classifier results

The best accuracy for HSIL was 0.823 (51/62), when the number of the augmented training data set was 1,488, the value of L2 regularization 0.175, the number of layers of the architecture 11 (Table II) and the image size 70×70 pixels. The sensitivity, specificity, positive predictive value, negative predictive value and Youden's J index were 0.800 (36/45), 0.882 (15/17), 0.947 (36/38), 0.625 (15/24) and 0.682, respectively (Table III). The accuracy, sensitivity, specificity, positive predictive value, negative predictive value and Youden's J index of the best modified ResNet-50 were 0.790 (49/62), 0.847 (39/46), 0.625 (10/16), 0.867 (39/45), 0.588 (10/17) and 0.472, respectively. The original conventional neural network was better than the modified ResNet-50, although not significantly. There were no differences between the gynecologic oncologists and the best AI in accuracy, sensitivity, specificity, positive or negative predictive value, as determined by a proportional test. Using confidence score, the area under the receiver-operating characteristic curve (ROC) of the best classifier for predicting HSIL was 0.824±0.052 (mean ± SE), and the 95% confidence interval 0.721–0.928. The ROC curve is shown in Fig. 1. The optimal cut-off point was 0.692.
Figure 1.

The receiver-operating characteristic curve of the best classifier for predicting high-grade squamous intraepithelial lesions. The value of the area under the curve is 0.824±0.052 (mean ± standard error) and the 95% confidence interval ranged between 0.721–0.928.

Comparison of AI-classifier with conventional colposcopy

The association among conventional colposcopy, the AI classifier and pathological results for the test dataset that was 20% of patients of both pathological HSIL and LSIL diagnosed by punch biopsy are shown in Tables IV–VI. Cohen's Kappa (52) coefficients of the conventional colposcopy and pathological results, the AI classifier and pathological results, and the conventional colposcopy and the AI classifier were 0.691, 0.561 and 0.437 (all P<0.0001), respectively. All were more than moderate agreements (54). Conventional colposcopy showed better agreement than the AI, although the difference was not significant. Classification took less than 0.2 sec per image.
Table IV.

Conventional colposcopy diagnosis and pathological results of the test data set.

Conventional colposcopy diagnosis

Lesion typeHSILLSIL
Pathological HSIL396
Pathological LSIL215

Cohen's Kappa coefficient was 0.691, P<0.0001. HSIL, high-grade squamous intraepithelial lesions; LSIL, low-grade squamous intraepithelial lesions; AI, artificial intelligence.

Table VI.

Conventional colposcopy diagnosis and AI colposcopy diagnosis for test data set.

AI colposcopy diagnosis

HSILLSIL
Conventional colposcopy HSIL329
Conventional colposcopy LSIL714

Cohen's Kappa coefficient was 0.437, P<0.0005. HSIL, high-grade squamous intraepithelial lesions; LSIL, low-grade squamous intraepithelial lesions; AI, artificial intelligence.

Discussion

We developed a classifier using deep learning with convolutional neural networks using images of cervical SILs to predict the pathological diagnosis. In the present study, the accuracy values achieved by the classifier and by gynecologic oncologists were 0.823 and 0.797, respectively (Table III). The sensitivity values of the classifier and gynecologic oncologists were 0.800 and 0.831, respectively. The specificity values of the classifier and gynecologic oncologists were 0.882 and 0.773, respectively. The accuracy and specificity of the classifier were superior to those of gynecologic oncologists, although the difference was not significant. Only moderate agreement was obtained between conventional colposcopy diagnosis and AI colposcopy diagnosis with the Kappa value of 0.437. McHugh reported that Cohen suggested 0.41 might be acceptable and the Kappa result be interpreted as follows: Values ≤0 as indicating no agreement and 0.01–0.20 as none to slight, 0.21–0.40 as fair, 0.41–0.60 as moderate, 0.61–0.80 as substantial and 0.81–1.00 (54). Thus, the Kappa value of 0.437 was acceptable. But, AI colposcopy that might have a potential would not be an alternative to conventional colposcopy without further studies. Several reports have used AI (55) for deep learning with convolutional neural networks in medicine (56). The accuracy values of this method with deep learning have been published and include 0.997 for the histopathological diagnosis of breast cancer (57), 0.90–0.83 for the early diagnosis of Alzheimer's disease (58), 0.83 for urological dysfunctions (59), 0.72 (60) and 0.50 (61) for colposcopy, 0.68–0.70 for localization of rectal cancer (62), 0.83 for the diagnostic imaging of orthopedic trauma (63), 0.98 for the morphological quality of blastocysts and evaluation by an embryologist (64), 0.65 for predicting live birth without aneuploidy from a blastocyst image (65) and 0.64–0.88 for predicting live birth from a blastocyst image of patients by age (66). Several studies have reported a limitation of conventional colposcopy. An investigation of the accuracy of colposcopically-directed biopsy reported a total biopsy failure rate, comprising both non-biopsy and incorrect selection of biopsy site, of 0.20 in CIN1, 0.11 in CIN2 and 0.09 in CIN3 (67). The colposcopic impression of high-grade CIN had a sensitivity of 0.54 and a specificity of 0.88, as determined by 9 expert colposcopists in 100 cervigrams (68). The sensitivity of an online colpophotographic assessment of high-grade disease (CIN2 and CIN3) by 20 colposcopists was 0.39 (69). Thus, conventional colposcopy does not provide good sensitivity, even by colposcopists. By contrast, the accuracy and sensitivity reported in this study for predicting HSIL from colposcopy images using deep learning was 0.823 and 0.800, respectively, which appears favorable. Since the classifier was not trained in colposcopy findings such as acetowhite epithelium, mosaic, punctuation, it may recognize some features of cervical SILs by itself in high-dimensional space. It is possible that the AI-classifier may recognize features that colposcopists do not, such as complexity of the shape of the lesion, relative or absolute brightness of acetowhite, distribution of punctuation density, and quantitative marginal evaluation of borders. The pathological results were obtained and defined by punch biopsy in this study, as it was not ethically recommended for patients with LSIL (CIN1) diagnosed by colposcopy to undergo conization or hysterectomy. If the pathological results were defined by conization or hysterectomy, the advanced lesion would have been be revealed, and thus both conventional colposcopy and the AI classifier may have demonstrated different results. In the present study, we only tried to compare the effectiveness of AI with that of conventional colposcopy for SIL. When the AI will be used to predict more advanced diseases, such as squamous cell carcinoma, adenocarcinoma and AIS, the pathological diagnosis should be provided by not punch biopsy but conization or hysterectomy. In clinical practice, it is important for clinicians to distinguish HSIL from LSIL in biopsy specimens. Further examination or treatment, such as conization, may be required for HSIL. When a reliable classifier indicates HSIL from colposcopy images in clinical practice, the clinician should consider biopsy. The accuracy values of the classifier and gynecologists for detecting HSIL were 0.823 and 0.797, respectively. The classifier might help untrained clinicians to avoid or reduce the risks of overlooking HSIL. When the AI-classifier can perform higher in terms of accuracy, sensitivity and specificity for classifying HSIL/LSIL, clinicians will be able to perform more precise practice, referencing AI. Furthermore, a gynecologist could reduce the time and effort it takes to become a colposcopy expert and, as a result, improve in other skills, training and activities. The architecture of the neural network has progressed. The LeNet study published in 1998 (70) consisted of 5 layers. AlexNet, published in 2012 (38), consisted of 14 and Google Net, published in 2014 (35), was constructed from a combination of micro networks. ResNet-50, published in 2015 (41), consisted of modules with a shortcut process. The Squeeze-and-Excitation Networks were published in 2017 (71). AI used for image recognition is still being developed. Progress in AI will allow us to achieve better results. Image information is one of the parameters that need to be investigated. Only 15×15 pixels are used to detect cervical cancer (72). In a colposcopy study (61), it was reported that the accuracy for images of 150×150 pixels was better than those for 32×32 or 300×300 pixels. Hence, image size remains an issue. We used 70×70 and 111×111 pixels for our images, in order to use the original neural networks and the modified ResNet-50, respectively. The original conventional neural network was better than the modified ResNet-50, although not significantly. We believe that a pixel size of 70×70 falls within the acceptable range. Regularization values are also important parameters for constructing a good classifier that avoids overfitting. If the regularization value is too low, overfitting occurs. If the value is too large, the classifier will not be trained well. Choosing the appropriate number of training datasets is also very important. If the number of training datasets is too high, the accuracy will be lower and more variances will be observed. The validation dataset, as well as L2 regularization, also prevent overfitting. The appropriate number of training datasets must be achieved to obtain a good classifier. More varied patterns of images may be needed for datasets. Ordinarily, 500–1,000 images are prepared for each class during image classification with deep learning (61,73). Such a large number of datasets will improve the accuracy and specificity of the classifier with deep learning. In the present study, a classifier was developed based on deep learning, which used images of uterine cervical SILs to predict pathological HSIL/LSIL. Its accuracy was 0.823. Although further study may be required to validate the classifier, we demonstrated that AI may have a clinical use in colposcopic examinations and may provide benefits to both patients and medical personnel.
Table V.

AI colposcopy diagnosis and pathological result for test data set.

AI colposcopy diagnosis

Lesion typeHSILLSIL
Pathological HSIL369
Pathological LSIL314

Cohen's Kappa coefficient was 0.561, P<0.0001. HSIL, high-grade squamous intraepithelial lesions; LSIL, low-grade squamous intraepithelial lesions; AI, artificial intelligence.

  33 in total

Review 1.  Logistic regression and artificial neural network classification models: a methodology review.

Authors:  Stephan Dreiseitl; Lucila Ohno-Machado
Journal:  J Biomed Inform       Date:  2002 Oct-Dec       Impact factor: 6.317

2.  Index for rating diagnostic tests.

Authors:  W J YOUDEN
Journal:  Cancer       Date:  1950-01       Impact factor: 6.860

3.  Sensitivity analysis in bayesian classification models: multiplicative deviations.

Authors:  M Ben-Bassat; K L Klove; M H Weil
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  1980-03       Impact factor: 6.226

Review 4.  The Lower Anogenital Squamous Terminology Standardization Project for HPV-Associated Lesions: background and consensus recommendations from the College of American Pathologists and the American Society for Colposcopy and Cervical Pathology.

Authors:  Teresa M Darragh; Terence J Colgan; J Thomas Cox; Debra S Heller; Michael R Henry; Ronald D Luff; Timothy McCalmont; Ritu Nayar; Joel M Palefsky; Mark H Stoler; Edward J Wilkinson; Richard J Zaino; David C Wilbur
Journal:  J Low Genit Tract Dis       Date:  2012-07       Impact factor: 1.925

5.  Automation of Detection of Cervical Cancer Using Convolutional Neural Networks.

Authors:  Vidya Kudva; Keerthana Prasad; Shyamala Guruvare
Journal:  Crit Rev Biomed Eng       Date:  2018

6.  Worldwide burden of cervical cancer in 2008.

Authors:  M Arbyn; X Castellsagué; S de Sanjosé; L Bruni; M Saraiya; F Bray; J Ferlay
Journal:  Ann Oncol       Date:  2011-04-06       Impact factor: 32.976

7.  Automated Melanoma Recognition in Dermoscopy Images via Very Deep Residual Networks.

Authors:  Lequan Yu; Hao Chen; Qi Dou; Jing Qin; Pheng-Ann Heng
Journal:  IEEE Trans Med Imaging       Date:  2016-12-21       Impact factor: 10.048

8.  Feasibility of artificial intelligence for predicting live birth without aneuploidy from a blastocyst image.

Authors:  Yasunari Miyagi; Toshihiro Habara; Rei Hirata; Nobuyoshi Hayashi
Journal:  Reprod Med Biol       Date:  2019-02-19

9.  Interrater reliability: the kappa statistic.

Authors:  Mary L McHugh
Journal:  Biochem Med (Zagreb)       Date:  2012       Impact factor: 2.313

10.  Deep learning as a tool for increased accuracy and efficiency of histopathological diagnosis.

Authors:  Geert Litjens; Clara I Sánchez; Nadya Timofeeva; Meyke Hermsen; Iris Nagtegaal; Iringo Kovacs; Christina Hulsbergen-van de Kaa; Peter Bult; Bram van Ginneken; Jeroen van der Laak
Journal:  Sci Rep       Date:  2016-05-23       Impact factor: 4.379

View more
  10 in total

1.  Computer-aided diagnostic system based on deep learning for classifying colposcopy images.

Authors:  Lu Liu; Ying Wang; Xiaoli Liu; Sai Han; Lin Jia; Lihua Meng; Ziyan Yang; Wei Chen; Youzhong Zhang; Xu Qiao
Journal:  Ann Transl Med       Date:  2021-07

2.  The application of deep learning based diagnostic system to cervical squamous intraepithelial lesions recognition in colposcopy images.

Authors:  Chunnv Yuan; Yeli Yao; Bei Cheng; Yifan Cheng; Ying Li; Yang Li; Xuechen Liu; Xiaodong Cheng; Xing Xie; Jian Wu; Xinyu Wang; Weiguo Lu
Journal:  Sci Rep       Date:  2020-07-15       Impact factor: 4.379

3.  The challenges of colposcopy for cervical cancer screening in LMICs and solutions by artificial intelligence.

Authors:  Peng Xue; Man Tat Alexander Ng; Youlin Qiao
Journal:  BMC Med       Date:  2020-06-03       Impact factor: 8.775

4.  Using Dynamic Features for Automatic Cervical Precancer Detection.

Authors:  Roser Viñals; Pierre Vassilakos; Mohammad Saeed Rad; Manuela Undurraga; Patrick Petignat; Jean-Philippe Thiran
Journal:  Diagnostics (Basel)       Date:  2021-04-17

5.  Deep learning model for classifying endometrial lesions.

Authors:  YunZheng Zhang; ZiHao Wang; Jin Zhang; CuiCui Wang; YuShan Wang; Hao Chen; LuHe Shan; JiaNing Huo; JiaHui Gu; Xiaoxin Ma
Journal:  J Transl Med       Date:  2021-01-06       Impact factor: 5.531

Review 6.  Artificial Intelligence in Cervical Cancer Screening and Diagnosis.

Authors:  Xin Hou; Guangyang Shen; Liqiang Zhou; Yinuo Li; Tian Wang; Xiangyi Ma
Journal:  Front Oncol       Date:  2022-03-11       Impact factor: 6.244

Review 7.  Smartphone-Based Visual Inspection with Acetic Acid: An Innovative Tool to Improve Cervical Cancer Screening in Low-Resource Setting.

Authors:  Jana Sami; Sophie Lemoupa Makajio; Emilien Jeannot; Bruno Kenfack; Roser Viñals; Pierre Vassilakos; Patrick Petignat
Journal:  Healthcare (Basel)       Date:  2022-02-18

8.  Radiomics Diagnostic Tool Based on Deep Learning for Colposcopy Image Classification.

Authors:  Yuliana Jiménez Gaona; Darwin Castillo Malla; Bernardo Vega Crespo; María José Vicuña; Vivian Alejandra Neira; Santiago Dávila; Veronique Verhoeven
Journal:  Diagnostics (Basel)       Date:  2022-07-12

9.  Application of deep learning to the classification of uterine cervical squamous epithelial lesion from colposcopy images combined with HPV types.

Authors:  Yasunari Miyagi; Kazuhiro Takehara; Yoko Nagayasu; Takahito Miyake
Journal:  Oncol Lett       Date:  2019-12-12       Impact factor: 2.967

10.  The Application of Artificial Intelligence-Assisted Colposcopy in a Tertiary Care Hospital within a Cervical Pathology Diagnostic Unit.

Authors:  Aleksandra Zimmer-Stelmach; Jan Zak; Agata Pawlosek; Anna Rosner-Tenerowicz; Joanna Budny-Winska; Michal Pomorski; Tomasz Fuchs; Mariusz Zimmer
Journal:  Diagnostics (Basel)       Date:  2022-01-04
  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.