Literature DB >> 34178558

Deep Metric Learning for Cervical Image Classification.

Anabik Pal1, Zhiyun Xue1, Brian Befano2, Ana Cecilia Rodriguez3, L Rodney Long1, Mark Schiffman3, Sameer Antani1.   

Abstract

Cervical cancer is caused by the persistent infection of certain types of the Human Papillomavirus (HPV) and is a leading cause of female mortality particularly in low and middle-income countries (LMIC). Visual inspection of the cervix with acetic acid (VIA) is a commonly used technique in cervical screening. While this technique is inexpensive, clinical assessment is highly subjective, and relatively poor reproducibility has been reported. A deep learning-based algorithm for automatic visual evaluation (AVE) of aceto-whitened cervical images was shown to be effective in detecting confirmed precancer (i.e. direct precursor to invasive cervical cancer). The images were selected from a large longitudinal study conducted by the National Cancer Institute in the Guanacaste province of Costa Rica. The training of AVE used annotation for cervix boundary, and the data scarcity challenge was dealt with manually optimized data augmentation. In contrast, we present a novel approach for cervical precancer detection using a deep metric learning-based (DML) framework which does not incorporate any effort for cervix boundary marking. The DML is an advanced learning strategy that can deal with data scarcity and bias training due to class imbalance data in a better way. Three different widely-used state-of-the-art DML techniques are evaluated- (a) Contrastive loss minimization, (b) N-pair embedding loss minimization, and, (c) Batch-hard loss minimization. Three popular Deep Convolutional Neural Networks (ResNet-50, MobileNet, NasNet) are configured for training with DML to produce class-separated (i.e. linearly separable) image feature descriptors. Finally, a K-Nearest Neighbor (KNN) classifier is trained with the extracted deep features. Both the feature quality and classification performance are quantitatively evaluated on the same data set as used in AVE. It shows that, unlike AVE, without using any data augmentation, the best model produced from our research improves specificity in disease detection without compromising sensitivity. The present research thus paves the way for new research directions for the related field.

Entities:  

Keywords:  Automated cervical visual examination; cervical cancer; deep metric learning; siamese network

Year:  2021        PMID: 34178558      PMCID: PMC8224396          DOI: 10.1109/access.2021.3069346

Source DB:  PubMed          Journal:  IEEE Access        ISSN: 2169-3536            Impact factor:   3.367


INTRODUCTION

Cervical cancer is a major cause of premature female morbidity with over half a million new cases and over three hundred thousand deaths reported in 2018.[1] Universally, this disease is caused by persistent infection with one or more from a dozen oncogenic types of the human papillomavirus (HPV). Early detection of HPV-induced precancer (the precursor to cancer) and providing appropriate treatment where necessary can reduce suffering and premature death. However, there is a significant scarcity of clinical and gynecological services and expertise as well as a lack of sufficient access to effective low-cost cervical screening programs in the low and middle-income countries (LMIC). This shortage correlates highly with regions where the highest death rates have been reported. The VIA method (Visual Inspection with Acetic Acid) is a low-cost and readily available screening method. A weak (3%-5%) acetic acid solution is applied to the cervix region which is then visually assessed by an expert. Whitening of cervical tissue around the transformation zone [1] indicates focal HPV infections. Appropriate treatment of the cervix or referral for further evaluation is recommended by the clinician. Although VIA is cheap and easily accessible, low reliability, accuracy, and high inter-observer variability have been reported [2]. Clinical colposcopes help in improving VIA performance with better illumination and optical magnification of the cervix region. However, they are very expensive and not commonly available in all settings. Moreover, these assessments also suffer from high intra- and inter-observer variability [3]. These challenges present an opportunity for research in developing powerful image analytics algorithms in an automated low-cost assistive screening system that is accurate, reliable, and effective. There are several challenges towards achieving this goal. First, there is a lack of an imaging standard. We find that images are often taken inconsistently, with varied illumination, poor focus, high specular reflection, and imperfect color tone [4], [5]. Designing hand-crafted statistical features for addressing these variables is limiting and error-prone. Modern deep learning-based classification algorithms can apply data-driven strategies to deal with it in a better way. However, there is a naturally occurring high class-imbalance in screening data due to low disease prevalence in the general population with many more controls (normal class) than cases (abnormal class) -typical ratios of 99:1 or higher in controls to cases are not uncommon. This makes the tasks more challenging. Recall that, while DCNN is an important addition for representation learning in computer vision [6], [7], these networks are trained in an end-to-end manner, and during training adjust weight matrices within the network layers. Images are processed through the layers and produce prediction maps for all possible classes. The training error is computed based on the ground-truth probability map for every class, with a differentiable error function, such as Cross-Entropy, Mean Square Error, etc. However, key drawbacks of training a DCNN with classification loss are that it is prone to bias toward the majority class, which tends to be comprised of images from normal women. Such an imbalance does not guarantee, without an appropriate training strategy, that the image embedding (i.e. feature map obtained from the last fully connected layer before the loss layer) are linearly separable. While, it may sometimes appear that we obtain good classification accuracy, due to the classifier’s biasedness toward majority classes, we are often unable to get good feature representation to support generalization to unseen data. Previously, Faster-RCNN, a deep learning-based Automatic Visual Evaluation (AVE) [8] method was proposed for detecting precancer cases. AVE uses a region proposal algorithm (Faster-RCNN) to localize the cervix boundary prior to the classification module. In developing AVE, the data skew was retained to increase statistical power for epidemiologic analyses but reduced to a 3:1 ratio of controls to cases. Consequently, in selecting absolute precancer cases the number of samples was also limited. With fewer data available for training, synthetic data augmentation was used during network training to overcome its impact on learning and classification performance. As a result, the trained model was likely over-fit to the data and less likely to adapt to naturally occurring variations in cervical image appearance and disease prevalence. In order to advance the prior AVE effort and pursue the first step toward addressing data skew, we develop a new method that operates on the full cervix image. We propose to train the convolutional neural network with deep metric learning (DML) for producing class-separated feature representation of the cervical images. Finally, a K-Nearest neighbor classifier is built with the deep features. The key contributions of this paper are as follows. We present a pioneering approach using deep metric learning for cervical precancer detection aimed at naturally occurring disease prevalence. We analyze the linear separability of learned image features both quantitatively and qualitatively. A detailed analysis of experimental results is conducted which demonstrates that the method improves specificity in disease detection without compromising sensitivity and paves the way for new research directions. The organization of the rest of the paper is as follows: Section II provides background on state-of-the-art approaches on cervical image analysis. A discussion about deep metric learning and the experimented approaches are available in Section III. The experimental setup and the analysis of experimental results are presented in Section IV and Section V. Finally, Section VI concludes the paper.

RELATED LITERATURE

The potential for automatic analysis of digital cervical images in revolutionizing screening for precancers has motivated the development of several automatic and semi-automatic image analysis algorithms. These include algorithms for anatomical landmark detection [9], cervix region detection [10], [11], cervix type detection [12], pre-cancerous lesion detection-segmentation [13-15] and disease diagnosis [16-18]. Since our main concern is detecting precancer (or worst disease condition) in cervical images, we restrict our literature review to the topically relevant algorithms. Early cervical image classification research mainly focused on the development of robust features to represent cervical images and classifier development. Commonly, hand-crafted image features were used as cervical image descriptors, such as (a) filter bank-based texture features, (b) pyramid color histogram features (in L*a*b* color space), (c) pyramid histogram of oriented gradients (PHOG), and (d) pyramid histogram of local binary patterns (PLBP). These features were subsequently used for developing classification algorithms, such as χ2 distance, support vector machine (SVM), random forest (RF), gradient boosting decision tree (GBDT), AdaBoost, logistic regression (LR), multi-layer perceptron (MLP), and k-Nearest Neighbors (kNN) [16], [17], [19-21]. Some approaches extracted features from the cervix region-of-interest (RoI) detected at an earlier stage of the algorithm [16], [17]. However, these hand-crafted color and texture features were rarely sufficiently robust in representing cervical images due to high variability in image quality and appearance. The variability is most confounding in color and object illumination which are critical for disease discrimination, but also includes focus [4], the region of interest coverage, the imaging device, time that the image was taken after application of acetic acid, and geographic region [22]. This has resulted in data-driven automatic supervised representation learning algorithms becoming an attractive choice for computer vision researchers [8], [23-29]. Training a DCNN model from scratch was proposed in [23]. Multimodal learning [24], where image data along with clinical records are processed together has also been attempted. Multi-scale CNN are proposed in [25], [26]. Multi-CNN decision feature integration is used in [27]. [28] proposed to use the Deep Belief network. Object detection networks are employed in [8], [29]. All these approaches focus on developing a discriminating model, or classifier, from the raw color intensity matrix of the input images. In contrast, our research focuses on cervical image representation with deep metric learning.

METHODS

The deep metric learning (DML) is a robust technique that can address two limitations of commonly used deep classification networks- (i) biasedness towards majority class [30-32] and (ii) over-fitted model development due to data scarcity [33], [34]. The training strategy of DML aims to produce the image embedding in such a way that they are closer if the images are sampled from the same class and distant otherwise and thus produces class-separated image representation. Also, unlike classification model training, the training loss is computed based on the embedding obtained from multiple images. In the literature, several DML approaches are proposed which are broadly designed based on the image sampling strategy, embedding distance computation, loss computation etc. [30-32]. In this paper, we develop a DML based framework (see Fig 1 (a)) for cervical image classification. In the proposed framework, firstly, the DML is performed with the training images and their labels. The learning objective of DML is to produce a deep model which can generate class-discriminating feature vectors from the training images. Note that the deep network does not contain any classification layer during DML (see Fig 1(b)). In the next stage, the trained DML model serves as a feature extractor and extracted feature vectors are then used to build a K-Nearest Neighbor (KNN) classifier. During the test phase, the embeddings of the test images are obtained from the trained DML network and then their class labels are predicted from the trained KNN. We opt for a KNN classifier since the size of the training data is small and the features are expected to be linearly separable.
FIGURE 1.

(a) Block diagram of the proposed system. Upper part denotes two-step training phase and lower part denotes test phase. (b) Block diagram of deep metric learning (DML). Images and their class labels are inputted and the mini-batch loss is computed based on the image embeddings.

In this paper, we vary learning objective functions (called loss functions) for deep model development with the DML. The loss functions associated with the chosen DML algorithms are described in the following paragraphs.

DEEP METRIC LEARNING WITH CONTRASTIVE LOSS

In this approach, a mini-batch is constructed with a randomly sampled pair of images. If the two images are sampled from the same class then the pair is called positive pair and if the images are sampled from different classes then the pair is called negative pair. The distance between a positive pair is called positive distance and the distance between a negative pair is called negative distance. The training loss is designed in a way such that the positive distance is minimized and the negative distance is maximized. Mathematically, the contrastive loss (L) is defined as: where m denotes the upper limit of positive distance, m denotes the lower limit of negative distance, d denotes positive distance, d denotes negative distance and .

DEEP METRIC LEARNING WITH N-PAIR EMBEDDING LOSS

Suppose, X1, Y1, X2, Y2 …, X, Y are N-pair of images sampled from N different classes where X, Y are images from i class. The training loss between two i class images X and Y is given by where f is the feature vector of image A; and representation transpose operation. The mini-batch loss is computed as the mean of all N classes, i.e., . In the present work, N = 2 as we are dealing with binary classification problem.

DEEP METRIC LEARNING WITH BATCH HARD SAMPLING

In this approach, P classes[2] are randomly chosen and from every class, S images are randomly sampled. In a minibatch, the loss function considers the hard samples i.e. the maximum of intra-class (or positive) distances and minimum of inter-class (or negative) distances. The training loss is designed in a way such that it is decreasing the intra-class distance as well as increasing inter-class distance. Mathematically, the Batch-hard sampling loss (L) is defined as: where m is a predefined threshold, f denotes feature vector of an image, D(x, y) represents distance between x and y and .

EXPERIMENTAL SETUP

DATA SET DESCRIPTION

This paper uses Cervigram®[3] image data set used in AVE research [8]. Every image in the data set was labeled either as a case (disease) or control (non-disease) based on the following diagnostic information: HPV status, naked-eye visual impression, colposcopic impression, cytological findings, histopathological analysis outcome. A sample image from both case and control class is shown in Fig 2.
FIGURE 2.

Samples of cervical images from the present data set. Left image from Control class. Right image from Case class.

In [8], the data set was partitioned into three nonoverlapping subsets: training, validation, and hold-out test sets, respectively. In our work, we make a random disjoint split of the training data into training and val1 data. The previous AVE validation data is termed val2. The val1 is used for parameter selection in the DML training. The val2 (i.e. validation data of [8]) is used for K value selection during classifier model building. The hold-out test data, which is the same as one used in AVE research, is used for comparing the classification performance with AVE. Details of data set splits are given in Table 1.
TABLE 1.

Age-stratified data set splits. The entries in the table denote the number of images.

Age-GroupTrainingVal1Val2Hold-out Test
ConrolCaseConrolCaseConrolCaseConrolCase
<2536125224992419
25-49324119532215358501045
>49115262646515223220
Total475157842824282817485

DEEP NETWORKS AND TRAINING STRATEGIES

Three state-of-the-art pre-trained networks, namely, ResNet-50 [35], NasNetMobile [36] and Mobile-Net [37] are selected as backbone networks. First, the softmax classification layer is removed from each backbone network. Then an L2 normalization layer is used after the last feature layer of the networks. Finally, the networks are trained with the chosen DML algorithms (Section III). After training, the L2-normalized output vector obtained from the trained network is used as the image embedding. We vary the parameters associated with DML training best results are found when the DML are trained with learning rate = 0.002, weight decay = 1e – 6, and momentum = 0.9, epoch = 50. The DML algorithms built with constrictive loss and batch hard loss need to set loss function parameters. In this paper, we vary these parameters and receive best performance for following parameters (a) constrictive loss: m = 0 and m = 0.25, (b) batch hard loss: S = 8, m = 0.25.

BASELINE ALGORITHMS

The state-of-the-art pre-trained models developed using the ImageNet data are our initial choice as the baseline feature extractor networks. Note that the limitation of classification networks for imbalanced data set is our key concern. Our next baseline network is the fine-tuned binary classification network with our data. For this model development, we use binary cross-entropy loss minimization for training this classification network. The performance of the chosen DML algorithms is compared with these baselines.

PERFORMANCE EVALUATION

The proposed system has two steps: (1) Training deep model for linearly separable image embedding extraction, and (2) Classifier model development. We evaluate both steps separately. The evaluation scheme is described in the following subsections.

EMBEDDING QUALITY ASSESSMENT

We assess the quality of the feature embedding using t-SNE plots [38] visualization. The t-SNE converts the high dimensional data (here image embedding) into 2-D data vectors which help to visualize them in a 2D-plane. It is a very popular choice for data visualization in spite of its limitations. The t-SNE plot provides only a geometric interpretation of separation in the embedding at the cost of significant information loss in reducing high-dimensional data into a 2D vector. Moreover, small differences between feature vectors cannot be determined from the plot. To offset this limitation, we also propose using the following two quantitative measures for assessing embedding quality.

MEAN K-PRECISION

The K-precision of a test sample T is given by the ratio , where k is the number of nearest neighbors of T of the same class selected among total K nearest neighbors from the training data. The mean K-precision is the mean of K-precision for all test images. The value of K-Precision lies in the range [0,1] and a higher value represents better performance.

CLASS-WISE MEAN N-PRECISION

Suppose, in the training data, there are N controls and N cases. The N-precision for a test data sample T belonging to the control class is given by the ratio , where n is the number of nearest neighbors of T in the training set which belongs to control class. Similarly, N-precision for a test data sample T belonging to the case class is given by the ratio , where n is the number of nearest neighbors of T in the training set which belongs to case class. The value of N-precision lies in range [0,1] and the higher value represent better performance. For a perfect model, all test data will have N-precision equals 1.

CLASSIFICATION PERFORMANCE EVALUATION

The classification performance is evaluated using class-wise accuracy. Class-wise accuracy is defined by the percentage of correct classifications achieved by the proposed model for each class. Note that the case accuracy refers to sensitivity/recall and the control accuracy refers to specificity. These two performance measures together can provide an idea about the biasedness of the model toward the majority class.

RESULTS AND DISCUSSION

The training loss and validation loss improvement for chosen DML algorithms are shown in Fig 3. According to Fig 3, the DML algorithm trained with Batch-hard loss has closer training and validation loss (val1) at the end of the training.
FIGURE 3.

Loss improvement during DML training.

A detailed analysis of the experimental results is given in this section. We divide the discussion into four sections. The first section discusses the separability of the image embedding for the chosen algorithms. Next, we discuss the effectiveness of the K-Nearest Neighbour classifier. Then, the performance of the algorithm on the hold-out test data is presented. Finally, the performance of the best model is compared with the state-of-the-art AVE results [8].

EMBEDDING QUALITY ASSESSMENT

Fig 4 shows 2D t-SNE plots of the image embedding obtained from the considered competing approaches. In the first row, i.e., subfigures (a, b, c), show that the images are poorly separated when features from the ImageNet pre-trained networks are used. The second row, i.e., subfigures (d, e, f), shows that features from the fine-tuned models increase the separability. Finally, the last three rows, i.e., subfigures (g-o), demonstrate the power of DML which produces increasingly well-separated training image feature representations. We find that for many scenarios, due to the inability of producing fully generalized models with the chosen techniques, val2 images are not guaranteed to be closer to the appropriate class. However, based on our experiments, we can assert that the DML algorithm has the potential to deal with the current image classification task in a better way.
FIGURE 4.

The t-SNE plots of feature embeddings.

The mean K-Precision for three different K values for all competing approaches for val2 data are given in Table 2. According to Table 2, the performance after fine-tuning with the training data is markedly improved over the pre-trained model. We also note that the performance of the deep networks built with contrastive loss based DML is comparable with the fine-tuned model. However, we see that the networks built with N-pair loss and Batch-hard sampling strategy outperform these competing methods.
TABLE 2.

Mean K-Precision in val2 data set. Network-wise best performing Mean K-Precision for different values of K are bold-faced.

Neighbour (K)NetworkMethod
Pre-trainedFine-tunedContrastiveN-Pair EmbeddingBatch-hard
1ResNet-500.65120.90430.89200.91050.9198
MobileNet0.70990.91050.90740.91670.9105
NasNet0.66360.87040.91050.92280.9321
3ResNet-500.66870.89920.89400.91150.9198
MobileNet0.66360.90430.89610.91670.9105
NasNet0.66050.87550.89610.92080.9321
5ResNet-500.67840.89260.89320.91050.9198
MobileNet0.66110.90310.89510.91670.9105
NasNet0.64510.87160.89630.92040.9321
The mean N-Precision values of the case and control for all competing methods for val2 data are listed in Table 3. Here, we see that in terms of separability the embeddings obtained from the fine-tuned deep models are much better than the respective pre-trained models. The DML algorithms noticeably improve the mean N-Precision values for both cases and control over the fine-tuned models. Our experimental results also show that the networks trained with N-Pair embedding loss or Batch-hard loss minimization produce much better mean N-Precision values for cases than the networks trained with contrastive loss minimization. This demonstrates that the deep model developed with DML algorithm with Batch-hard or N-Pair Embedding loss minimization is valuable for producing separable image embedding for our classification task. We see this as a significant finding because real-world data are likely to be highly imbalanced with many more controls than cases. It is desirable to build classification models that adopt strategies built around this precondition thereby resulting in realistic and usable decisions.
TABLE 3.

Class-wise Mean N-Precision in val2 data set. The best performing feature representation method is chosen based on the average of Case and Control’s Mean N-precision. Network-wise best performing feature representation methods are bold-faced.

NetworkPre-trainedFine-tunedContrastiveN-Pair EmbeddingBatch-hard
ControlCaseControlCaseControlCaseControlCaseControlCase
ResNet-500.75570.24580.88140.42730.97620.55930.97900.81190.98040.8361
MobileNet0.73830.33050.87000.50270.96790.63850.97500.87250.97500.8482
NasNet0.73810.32620.87660.54430.97140.64610.97090.88460.96960.9088

CLASSIFICATION PERFORMANCE

The class-wise K-Nearest Neighbor (KNN) classification accuracy for three different values of K in val2 data are shown in Table 4. Here, we see that the KNN classifiers trained with the pre-trained deep models are not suitable for our task. The performance of the DML algorithm trained with contrastive loss is comparable with fine-tuned models. For both approaches, keeping the network fixed results in good accuracy for controls but is found to be poor for cases. This is an indication of the classifier’s bias towards the majority class. Again, we see that for every network that was studied, the DML algorithm trained with N-pair embedding loss or Batch-hard loss is better at overcoming data imbalance and consequently the diminishing classifier bias. We surmise that the potential source of bias might be due to the model over-fitting to the training data. Finally, Table 4 shows that the NasNet model trained with Batch-hard loss minimization is our best deep model, and 1-NN (K = 1) can be considered as the best classifier because higher values of K increases the complexity and is unable to improve the classification performance.
TABLE 4.

KNN classification accuracy for val2 data set in percentage (%). The best performing classification model is chosen based on the average of Case and Control’s classification accuracy. The bold-faced numbers represent the network-wise best performing models for different values of K.

KNetworkPre-trainedFine-tunedContrastiveN-Pair EmbeddingBatch-hard
ControlCaseControlCaseControlCaseControlCaseControlCase
K=1ResNet-5078.9324.3995.4575.6196.2868.2994.6376.8394.6384.15
MobileNet80.9941.4695.0479.2793.3973.1792.1586.5992.5682.93
NasNet76.4536.5991.7473.1795.0468.2991.7489.0291.3291.46
K=3ResNet-5088.4318.2996.2873.1796.6968.2995.8780.4994.6384.15
MobileNet82.6432.9395.4575.6194.6375.6192.1586.5992.5682.93
NasNet84.3030.4993.3976.8395.8768.2991.7489.0291.3291.46
K=5ResNet-5095.0410.9895.0471.9596.6968.2994.2180.4994.6384.15
MobileNet83.8828.0596.2875.6195.0479.2792.1586.5992.5682.93
NasNet87.1926.8394.6371.9595.4569.5192.5687.8091.3291.46

PERFORMANCE ON HOLD-OUT TEST SET

The mean K-Precision and class-wise mean N-Precision values for the hold-out test set are presented in Table 5 and Table 6, respectively. According to Table 5 all DML algorithms produce very good mean K-Precision for different values of K. As the hold-out test data is highly skewed (96:1) towards the control class and so mean K-Precision is not an effective measure as the good mean K-Precision may come from biased feature representation. Hence, we focus on Table 6 for performance comparison. According to Table 6, the mean N-Precision for control is good but the mean N-Precision for the case is not good for all networks. We see the balanced performance is obtained for the Batch-hard sampling approach and for NasNet it produces the best result.
TABLE 5.

Mean-K Precision on hold-out test set. Network-wise best results are bold-faced.

KNetworkContrastiveN-Pair EmbeddingBatch-hard
K=1ResNet-500.92130.89640.8962
MobileNet0.88850.89180.8845
NasNet0.91390.89610.8514
K=3ResNet-500.92740.89610.8962
MobileNet0.88770.89180.8845
NasNet0.91300.89590.8514
K=5ResNet-500.93050.89590.8962
MobileNet0.88900.89180.8845
NasNet0.91180.89580.8514
TABLE 6.

N-Precision for hold-out test set. The best performing DML model is chosen based on the average of Case and Control’s mean N-precision. The network-wise best results are bold-faced.

NetworkContrastiveN-Pair EmbeddingBatch-hard
ControlCaseControlCaseControlCase
ResNet-500.96640.39510.96450.70140.96460.7365
MobileNet0.95490.50610.96300.73650.96060.7598
NasNet0.95750.52270.96430.73650.94950.8300

COMPARISON WITH THE STATE-OF-THE-ART

In [8], for the same data set, only the area under curve (AUC) values of the Receiver operating curve (ROC) were used to evaluate the classification model built with the Faster R-CNN algorithm. In this paper, for comparison purposes, we compute the age group-wise classification accuracies on the hold-out test set from the previous class prediction outcomes. Then the performance of the previous algorithm is compared with the best model produced from this research. According to the discussion presented in Section V-B, NasNet trained with Batch-hard loss minimization is our best feature extractor model and 1-NN is our best classifier. The age group-wise classification performance for our best model and the model built with the AVE algorithm is presented in Table 7. We see that the overall performance of our system outperforms the previously reported result. It is important to mention that we use the entire image and its class label. In contrast, the Faster R-CNN-based AVE algorithm uses additional annotation that localizes the cervix region of interest (ROI) during model training. We conclude that our method improves overall performance. We also compute the age-stratified Kappa statistics between our best model with previously reported Faster RCNN results, which are reported in Table 8.
TABLE 7.

Comparison with state-of-the-art on hold-out test data. This table shows overall and age-stratified (95% CI with exact binomial) comparison of best DML model with Faster RCNN [8]. Reported age stratified analysis excludes nine (9) women as their ages are missing.

Age GroupMethodPredicted ClassesGround Truth
CaseCOL%95% CIControlCOL%95% CI
AllFaster-RCNNCase7183.5%73.9% - 90.7%139217.0%16.2% - 17.9%
Control1416.5%9.3% - 26.1%678283.0%82.1% - 83.8%
BH-NasNet-1-NNCase7183.5%73.9% - 90.7%121314.8%14.1% - 15.6%
Control1416.5%9.3% - 26.1%696185.2%84.4% - 85.9%
<25Faster-RCNNCase1578.9%54.4% - 93.9%21222.9 %20.3% - 25.8%
Control421.1%6.1% - 45.6%71277.1%74.2% - 79.7%
BH-NasNet-1-NNCase1473.7%48.8% - 90.9%22023.8%21.1% - 26.7%
Control526.3%9.1% - 51.2%70476.2%73.3% - 78.9%
25-49Faster-RCNNCase4395.6%84.9% - 99.5%80016.0%15.0% - 17.0%
Control24.4%0.5% - 15.1%421084.0%83.0% - 85.0%
BH-NasNet-1-NNCase4191.1%78.8% - 97.5%69213.8%12.9% - 14.8%
Control48.9%2.5% - 21.2%431886.2%85.2% - 87.1%
>50Faster-RCNNCase1260.0%36.1% - 80.9%38017.0%15.5% - 18.6%
Control840.0%19.1% - 63.9%185283.0%81.4% - 84.5%
BH-NasNet-1-NNCase1575.0%50.9% - 91.3%30113.5%12.1% - 15.0%
Control525.0%8.7% - 49.1%193186.5%85.0% - 87.9%
TABLE 8.

Confusion matrix of the best DML model (BH-NasNet-1-NN) and comparison with state-of-the-art performance on hold-out test data. This table shows overall and age stratified Kappa statistics between best DML model and Faster RCNN [8]. Reported age stratified analysis excludes nine (9) women as their ages are missing.

Age groupFaster RCNNBH-NasNet-1-NNKappa
ControlCase
All agesControl66131830.76
Case3621101
<25Control676400.79
Case33194
25-49Control4120920.78
Case202641
>50Control1809510.70
Case127265

CONCLUSION

This paper takes a pioneering initiative to study the effectiveness of the deep metric learning algorithm for cervical image classification. Our experimental results show that the deep metric learning with Batch-hard loss minimization performs better than the previously proposed AVE method on the hold-out test set. Additionally, the present framework diminishes the image level ROI annotation labor. While our results are indeed better, we note that some misclassification still exists. The probable reason for this is the possible lack of proper generalizability during training. We believe that using more advanced metric learning techniques could overcome this deficit and is left for future work. The real-world application for the proposed system is to serve as an intelligent assistant for the clinician evaluating the woman. Also, the images used in the envisioned system could be acquired using a variety of devices, such as a smartphone, digital camera, or a colposcope enabled with digital image capture capability. These are likely to introduce additional image appearance variability, as noted earlier. Our future work shall also include steps to address this variability in addition to any data imbalance and regional variations in the appearance of the cervix.
  13 in total

Review 1.  Digital tools for collecting data from cervigrams for research and training in colposcopy.

Authors:  Jose Jeronimo; L Rodney Long; Leif Neve; Bopf Michael; Sameer Antani; Mark Schiffman
Journal:  J Low Genit Tract Dis       Date:  2006-01       Impact factor: 1.925

2.  Interobserver agreement in the evaluation of digitized cervical images.

Authors:  Jose Jeronimo; L Stewart Massad; Philip E Castle; Sholom Wacholder; Mark Schiffman
Journal:  Obstet Gynecol       Date:  2007-10       Impact factor: 7.661

3.  An Observational Study of Deep Learning and Automated Evaluation of Cervical Images for Cancer Screening.

Authors:  Liming Hu; David Bell; Sameer Antani; Zhiyun Xue; Kai Yu; Matthew P Horning; Noni Gachuhi; Benjamin Wilson; Mayoore S Jaiswal; Brian Befano; L Rodney Long; Rolando Herrero; Mark H Einstein; Robert D Burk; Maria Demarco; Julia C Gage; Ana Cecilia Rodriguez; Nicolas Wentzensen; Mark Schiffman
Journal:  J Natl Cancer Inst       Date:  2019-09-01       Impact factor: 13.506

4.  Automated and interactive lesion detection and segmentation in uterine cervix images.

Authors:  Amir Alush; Hayit Greenspan; Jacob Goldberger
Journal:  IEEE Trans Med Imaging       Date:  2010-02       Impact factor: 10.048

5.  Automatic CIN Grades Prediction of Sequential Cervigram Image Using LSTM With Multistate CNN Features.

Authors:  Zijie Yue; Shuai Ding; Weidong Zhao; Hao Wang; Jie Ma; Youtao Zhang; Yanchun Zhang
Journal:  IEEE J Biomed Health Inform       Date:  2019-06-13       Impact factor: 5.772

6.  A demonstration of automated visual evaluation of cervical images taken with a smartphone camera.

Authors:  Zhiyun Xue; Akiva P Novetsky; Mark H Einstein; Jenna Z Marcus; Brian Befano; Peng Guo; Maria Demarco; Nicolas Wentzensen; Leonard Rodney Long; Mark Schiffman; Sameer Antani
Journal:  Int J Cancer       Date:  2020-05-19       Impact factor: 7.396

7.  Multi-feature based Benchmark for Cervical Dysplasia Classification Evaluation.

Authors:  Tao Xu; Han Zhang; Cheng Xin; Edward Kim; L Rodney Long; Zhiyun Xue; Sameer Antani; Xiaolei Huang
Journal:  Pattern Recognit       Date:  2016-09-22       Impact factor: 7.740

Review 8.  Performance of visual inspection with acetic acid for cervical cancer screening: a qualitative summary of evidence to date.

Authors:  Lynne Gaffikin; Margo Lauterbach; Paul D Blumenthal
Journal:  Obstet Gynecol Surv       Date:  2003-08       Impact factor: 2.347

9.  Automatic detection of anatomical landmarks in uterine cervix images.

Authors:  Hayit Greenspan; Shiri Gordon; Gali Zimmerman; Shelly Lotenberg; Jose Jeronimo; Sameer Antani; Rodney Long
Journal:  IEEE Trans Med Imaging       Date:  2009-03       Impact factor: 10.048

10.  Cross-Dataset Evaluation of Deep Learning Networks for Uterine Cervix Segmentation.

Authors:  Peng Guo; Zhiyun Xue; L Rodney Long; Sameer Antani
Journal:  Diagnostics (Basel)       Date:  2020-01-14
View more
  1 in total

1.  Oral Cavity Anatomical Site Image Classification and Analysis.

Authors:  Zhiyun Xue; Paul C Pearlman; Kelly Yu; Anabik Pal; Tseng-Cheng Chen; Chun-Hung Hua; Chung Jan Kang; Chih-Yen Chien; Ming-Hsui Tsai; Cheng-Ping Wang; Anil K Chaturvedi; Sameer Antani
Journal:  Proc SPIE Int Soc Opt Eng       Date:  2022-04-04
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.