Literature DB >> 29619277

Deep Learning Nuclei Detection in Digitized Histology Images by Superpixels.

Sudhir Sornapudi1, Ronald Joe Stanley1, William V Stoecker2, Haidar Almubarak1, Rodney Long3, Sameer Antani3, George Thoma3, Rosemary Zuna4, Shelliane R Frazier5.   

Abstract

BACKGROUND: Advances in image analysis and computational techniques have facilitated automatic detection of critical features in histopathology images. Detection of nuclei is critical for squamous epithelium cervical intraepithelial neoplasia (CIN) classification into normal, CIN1, CIN2, and CIN3 grades.
METHODS: In this study, a deep learning (DL)-based nuclei segmentation approach is investigated based on gathering localized information through the generation of superpixels using a simple linear iterative clustering algorithm and training with a convolutional neural network.
RESULTS: The proposed approach was evaluated on a dataset of 133 digitized histology images and achieved an overall nuclei detection (object-based) accuracy of 95.97%, with demonstrated improvement over imaging-based and clustering-based benchmark techniques.
CONCLUSIONS: The proposed DL-based nuclei segmentation Method with superpixel analysis has shown improved segmentation results in comparison to state-of-the-art methods.

Entities:  

Keywords:  Cervical cancer; cervical intraepithelial neoplasia; convolutional neural network; deep learning; image processing; segmentation; superpixels

Year:  2018        PMID: 29619277      PMCID: PMC5869967          DOI: 10.4103/jpi.jpi_74_17

Source DB:  PubMed          Journal:  J Pathol Inform


INTRODUCTION

The reconstruction of medical images into digital form has propelled the fields of medical research and clinical practice.[1] Image processing for histopathology image applications still has numerous challenges to overcome, especially in accurate nuclei detection. Cervical cancer is the fourth most prevalent female cancer globally.[2] Over 500,000 new cases of this cancer are reported annually, especially in Africa; over half of this total eventuates in death.[2] There is a cure for cervical cancer if it is detected early. The gold standard for early cervical cancer diagnosis is the microscopic evaluation of histopathology images by a qualified pathologist.[3456] The severity of cervical cancer increases as the immature atypical cells in the epithelium region increase. Based on this observation, cancer affecting squamous epithelium is classified as normal or one of the three increasingly premalignant grades of cervical intraepithelial neoplasia (CIN): CIN1, CIN2, and CIN3[456] [Figure 1]. Normal means there is no CIN; CIN1 corresponds to mild dysplasia (abnormal change); CIN2 denotes moderate dysplasia; and CIN3 corresponds to severe dysplasia.
Figure 1

Cervical intraepithelial neoplasia grades Left to right: (a) Normal, (b)cervical intraepithelial neoplasia 1, (c) cervical intraepithelial neoplasia 2, (d)cervical intraepithelial neoplasia 3

Cervical intraepithelial neoplasia grades Left to right: (a) Normal, (b)cervical intraepithelial neoplasia 1, (c) cervical intraepithelial neoplasia 2, (d)cervical intraepithelial neoplasia 3 With increasing CIN grade, the epithelium has been observed to show delayed maturation with an increase in immature atypical cells from bottom (basal membrane) to top of the epithelium region.[678910] This can be observed from Figure 1. Atypical immature cells are most dense in the bottom region of the epithelium for CIN1 [Figure 1b]. For CIN2, two-thirds of the bottom region is affected by the atypical immature cells [Figure 1c]. Finally, for CIN3, the atypical immature cells are densely spread over the whole epithelium region [Figure 1d]. At present, cervical tissue is analyzed manually by the pathologists with significant experience with cervical cancer. These pathology specialists are few, and it takes considerable time to scan the tissue slides. This calls for automatic histology image classification, which could alleviate scarce professional resources for image classification, particularly in developing countries where the burden of cervical cancer is greatest. A critical challenge for automatic classification is the accurate identification of nuclei, the small dark structures which undergo changes as the CIN progresses [Figure 1]. Epithelial nuclei provide critical features needed to classify cervical images. Although CIN grade classification can be done by applying deep learning (DL) techniques directly on the image data without the use of nuclei-based features, the accuracy of the classification can be further improved by fusing a feature-based trained neural network models with the DL model. The classification based on the features extracted from the histology images has shown good results in previous studies.[1112] Hence, the detection of nuclei is crucial for correct results. Detection accuracy can be limited by variations in tissue and nuclei staining, image contrast, noisy stain blobs, overlapping nuclei, and variation in nuclei size and shape, with the latter being more prominent with higher CIN grades. In recent years, various algorithms have been proposed to segment nuclei and to extract the nuclei features from digitized medical images. The accuracy of algorithms to identify nuclei may be measured in two ways. The first measure is called nucleus detection or object-based detection. This nucleus-based scoring counts whether a ground-truth nucleus is detected or not. The second method is called nucleus segmentation; nuclei segmentation is pixel-based scoring which counts accuracy pixel by pixel. Recent reviews by Xing and Yang[13] and Irshad et al.[14] summarized techniques in this fast-evolving field for both nuclei detection and segmentation. The review by Irshad et al. provided additional material on nuclear features; the review by Xing and Yang included additional recent studies; both reviews gave detailed descriptions of methods and results for the nuclei detection for many types of histopathology images including brain, breast, cervix, prostate, muscle, skin, and leukocyte.[713] In the following, we summarize selected recent methods to find nuclei in histopathology images in general, followed by specific methods to find nuclei in cervical images. For the general domain of histopathology images, recent studies have employed conventional techniques, various DL techniques, and techniques combining both methods. A graph-cut technique was followed by multiscale Laplacian-of-Gaussian (LoG) filtering, adaptive scale selection, and a second graph-cut operation.[15] Generalized LoG filters were used to detect elliptical blob centers; watershed segmentation was used to split touching nuclei.[16] The generalized LoG filter technique was modified using directional LoG filters followed by adaptive thresholding and mean-shift clustering.[17] A convolutional neural network (CNN) nuclear detection model called “deep voting” used voting based on location of patches and weights based on confidence in the patches to produce final nuclei locations.[18] Stacked sparse autoencoder (SSAE) DL was used for nuclei detection and compared to other DL techniques using CNN variations.[19] SSAE sensitivity was similar to that obtained for the optimal CNN; specificity compared favorably to CNN.[19] Another voting approach to overcome variable nuclear staining exploited nuclear symmetry.[20] An additional voting approach used adaptive thresholding for seed finding followed by elliptical modeling and watershed technique.[21] Canny edge detection was followed by multi-pass directional voting; results surpassed those of the SSAE. A CNN was combined with region merging and a sparse shape and local repulsive deformable model[22] with good results. In the domain of cervical cytology and histopathology, automated localization of the cervical nuclei used the converging squares algorithm.[23] The Hough transform was implemented to detect the nuclei based on shape features.[24] Cervical cells were classified using co-occurrence matrix textural feature extraction and morphological transforms.[25] Analysis of cell nuclei segmentation was performed through Bayesian interpretation after segmentation by a Viterbi search-based active contour method. Segmentation was also accomplished by a region grid algorithm through contour detection around the nuclei boundary.[26] Nuclei were segmented using level-set active contour methods.[2728] Intensity and color information were used for nuclei enhancement and segmentation.[29] A DL framework was used for segmentation of cytoplasm and nuclei.[30] K-means clustering was used for nuclei feature extraction followed by classification based on fusion.[12] A multi-scale CNN followed by graph partitioning was used for nuclei detection in cervical cytology images.[31] Transfer learning to recognize cervical cytology nuclei using the CaffeNet architecture was trained first on ImageNet then, using the trained network, retrained on cervical slide images, containing one cell per slide.[32] Semantic pixel-wise labeling[33] for the detection of nuclei increases computationally expensive since every pixel is individually labeled through a series of encoder and decoder stacks. U-Net[34] utilizes up-sampling approach with deconvolution layers with 23 convolutional layers, which makes the network use more memory and more computations. The nuclei segmentation research here employs DL to extract nuclei patches, a simple linear iterative cluster (SLIC) model and a CNN to classify the obtained superpixel data. A group of similar pixels (superpixels) are classified, requiring reduced memory compared to the pixel-wise approach, also reducing the number of parameters to be tuned. Scoring in the current study, object-based detection, is based upon whether nuclei are correctly detected or not.

METHODS

Biologically inspired CNNs operate upon a digital image, convolving image arrays with the image, producing feature vectors serving as parameters to the CNN. The automatically determined feature vectors serve as weights; these are modified with each iteration as the network learns by training. The primary goal of this paper is to segment the nuclei in the epithelium region of cervical cancer histology images by considering local features instead of features from the whole image. This local information is used to classify whether the segment contains nuclei or background. The CNNs use image vectors as inputs and learn different feature vectors, which ultimately solve the classification problem. The proposed methodology is depicted in Figure 2.
Figure 2

Proposed methodology

Proposed methodology To make use of localized information, small image patches are obtained from the original image using a superpixel extraction method. Superpixel algorithms are devised to group pixels with similar properties into regions to form clusters. Optimal superpixels avoid oversegmentation without information gain, which is present at the pixel level, and undersegmentation with information loss, if superpixels are too large. The SLIC algorithm is chosen as it generates superpixels based on color (intensity) and distance proximities with respect to each pixel.

Preprocessing

Before extracting superpixels, the original image is preprocessed using a Gaussian smoothing filter to smooth the input image to reduce Gaussian and impulse valued noises, which are mainly generated during image capture from the slides and digitization process.[35] The results of oversegmented images through superpixel generation also indicate the importance of smoothing the images. The filter's impulse response is the Gaussian function, which decays rapidly, so as to select narrow windows to avoid the loss of image information. This function divides the image into its respective windows and applies the cost function. The two-dimensional Gaussian function is applied on the input image using a built-in MATLAB® function. The standard deviation can be user-defined; here, we use the default value of two. The Gaussian filter is applied instead of a trimmed mean filter because the Gaussian filter processes our images 3184.16 times faster than the trimmed mean filter. When the outputs of the algorithms were compared, the output using the Gaussian filter gave a better superpixel result as compared to the output obtained using the trimmed mean filter. The darker nuclei are in general surrounded with red-stained cytoplasm inside a cell and the background region is not stained. Hence, the red, green, blue (RGB) color space of the image is converted to CIE LAB color space[3136] to improve the contrast between nuclei, cytoplasm, and background. The contrast is further enhanced using a linear transformation, increasing the scale of pixel intensity from [rmin, rmax] to (0, 255). A morphological closing operation is applied on the luminance plane (L component) of the resultant CIE LAB color image to remove any small holes and to smooth boundaries. The L component represents the perceived brightness, which further increases image contrast. These operations produce the initial binary nuclei mask to aid in extracting superpixels from the image. The generated binary mask reduces computational overload and reduces challenges due to noise and other variations in cervical histopathology images, such as variable staining present in cervical tissue, to provide a binary mask overlay to guide the next step in superpixel generation.

Superpixel extraction

Superpixels are generated automatically for the test images. An SLIC algorithm[37] is used to extract superpixels rather than other state-of-the-art methods[38394041] because it is faster, is more memory efficient, has better adherence to boundaries, and improves segmentation performance. Furthermore, it considers both color and distance properties which are appropriate with color orientation of the nuclei around a small region. A labeled matrix, with size equal to that of the original image, is obtained as an output from the SLIC function. A manually generated epithelium mask, which is verified by an expert pathologist (RZ), is then applied on the labeled matrix to remove the unwanted region. The resultant matrix is again relabeled. The minimum size for superpixels, 200 pixels, is chosen to be larger than the largest nucleus and smaller than the patch size (256 pixels). The patch width and height (16 pixels) are chosen to contain all superpixels and all nuclei, as shown in Figure 3, so that the whole superpixel region is covered while creating a 16 × 16 × 3 RGB patch image dataset for training the CNN.
Figure 3

A portion of original image with superpixels. Nuclei do not exceed 16 pixels in height or width or 200 pixels in area

A portion of original image with superpixels. Nuclei do not exceed 16 pixels in height or width or 200 pixels in area The centroid of each superpixel is computed. With respect to that centroid, a 16 × 16 × 3 image patch is formed as shown in Figure 4. A patch is said to be a part of the nuclei region if nuclei comprise at least 10% of its area. The nuclei region is given the highest priority as compared to the cytoplasm and background. The problem of generating 16 × 16 × 3 patches from superpixels at the edges of the image is solved by mirroring the image.
Figure 4

Generation of 16 × 16 × 3 red, green, blue image from superpixel

Generation of 16 × 16 × 3 red, green, blue image from superpixel Finally, 16 × 16 × 3 RGB input images are obtained from the superpixels of the original image. As DL benefits from more examples, data augmentation is performed by rotating the original image by 180° and extracting 16 × 16 patches.

Data generation

Data generation is done carefully to prepare both training and test image datasets. For our experiment, a total of 12 images, six images each from the 71-image dataset and 62-image dataset, are used for training the network. The remaining 121 images are used in the testing phase. Thus, the training and test sets used for generating results reported in this study are disjoint. Nuclei segmentation has been investigated in the previous studies using the 71-image[22] and 62-image[25] datasets, providing benchmarks for this study. Training images are carefully chosen so that the network understands how to handle different kinds of images. Observation of images from the datasets discloses three types of images: images with light nuclei and light cytoplasm, images with darker nuclei and moderate cytoplasm, and images with darker nuclei and thicker cytoplasm as shown in Figure 5. To balance the training set for the CNN, six images from each dataset, two images for each of the three image types are included, totaling 12 images.
Figure 5

Images with lighter nuclei (left), darker nuclei with lighter cytoplasm (center), darker nuclei with thicker cytoplasm (right)

Images with lighter nuclei (left), darker nuclei with lighter cytoplasm (center), darker nuclei with thicker cytoplasm (right) Classifying whether nuclei are present or not in the 16 × 16 × 3 patch is a binary classification problem. The patch target label is obtained from the binary nuclei masks that are already available in the database. Some of the portions of the nuclei masks are modified so that the target labels represent exact ground truth values. The extracted 16 × 16 × 3 patches are as shown in Figure 6. The label “0” denotes nuclei and the label “1” denotes background. A total of 377,012 patches are obtained using preprocessing steps as shown in Figure 7 (left) for 12 original images comprising both nuclei and background.
Figure 6

Sample 16 × 16 × 3 red, green, blue images and their 16 × 16 V-plane images

Figure 7

Generation of training dataset (left) and test dataset (right)

Sample 16 × 16 × 3 red, green, blue images and their 16 × 16 V-plane images Generation of training dataset (left) and test dataset (right) The test data are generated by preprocessing the image [Figure 7, right]. The luminance plane is used to generate superpixels, and then, 16 × 16 × 3 image patches are formed for each individual original image.

Convolutional neural network

As a prestep to train the CNN, all small image patches are converted to the HSV color plane and then the value plane (V-plane) is extracted. Before selecting the V-plane, various color planes are observed manually and are also used to train the network. The V-plane and the L-plane (luminance plane) gave promising results. The V-plane is considered for this experiment as shown in Figure 6. The V component indicates the quantity of light reflected and is useful for extraction from the patches because the nuclei are typically blue-black and reflect only a small amount of light. To classify the presence of nuclei, the CNN is trained with the features that were generated by convolutional layers using raw pixel input data. The first stage was a shallow CNN with one convolutional layer and a following max pool layer. A total of 36,478 image patches (extracted from two images) were processed for a quick quality check. To classify the presence of nuclei, the CNN is trained with the features that were generated by convolutional layers using raw pixel input data. A remarkable improvement in the validation accuracy was observed when a deep CNN architecture [modified LeNet-5[42] model with varied layers and hyper-parameters as shown in Figure 8] was considered with multiple convolutional, max pooling, and dropout layers at the beginning of the network and three dense neural networks (convolution and dense layers with a nonlinear ReLU activation function[43]) at the end of the network. The two neurons in the output layer are activated with a SoftMax function.
Figure 8

Convolutional neural network architecture

Convolutional neural network architecture This produced 98.1% validation accuracy on two input images. Later, 10 more images were included to make the network learn to classify nuclei in different environments as shown in Figure 5. Upon training with 377,012 patches of 16 × 16 size (extracted from 12 full-size images), a validation accuracy of 95.70% is achieved. The obtained dataset of inputs and target labels is used to train CNNs with different architectures, and the following architecture [Figure 8] gave best results with higher validation accuracy on test images that were part of the training data. The training dataset is used to fit the CNN model. A validation dataset, consisting of 20% of the training dataset, is helpful to estimate the prediction error for best model selection. Categorical accuracy (Lμ) is computed between targets (ti, c) and prediction (pi, c) produced from the validation dataset. The weights are initialized randomly using Glorot weight initialization.[44] An adaptable learning rat ∈ (0.0001, 0.03) and momentum with range ∈ (0.9, 0.999) are applied to the network while training for 2000 epochs. The architecture produced a validation accuracy of 95.70% at the end of the 2000th epoch. The network is trained for 2000 epochs since further training appears not to decrease validation loss [Figure 9].
Figure 9

Training loss, validation accuracy, and validation loss versus epochs

Training loss, validation accuracy, and validation loss versus epochs The error on the training set is denoted as training loss. Validation loss is the error as a result of running the validation set through the previously trained CNN. Figure 9 represents a drop in training and validation error as the number of epochs increase. This is a clear indication that the network is learning from the data that are given as an input to the network. Figure 10 (left) shows all 37 × 3 × 3 first-layer convolutional feature vectors obtained from the trained network. The initial layer of the convolutional network mainly focuses learning on the edge and curve features of the input image. Figure 10 (right) represents the result of the convolution of the feature vectors with the 16 × 16 image producing a 32 × 14 × 14 image.
Figure 10

32 × 3 × 3 convolutional neural network filters and 32 × 14 × 14 convolved output in first layer

32 × 3 × 3 convolutional neural network filters and 32 × 14 × 14 convolved output in first layer The trained network model is saved along with the weights and filter coefficients. This saved model is loaded back to test on the remaining images of the 71-image and 62-image datasets (121 images) by classifying individual patches generated from each image to assess nuclei detection accuracy. The location of every superpixel extracted from the original image is saved as a labeled image. The results of classification are mapped with the labeled image to finally obtain a binary nuclei mask from the corresponding original image. The nuclei detection rate on the test images is then calculated by manually counting all 108,635 original ground truth nuclei truly detected and those falsely detected by the algorithm.

Experimental results and analysis

Experimental results

The proposed algorithm is applied on both 71-image and 62-image datasets, using six images from each of the datasets for training the CNN. The remaining images are used for testing the trained model. The training set and test set are disjoint. Figure 11 depicts the nuclei mask generated with nuclei mask boundaries marked in green.
Figure 11

Nuclei masks (green) superimposed on the original image

Nuclei masks (green) superimposed on the original image The DL algorithm applied to both the 71-image dataset and the 62-image dataset shows overall segmentation accuracy of 97.11% and 93.33%, respectively. Finally, the overall segmentation accuracy of the combined set is 95.97%. The accuracy of nuclei detection is calculated on a per-nuclei basis by manually recording the true positive (TP) (i.e., the number of nuclei successfully detected), false negative (FN) (i.e., the number of nuclei not detected), and false positive (FP) (i.e., number of nonnuclei objects found). Using FP and FN totals, accuracy measures are calculated,[25] including precision, recall, accuracy (μ), dice similarity coefficient (DSC), harmonic mean of precision and recall (F1), and Jacquard index, Equations 1–6. Table 1 shows these accuracy measures for the 62, 71, and combined datasets.
Table 1

Nuclei detection results using the deep learning superpixel approach1

Nuclei detection results using the deep learning superpixel approach1 It is observed that if smaller size superpixels are considered, that is, if finer localization is done, the final nuclei masks are better. In addition, a deeper CNN shows improved classification results when compared to a shallow CNN.

Analysis of results

Here, the results obtained above are compared with the results from benchmark algorithms. The following images represent the FP and FN cases. Figure 12 (left) represents a FP condition where false nuclei detection is observed. The circled portion shows the region where there is no nucleus present in the original image but detected as nucleus present with a green contour around the FP object boundary. Figure 12 (right) shows a nucleus misclassified as background. The undetected nucleus is marked in the original image, but there is no contour around the marked nucleus. Both FP and FN cases lower overall object-based detection accuracy.
Figure 12

Examples of false positive (left) and false negative (right) results. Note variable staining

Examples of false positive (left) and false negative (right) results. Note variable staining Equations 1-6. Nuclei detection accuracy given TP and TN. The presence of red stains on the image samples always poses challenges in nuclei detection as the stains are falsely detected as nuclei by various algorithms, yet some nuclei may lie under red stains. The proposed algorithm has overcome this challenge by detecting the nuclei even under the red stains [Figure 13]. The training process of the CNN allows learning about this feature from the ground truth images.
Figure 13

Nuclei detected even under red stains

Nuclei detected even under red stains

Comparison of results

This article presents a DL-based nuclei segmentation approach, using superpixel extraction followed by a CNN classifier. The algorithm has achieved an overall accuracy μ of 96.0% on the combined set, with 97.11% accuracy achieved on the 71-image dataset [Table 1], outperforming previous cervical histopathology nuclei detection approaches. Previously, segmentation based on K-means clustering followed by mathematical morphology operations[12] produced an overall recall estimated at 89.5% on the 62-set of images. The level set method and fuzzy C-means clustering[28] approach on the 71-image dataset achieved 96.47% accuracy in comparison to the current 97.11% accuracy. Some recent results in cervical cytology nuclei detection have produced very high nuclei detection (object-based results) [Table 2].[3045] Nuclei detection in cervical cytology images is not comparable to nuclear detection in histopathology images. As Irshad et al. noted, nuclei segmentation “is particularly difficult on pathology images.”[14] Cervical cytology images have “well-separated nuclei and the absence of complicated tissue structures,” while most nuclei in histopathology images are “often part of structures presenting complex and irregular visual aspects.”[14] In addition, we have found that cytology images have a greater contrast and fewer nuclear mimics.
Table 2

Cervical nuclear detection versus current deep learning superpixel approach1

Cervical nuclear detection versus current deep learning superpixel approach1 Table 2 compares the current DL superpixel nuclei results with previous cervical nuclei detection studies, with results for all studies using object-based scoring. The current method outperforms the previous cervical histopathology study. Table 3 compares the current study with recent histopathology nuclei detection studies reported for various tissues, using object-based scoring. This object identification accuracy, in comparison to pixel-based nuclear outline accuracy, may be the better of the two measures, because once a nucleus is known with high assurance, then outlines, texture, and other characteristics can be scored. The current method for nuclei object detection outperforms all previous approaches.
Table 3

General nuclei detection results versus current deep learning superpixel approach1

General nuclei detection results versus current deep learning superpixel approach1 There has been a noticeable trend recently in the number of studies using DL for nuclei detection. DL is a powerful technique for nuclei detection; with sufficient numbers of nuclei, DL yields superior performance.[13] Yet, the general enthusiasm about DL techniques should be tempered with the reality that datasets often have insufficient samples to allow learning of nuclei characteristics that vary significantly; besides nuclei size, shape, and internal features, nuclear staining varies widely.[20] Since pathologist time is a scarce resource, the number of pathologist-marked nuclei in databases remains over two orders of magnitude lower than the numbers of test nuclei in large test sets; nuclei detection results are often estimated from samples of marked nuclei.[12] In some recent studies, detection accuracy for conventional techniques, which included incorporation of higher level knowledge, e.g., nuclear edge symmetry, surpassed DL results [Table 3].[912142046] Other studies in histopathology have surpassed DL results by combining conventional techniques with DL techniques. Zhong et al. fused information from supervised and DL approaches. In comparing multiple machine-learning strategies, it was found that the combination of supervised cellular morphology features and predictive sparse decomposition DL features provided the best separation of benign and malignant histology sections.[47] Wang et al. were able to detect mitosis in breast cancer histopathology images using the combined manually tuned cellular morphology data and convolutional neural net features.[48] Arevalo et al. added an interpretable layer they called “digital staining,” to improve their DL approach to classification of basal cell carcinoma.[49] Of interest, the handcrafted layer finds the area of interest, reproducing the high-level search strategy of the expert pathologist. Additional higher level knowledge has been used to separate nuclei which touch or overlap in multiple studies. However, the higher-level knowledge which pathology specialists use most extensively is the overall architecture present in the arrangement of cells and nuclei in the histopathology image. Thus, certain patterns, such as the gradient of nuclear atypia from basal layer to surface layer in carcinoma in situ, changes as the CIN grade increases, and different patterns of a certain type of cancer, can all provide critical diagnostic information. There is an interaction between these higher-level patterns and nuclei detection; not all nuclei are of equal importance in contributing to the diagnosis. Future studies could incorporate higher-level architectural patterns in the detection of critical cellular components such as nuclei. Thus, higher-level architectural knowledge such as nuclear distribution obtained by conventional image processing techniques fused with DL techniques will be used to advantage in automated diagnosis in the future. Since much higher-level histopathology knowledge is domain-specific, the longstanding goal of applying a single method to multiple histopathology domains remains elusive.

CONCLUSION

The proposed method of DL-based nuclei segmentation with superpixel analysis has shown improved segmentation results in comparison to state-of-the-art methods. The proposed method, oversegmenting the original image by generating superpixels, allows the CNN to learn the localized features better in the training phase. The trained model is finally applied on a larger dataset. Future work includes the application of other CNN architectures as well as fusion with higher-level knowledge with the CNN classifier. Features obtained from the detected nuclei will be used in automatic CIN classification.

Financial support and sponsorship

This research was supported (in part) by the Intramural Research Program of the National Institutes of Health, National Library of Medicine, and Lister Hill National Center for Biomedical Communications.

Conflicts of interest

There are no conflicts of interest.
  24 in total

1.  Mitosis detection in breast cancer pathology images by combining handcrafted and convolutional neural network features.

Authors:  Haibo Wang; Angel Cruz-Roa; Ajay Basavanhally; Hannah Gilmore; Natalie Shih; Mike Feldman; John Tomaszewski; Fabio Gonzalez; Anant Madabhushi
Journal:  J Med Imaging (Bellingham)       Date:  2014-10-10

2.  Improved automatic detection and segmentation of cell nuclei in histopathology images.

Authors:  Yousef Al-Kofahi; Wiem Lassoued; William Lee; Badrinath Roysam
Journal:  IEEE Trans Biomed Eng       Date:  2009-10-30       Impact factor: 4.538

3.  An unsupervised feature learning framework for basal cell carcinoma image analysis.

Authors:  John Arevalo; Angel Cruz-Roa; Viviana Arias; Eduardo Romero; Fabio A González
Journal:  Artif Intell Med       Date:  2015-04-23       Impact factor: 5.326

4.  Automatic Nuclei Detection Based on Generalized Laplacian of Gaussian Filters.

Authors:  Richard Berendt; Naresh Jha; Mrinal Mandal
Journal:  IEEE J Biomed Health Inform       Date:  2016-03-21       Impact factor: 5.772

5.  Segmentation of cytoplasm and nuclei of abnormal cells in cervical cytology using global and local graph cuts.

Authors:  Ling Zhang; Hui Kong; Chien Ting Chin; Shaoxiong Liu; Zhi Chen; Tianfu Wang; Siping Chen
Journal:  Comput Med Imaging Graph       Date:  2014-02-21       Impact factor: 4.790

6.  Histology image analysis for carcinoma detection and grading.

Authors:  Lei He; L Rodney Long; Sameer Antani; George R Thoma
Journal:  Comput Methods Programs Biomed       Date:  2012-03-20       Impact factor: 5.428

7.  When machine vision meets histology: A comparative evaluation of model architecture for classification of histology sections.

Authors:  Cheng Zhong; Ju Han; Alexander Borowsky; Bahram Parvin; Yunfu Wang; Hang Chang
Journal:  Med Image Anal       Date:  2016-09-09       Impact factor: 8.545

Review 8.  Robust Nucleus/Cell Detection and Segmentation in Digital Pathology and Microscopy Images: A Comprehensive Review.

Authors:  Fuyong Xing; Lin Yang
Journal:  IEEE Rev Biomed Eng       Date:  2016-01-06

9.  Nuclei-Based Features for Uterine Cervical Cancer Histology Image Analysis With Fusion-Based Classification.

Authors:  Peng Guo; Koyel Banerjee; R Joe Stanley; Rodney Long; Sameer Antani; George Thoma; Rosemary Zuna; Shelliane R Frazier; Randy H Moss; William V Stoecker
Journal:  IEEE J Biomed Health Inform       Date:  2015-10-26       Impact factor: 5.772

10.  Cancer incidence and mortality worldwide: sources, methods and major patterns in GLOBOCAN 2012.

Authors:  Jacques Ferlay; Isabelle Soerjomataram; Rajesh Dikshit; Sultan Eser; Colin Mathers; Marise Rebelo; Donald Maxwell Parkin; David Forman; Freddie Bray
Journal:  Int J Cancer       Date:  2014-10-09       Impact factor: 7.396

View more
  11 in total

1.  Fast whole-slide cartography in colon cancer histology using superpixels and CNN classification.

Authors:  Frauke Wilm; Michaela Benz; Volker Bruns; Serop Baghdadlian; Jakob Dexl; David Hartmann; Petr Kuritcyn; Martin Weidenfeller; Thomas Wittenberg; Susanne Merkel; Arndt Hartmann; Markus Eckstein; Carol Immanuel Geppert
Journal:  J Med Imaging (Bellingham)       Date:  2022-03-14

2.  Automated Cervical Digitized Histology Whole-Slide Image Analysis Toolbox.

Authors:  Sudhir Sornapudi; Ravitej Addanki; R Joe Stanley; William V Stoecker; Rodney Long; Rosemary Zuna; Shellaine R Frazier; Sameer Antani
Journal:  J Pathol Inform       Date:  2021-06-09

Review 3.  Artificial intelligence in digital pathology - new tools for diagnosis and precision oncology.

Authors:  Kaustav Bera; Kurt A Schalper; David L Rimm; Vamsidhar Velcheti; Anant Madabhushi
Journal:  Nat Rev Clin Oncol       Date:  2019-08-09       Impact factor: 66.675

Review 4.  A Review of Computational Methods for Cervical Cells Segmentation and Abnormality Classification.

Authors:  Teresa Conceição; Cristiana Braga; Luís Rosado; Maria João M Vasconcelos
Journal:  Int J Mol Sci       Date:  2019-10-15       Impact factor: 5.923

5.  DeepCIN: Attention-Based Cervical histology Image Classification with Sequential Feature Modeling for Pathologist-Level Accuracy.

Authors:  Sudhir Sornapudi; R Joe Stanley; William V Stoecker; Rodney Long; Zhiyun Xue; Rosemary Zuna; Shellaine R Frazier; Sameer Antani
Journal:  J Pathol Inform       Date:  2020-12-24

6.  Automated Diagnosis of Cervical Intraepithelial Neoplasia in Histology Images via Deep Learning.

Authors:  Bum-Joo Cho; Jeong-Won Kim; Jungkap Park; Gui-Young Kwon; Mineui Hong; Si-Hyong Jang; Heejin Bang; Gilhyang Kim; Sung-Taek Park
Journal:  Diagnostics (Basel)       Date:  2022-02-21

7.  Hybrid AI-assistive diagnostic model permits rapid TBS classification of cervical liquid-based thin-layer cell smears.

Authors:  Xiaohui Zhu; Xiaoming Li; Kokhaur Ong; Wenli Zhang; Wencai Li; Longjie Li; David Young; Yongjian Su; Bin Shang; Linggan Peng; Wei Xiong; Yunke Liu; Wenting Liao; Jingjing Xu; Feifei Wang; Qing Liao; Shengnan Li; Minmin Liao; Yu Li; Linshang Rao; Jinquan Lin; Jianyuan Shi; Zejun You; Wenlong Zhong; Xinrong Liang; Hao Han; Yan Zhang; Na Tang; Aixia Hu; Hongyi Gao; Zhiqiang Cheng; Li Liang; Weimiao Yu; Yanqing Ding
Journal:  Nat Commun       Date:  2021-06-10       Impact factor: 14.919

8.  Deep Convolution Neural Network for Malignancy Detection and Classification in Microscopic Uterine Cervix Cell Images.

Authors:  Shanthi P B; Faraz Faruqi; Hareesha K S; Ranjini Kudva
Journal:  Asian Pac J Cancer Prev       Date:  2019-11-01

9.  Tailored for Real-World: A Whole Slide Image Classification System Validated on Uncurated Multi-Site Data Emulating the Prospective Pathology Workload.

Authors:  Julianna D Ianni; Rajath E Soans; Sivaramakrishnan Sankarapandian; Ramachandra Vikas Chamarthi; Devi Ayyagari; Thomas G Olsen; Michael J Bonham; Coleman C Stavish; Kiran Motaparthi; Clay J Cockerell; Theresa A Feeser; Jason B Lee
Journal:  Sci Rep       Date:  2020-02-21       Impact factor: 4.379

10.  Evaluating Very Deep Convolutional Neural Networks for Nucleus Segmentation from Brightfield Cell Microscopy Images.

Authors:  Mohammed A S Ali; Oleg Misko; Sten-Oliver Salumaa; Mikhail Papkov; Kaupo Palo; Dmytro Fishman; Leopold Parts
Journal:  SLAS Discov       Date:  2021-06-24       Impact factor: 3.341

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.