Literature DB >> 32603329

Deep learning approach to describe and classify fungi microscopic images.

Bartosz Zieliński1,2, Agnieszka Sroka-Oleksiak3,4, Dawid Rymarczyk1,2, Adam Piekarczyk1, Monika Brzychczy-Włoch4.   

Abstract

Preliminary diagnosis of fungal infections can rely on microscopic examination. However, in many cases, it does not allow unambiguous identification of the species due to their visual similarity. Therefore, it is usually necessary to use additional biochemical tests. That involves additional costs and extends the identification process up to 10 days. Such a delay in the implementation of targeted therapy may be grave in consequence as the mortality rate for immunosuppressed patients is high. In this paper, we apply a machine learning approach based on deep neural networks and bag-of-words to classify microscopic images of various fungi species. Our approach makes the last stage of biochemical identification redundant, shortening the identification process by 2-3 days, and reducing the cost of the diagnosis.

Entities:  

Mesh:

Year:  2020        PMID: 32603329      PMCID: PMC7326179          DOI: 10.1371/journal.pone.0234806

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

Yeast and yeast-like fungi are a component of natural human microbiota [1]. However, as opportunistic pathogens, they can cause surface and systemic infections [2]. The leading causes of the fungal infections are impaired function of the immune system and imbalanced microbiota composition in the human body. Other factors of fungal infections include steroid treatment, invasive medical procedures, and long-term antibiotic treatment with a broad spectrum of antimicrobial agents [3-5]. The standard procedure in mycological diagnostics begins with collecting various types of test materials like swabs, scraps of skin lesions, urine, blood, or cerebrospinal fluid. Next, the clinical materials (marked as B in Fig 1) are directly cultured on special media, while the blood and cerebrospinal fluid samples (marked as A in Fig 1) require prior cultivation in automated closed systems for additional 2-3 days. Material incubates under specific temperature conditions (usually for 2-4 days in case of yeast-like fungi). The initial identification of fungi bases on the assessment of the cells’ shapes observed under the microscope as well as the growth rate, type, shape, color, and the smell of the colonies. Such analysis allows the assignment to fungi type; however, identification of the species is usually impossible due to the significant similarity between them. Because of that, further analysis consisting of biochemical tests, is necessary. As a result, the entire diagnostic process from the moment of culture to species identification can last 4-10 days (see Fig 1).
Fig 1

Standard and computer-aided mycological diagnosis.

Standard mycological diagnostics (I) require analysis with biochemical tests. As a result, the entire diagnostic process can last 4-10 days. In our computer-aided approach (II), biochemical tests are replaced with a machine learning approach that predicts fungi species based only on microscopic images. It shortens the diagnosis by 2-3 days.

Standard and computer-aided mycological diagnosis.

Standard mycological diagnostics (I) require analysis with biochemical tests. As a result, the entire diagnostic process can last 4-10 days. In our computer-aided approach (II), biochemical tests are replaced with a machine learning approach that predicts fungi species based only on microscopic images. It shortens the diagnosis by 2-3 days. In this paper, we apply a machine learning approach based on deep neural networks and bag-of-words approaches to classify microscopic images of various fungus species. As a result, the last stage of biochemical identification is unnecessary, which shortens the identification process by 2-3 days and reduces the cost of diagnosis. It allows accelerating the decision about the introduction of an appropriate antifungal drug, which prevents the progression of the disease and shortens the time of patient recovery. According to our best knowledge, there are no other methods for classifying fungi species based only on microscopic images. Existing methods involve techniques such as morphological identification of a type of fungi [6], fluorescence in situ hybridization (FISH) [7], biochemical techniques, molecular approaches, such as PCR [8], and sequencing [9]. However, all of them are costly. On the other hand, our method bases on basic microbiological staining (Gram staining) and a simple microscope equipped with a camera, and takes only a few minutes, which makes it easily applicable in many laboratories. The paper is structured as follows. First, we introduce a fungus database and describe a classification method based on deep neural networks and bag-of-words methods. Then, we present experimental setup, results, and conclusion.

Materials and methods

Materials

One of the most common fungal infections is candidiasis [5], mainly caused by Candida albicans (50-70% of cases) [10]. Other species responsible for the diseases are Candida glabrata [2, 3], Candida tropicalis [4], Candida krusei [11], and Candida parapsilosis [3, 4]. In high-risk patients, severe infections can also be caused by Cryptococcus neoformans [12] and Saccharomyces phylum [13]. Taking those facts into consideration, we prepared database, which consists of five yeast-like fungal strains: Candida albicans ATCC 10231 (CA), Candida glabrata ATCC 15545 (CG), Candida tropicalis ATCC 1369 (CT), Candida parapsilosis ATCC 34136 (CP), and Candida lustianiae ATCC 42720 (CL); two yeast strains: Saccharomyces cerevisae ATCC 4098 (SC) and Saccharomyces boulardii ATCC 74012 (SB); and two strains belonging to the Basidiomycetes: Maalasezia furfur ATCC 14521 (MF) and Cryptococcus neoformans ATCC 204092 (CN). All strains are from the American Type Culture Collection. The species in our database highly overlap with the most common fungal infections; however, they are not identical due to the limitations of our repository. The strains were cultured on Sabouraud agar at 37°C for 48h (together with olive oil in the case of Maalaseizia furfur). After this time, microscopic preparations were made (2 preparations for each fungal strain) and stained with Gram method. Images were taken using an Olympus BX43 microscope with 100 times a super-apochromatic objective under oil-immersion. The photographic documentation was then produced with an Olympus BP74 camera and CellSense software (Olympus). Altogether, our Digital Images of Fungus Species database (DIFaS) contains 180 images (9 strains × 2 preparations × 10 images) of resolution 3600 × 5760 × 3 with 16-bits intensity range in every pixel. In Fig 2, we present three random thumbnails for each of the registered strains.
Fig 2

Sample images from DIFaS database.

Three random images for each of the strains from DIFaS database.

Sample images from DIFaS database.

Three random images for each of the strains from DIFaS database.

Method

Deep Neural Networks (DNN) have shown human-level performance in case of large amounts of training data; however, they are limited when it comes to the application on small datasets due to the large numbers of parameters. Therefore, in this work, we consider two types of domain adaptation, both based on DNN features initially pre-trained on a different task (i.e., instance classification [14]). As a baseline method, we fine-tune the classifier’s block of the well-known network architectures, i.e., AlexNet [15], DenseNet169 [16], InceptionV3 [17], and ResNet [18] (with frozen features’ block). As we present in results, such architectures are not optimal due to the small training set (see S1 and S3 Tables). Hence, we propose to apply the deep bag-of-words multi-step algorithm shown in Fig 3. In contrast to baseline methods, which utilize “shallow” Neural Network to previously calculated features, our strategies aggregate those features using one of the bag-of-words approaches and then classify them with Support Vector Machine (SVM). Such a policy, previously applied to texture recognition [19] and bacteria colony classification [20], is more accurate than the baseline methods; however, it is not well known. Therefore, to make this paper self-contained, below, we describe its successive steps.
Fig 3

Deep bag-of-words algorithm for mycological diagnosis.

The multi-step algorithm produces robust image features using previously trained deep neural network, aggregates them using one of the bag-of-words approaches, and classifies them with Support Vector Machine.

Deep bag-of-words algorithm for mycological diagnosis.

The multi-step algorithm produces robust image features using previously trained deep neural network, aggregates them using one of the bag-of-words approaches, and classifies them with Support Vector Machine. To generate robust image representation, AlexNet [15], InceptionV3 [17] or ResNet [18] pre-trained on ImageNet [14] database are used. Another option would be to use conventional handcrafted descriptors (like ORB [21] or DSIFT [22]); however, they are usually outperformed by deep features. Considered network architectures consist of two parts: convolutional layers, which are responsible for extracting image features (so-called features’ block), and fully connected layers, which are responsible for the classification (so-called classifier’s block). Classifier’s block cannot be directly used because it was trained for other types of images; however, features’ block encodes more general, reusable information. Therefore, removing the classifier block from the network and preserving convolutional layers allows us to generate robust image features. In the case of AlexNet, we obtain a set of points in 256-dimensional space, whose number depends on the input image’s resolution (e.g., in case of resolution 500 × 500 pixels, 169 points (13 ⋅ 13) are generated). Since the classified patches are always of the same size, their features’ blocks could be used directly by the classifier. It, however, would lead to vast data dimensionality (i.e., the feature vector of size 43264), which according to our experiments, results in the lack of generalization, primarily due to the relatively small size of the training set (100 images). Therefore, to obtain a more reliable representation of patches, we pool the acquired set of points using Bag of Words, BoW [23, 24], or its more expressive modification called Fisher Vector, FV [25]. The idea behind both of them is to aggregate a set of points (representing the patch) with a so-called codebook. The codebook is usually generated from the subset of training data in an unsupervised manner using a clustering algorithm (e.g., k-Means or Expectation Maximization [26]). Given a codebook, the set of 256-dimensional points obtained with AlexNet for a particular image is encoded by assigning points to the nearest codeword. In traditional Bag of Words, this encoding leads to a codeword histogram, i.e., a histogram for which each codeword contains points closest to this codeword. In the case of the Fisher Vector, the clusters are replaced with a Gaussian Mixture Model (GMM), and the representation encodes the log-likelihood gradients with respect to the parameters of this model. In this paper, we will use notations deep Bag of Words and deep Fisher Vector to refer to those two types of pooling methods. To make this article self-contained, we recall definitions of BoW and FV in S1 Appendix. As a result of pooling, one fixed-size vector is obtained for each of the analyzed patches, which can be classified with any machine learning methods to distinguish between various fungus species. We decided to use Support Vector Machine and Random Forest classifiers for this step.

Experimental setup and results

For the experiments, we split our DIFaS database (9 strains × 2 preparations × 10 images) into two subsets, so that both of them contain images of all strains, but from different preparation. It is because each preparation has its characteristics, and according to our previous studies [20], using images from the same preparation both in training and test set can result in overstated accuracy. As an example, let us consider the background-size, which depends on the size of the colony moved by inoculation loop from Sabouraud agar to preparation. Because there are only two preparations for each species in the dataset, the classifier could end up learning clinically irrelevant background-size instead of relevant fungus features. Therefore, images from particular preparation should not be shared between training and test set. We decided to use 2-fold cross-validation (one fold with 90 images from the first preparations and the second fold with 90 images from the second preparations). Moreover, we decided to classify patches instead of the whole image (see Image preprocessing for details) and introduce additional class corresponding to the background (BG) to compensate for the preparation characteristic on the final result. Nevertheless, we report accuracy for both patch-level and scan-level classification (the latter with majority voting). For each fold, we optimize the following parameters using internal 5-fold cross-validation: number of clusters in BoW ∈ [5, 10, 20, 50, 100, 200, 500]; number of clusters in FV ∈ [5, 10, 20, 50]; SVM kernel ∈[linear, RBF]; SVM C ∈ [1, 10, 100, 1000]; SVM γ ∈ [0.001, 0.0001]. As the evaluation metric for grid search optimization, we use the accuracy classification score. Best results were obtained for FV with 10 clusters and SVM with RBF kernel, C = 1, and γ = 0.0001. We performed all the experiments on a workstation with one 12 GB GPU and 256 GB RAM. On average, feature extraction, pooling, and classification take from 1 to 2 hours when training deep Fisher Vector. Such performance was possible thanks to the adaptation of the VLFeat library [27]. For comparison, the fine-tuning of the well-known architectures takes from 70 to 85 hours (see Table 1). Processing time in case of baseline methods was measured by multiplying the average time of an epoch by the number of epochs till the early stopping (i.e., the increase in validation loss). In the case of deep bag-of-words approaches, processing time was computed as a sum of all three steps of the algorithm (i.e., obtaining image representation, pooling, and classification).
Table 1

Test accuracy of patch-based classification averaged over two runs (for two subsets described in Experimental setup).

MethodCACGCLCNCPCTMFSBSCBGTotalTraining time (s)
AlexNet78.6 ± 4.380.0 ± 1.455.8 ± 1.463.4 ± 14.775.0 ± 7.935.0 ± 6.471.6 ± 16.667.9 ± 5.072.1 ± 2.190.0 ± 0.171.6 ± 2.4250600
DenseNet16967.9 ± 17.972.1 ± 0.753.6 ± 0.756.3 ± 13.360.0 ± 12.781.4 ± 0.768.7 ± 5.685.0 ± 5.068.8 ± 8.681.2 ± 2.072.9 ± 0.6271600
InceptionV367.3 ± 1.555.0 ± 2.259.3 ± 6.467.0 ± 18.264.3 ± 1.481.4 ± 2.961.1 ± 11.155.0 ± 0.789.3 ± 1.384.7 ± 3.169.9 ± 1.9309400
ResNet1891.4 ± 3.667.1 ± 5.767.1 ± 5.761.9 ± 7.565.0 ± 0.769.8 ± 2.854.8 ± 12.893.6 ± 3.293.5 ± 1.293.1 ± 0.575.9 ± 2.6263900
ResNet5086.4 ± 3.557.9 ± 2.189.3 ± 3.561.8 ± 11.560.0 ± 7.164.5 ± 6.066.4 ± 19.379.3 ± 0.764.3 ± 12.490.4 ± 2.773.9 ± 2.6276500
AlexNet BoW RF87.3 ± 9.739.0 ± 0.388.3 ± 7.064.0 ± 17.079.0 ± 11.079.3 ± 10.755.1 ± 3.393.3 ± 2.081.3 ± 6.092.6 ± 1.076.7 ± 1.0156
InceptionV3 BoW RF44.3 ± 7.751.0 ± 13.760.3 ± 0.334.2 ± 17.828.0 ± 0.740.7 ± 11.313.3 ± 1.439.7 ± 11.736.3 ± 8.383.5 ± 2.044.6 ± 0.3155
ResNet18 BoW RF54.3 ± 2.339.7 ± 8.380.7 ± 12.746.2 ± 20.172.0 ± 3.361.3 ± 11.342.1 ± 2.958.7 ± 8.768.0 ± 4.091.6 ± 2.062.7 ± 2.4158
AlexNet BoW SVM92.3 ± 3.044.3 ± 18.388.7 ± 9.362.2 ± 22.283.0 ± 5.771.0 ± 15.076.1 ± 11.885.0 ± 4.376.3 ± 7.089.5 ± 3.677.6 ± 1.25124
InceptionV3 BoW SVM44.3 ± 10.332.3 ± 5.757.7 ± 2.333.3 ± 17.628.3 ± 1.039.7 ± 6.313.0 ± 1.734.7 ± 8.032.0 ± 2.078.8 ± 2.140.8 ± 1.35153
ResNet18 BoW SVM59.0 ± 3.732.3 ± 5.784.0 ± 11.354.7 ± 22.066.0 ± 0.062.0 ± 4.039.7 ± 1.059.3 ± 2.073.0 ± 12.392.9 ± 1.863.4 ± 0.75195
AlexNet FV RF83.3 ± 10.754.3 ± 31.781.7 ± 0.349.3 ± 32.078.0 ± 12.073.0 ± 15.776.5 ± 5.889.3 ± 2.074.7 ± 6.788.0 ± 0.975.8 ± 0.4166
InceptionV3 FV RF40.7 ± 9.353.3 ± 11.360.3 ± 5.037.3 ± 21.727.7 ± 6.347.7 ± 13.016.5 ± 0.235.3 ± 13.332.0 ± 10.781.4 ± 4.144.6 ± 0.7168
ResNet18 FV RF63.7 ± 2.337.0 ± 5.778.7 ± 12.752.5 ± 21.173.3 ± 4.064.3 ± 5.061.8 ± 6.861.0 ± 9.769.3 ± 7.392.7 ± 2.066.4 ± 2.1167
AlexNet FV SVM93.7 ± 2.353.7 ± 16.390.7 ± 4.059.6 ± 15.277.7 ± 14.387.7 ± 9.782.8 ± 6.997.3 ± 1.381.3 ± 10.091.1 ± 2.582.4 ± 0.21541
InceptionV3 FV SVM46.0 ± 14.045.0 ± 20.358.3 ± 3.042.2 ± 20.424.0 ± 5.343.7 ± 3.713.0 ± 1.126.7 ± 5.076.7 ± 3.741.3 ± 1.941.3 ± 1.91511
ResNet18 FV SVM71.3 ± 11.335.3 ± 1.359.3 ± 10.351.6 ± 31.177.0 ± 8.376.7 ± 4.757.5 ± 0.573.3 ± 4.777.7 ± 9.794.5 ± 1.371.3 ± 1.51535
The remaining part of this section is structured as follows. First, we describe image preprocessing, including contrast stretching and background removal. Then, we describe the results obtained for patch-based classification using deep bag-of-words approaches and compare them with the well-known network architectures. To explain the outcomes of deep BoW, we introduce an in-depth explanatory analysis of the obtained codebooks together with the microbiological feedback. We continue this investigation for a deep FV approach. Finally, we present results obtained for scan-based classification, computed by aggregating patch-based scores. The code implemented in Python with PyTorch library is available at https://github.com/bziiuj/fungus.

Image preprocessing

DIFaS database contains 180 images of relatively high resolution and intensity range (from 0 to 65535); however, the actual pixel values are usually between 0 and 1000 (see Fig 4a). Therefore, in the first step of preprocessing, we compute the lower and upper-intensity limits (separately for every image) and use them for contrast stretching (see Fig 4b). Moreover, images are scaled to the range [0, 1].
Fig 4

Example image, its histogram, and foreground-background mask.

Original (a) and stretched (b) histogram of the 16-bits image; the stretched image itself (c); and its foreground-background masks with the various foreground to background proportions (d-f). Center locations of foreground and background patches are marked as yellow and green, respectively, while blue color corresponds to the area between them (omitted during classification).

Example image, its histogram, and foreground-background mask.

Original (a) and stretched (b) histogram of the 16-bits image; the stretched image itself (c); and its foreground-background masks with the various foreground to background proportions (d-f). Center locations of foreground and background patches are marked as yellow and green, respectively, while blue color corresponds to the area between them (omitted during classification). To overcome the issues with preparation characteristic (e.g., background-size), as the second step of preprocessing, we extract and classify only image patches with the reasonable foreground to background proportions (FBP), so patches with a rational number of foreground pixels. To obtain foreground-background segmentation on the pixel level, we apply thresholding (with threshold equal 0.5) to a grayscaled and blurred version of the scanned image. Such a simple segmentation is sufficient and works for all the images from the dataset (see S1 and S2 Figs) because the background is always much brighter than the areas with fungi cells. We tested three possible options of FBP: 1: 2, 1: 1, and 2: 1 (see Fig 4d–4f). Based on empirical studies, we decided to use FBP equal 2: 1, which gains around 1.5% comparing to the other options. As a result, we obtain rough segmentations with approximated locations of foreground patches (those with FBP greater than 2: 1) and background patches (those with FBP smaller than 1: 100). Additionally, we experimented with two image scales: the original images and images scaled by factor 0.5 (with bicubic interpolation), concluding that the latter gains around 4% comparing to the former.

Patch-based classification

In this experiment, we use baseline models (well-known network architectures) as well as deep Bag of Words and deep Fisher Vector models to classify each patch of the image separately. As baseline models, we fine-tune the classifier’s block of the well-known network architectures, such as AlexNet [15], DenseNet169 [16], InceptionV3 [17], and ResNet [18] for 100 epochs (with frozen features’ block). Every baseline model was previously pre-trained on the ImageNet database [14]. Before running all the experiments, we experimentally chose the optimal FBP (2: 1), patch size (500 × 500 pixels), and image scale (0.5) using grid search optimization. The number of foreground patches overlapped by less than 50% oscillates between 2000 and 3000, depending on the fold. Moreover, the number of patches significantly varies depending on the strains (see S1 and S3 Tables). Therefore, we apply data augmentation (rotations, mirror reflection, and random noise) for better regularization. The overall comparison of tested methods is presented in Table 1. One can observe that deep Fisher Vector works better than all the other techniques, including deep Bag of Words. However, its accuracy drops dramatically in the case of Candida glabrata (CG) and Cryptococcus neoformans (CN). In the case of CN it is most probably caused by a reduced number of samples, while in the case of CG due to its more substantial variance in the arrangement, appearance, and quantity (especially between two preparations, see S1 and S2 Figs). Moreover, CG images are hard to classify due to partial discoloration (pink color instead of purple) and vast overlapping of cells. As a result, CG is often classified as Candida lustianiae (CL) belonging to the same genus (see confusion matrix in Fig 5b). However, the classification error should decrease if the biological material of microscopic preparation has the smallest possible density with separated cells, as overlapping is the leading cause of blurriness.
Fig 5

Test confusion matrices.

Normalized test confusion matrices for the best baseline (ResNet18) and the best deep bag-of-words approach (deep Fisher Vector).

Test confusion matrices.

Normalized test confusion matrices for the best baseline (ResNet18) and the best deep bag-of-words approach (deep Fisher Vector). To further understand the reason for incorrect classification, we prepare a qualitative confusion matrix for deep Fisher Vector to show examples of correctly and incorrectly classified patches (see Fig 6). We observe a high morphological similarity between misclassified species belonging to genus Candida, Cryptococcus, and Saccharomyces, especially if the preparation with the biological material is discolored. Moreover, one can notice that deep Fisher Vector can return two different results for two highly overlapped patches from the same scan. It is usually caused by the artifacts in the background, such as purple trail in Candida lustianiae (CL), predicted as Cryptococcus neoformans (CN), or Maalasezia furfur (MF) in Fig 6. The other incorrect classifications appear due to the small number of incomplete (fragmented) cells (like Candida glabrata (CG) predicted as CN).
Fig 6

The qualitative confusion matrix for deep Fisher Vector (with AlexNet and SVM).

Each cell contains at most two patches of a “true” strain (represented by rows) that were classified as “predicted” strain (given by columns). Notice that the patches highly overlap, and they are often misclassified due to the artifacts in the background (such as purple trail).

The qualitative confusion matrix for deep Fisher Vector (with AlexNet and SVM).

Each cell contains at most two patches of a “true” strain (represented by rows) that were classified as “predicted” strain (given by columns). Notice that the patches highly overlap, and they are often misclassified due to the artifacts in the background (such as purple trail).

Analysis of deep Bag of Words clusters

In this section, we first analyze deep Bag of Words pooling step by visualizing clusters using the patches nearest to their centroids. Then, based on those patches, we introduce a description of the considered species using properties pre-defined by the microbiologists. Finally, we present the mean deep BoW for every species. To make our analysis clearer, in this section, we limit deep BoW to 10 clusters, although its optimal number obtained with grid search optimization is 50. Moreover, the presented properties are introduced only to explain the intrinsic rules of the method. They are not used in the automatic classification, which requires only a scan image as an input. Ten nearest neighbors of ten deep BoW centroids obtained with k-Means algorithm are presented in Fig 7. One can observe that they share common features and, therefore, can be used to determine which visual properties are essential for the classifier. We consider the following properties (see Table 2): brightness (dark or bright), size (small, medium or large), shape (circular, oval, longitudinal or variform), arrangement (regular or irregular), appearance (singular, grouped or fragmentary), color (pink, purple, blue or black), and quantity (low, medium or high). As a result, the standard set of parameters used to describe the species (size, shape, arrangement, and appearance) was significantly extended.
Fig 7

Ten nearest neighbors of deep Bag of Words centroids.

Table 2

A visual description of deep Bag of Words centroids from Fig 7.

Cluster No.BrightnessSizeShapeArrangementAppearancecolorQuantity
0brightsmallovallongitudinalregulargroupedfragmentaryblackpinkhigh
1darkmediumovalcircularirregulargroupedblacklow
2darklargelongitudinalvariformirregulargroupedfragmentaryblackmedium
3darkmediumvariformovalirregulargroupedfragmentaryblackmedium
4darklargelongitudinalirregulargroupedfragmentaryblackbluemedium
5brightsmalllongitudinalovalirregulargroupedbluepurplemedium
6darkmediumlongitudinalovalirregulargroupedfragmentaryblackmedium
7brightsmalllongitudinal ovalirregular regulargrouped fragmentarypurplemedium
8darkmediumlongitudinalovalirregulargroupedfragmentaryblackhigh
9darkmediumovalirregulargroupedblacklow
To investigate which visual properties are essential for the classifier, we calculate mean deep Bag of Words representation for every species (see Fig 8) and then examine how the visual information about their main clusters corresponds to the knowledge of a microbiologist. The main conclusions she drew are as follows:
Fig 8

Mean deep Bag of Words for individual species together with the variance.

species of the genus Candida mainly belong to cluster 2 with black cells of medium or large size, and oval or longitudinal shape; Maalasezia furfur has been assigned to clusters 0, 2, 5 and 8, mostly representing the black and longitudinal shape of various size; Saccharomyces boulardii and Saccharomyces cerevisiae are mainly described by clusters 1, 2, 4 and 8, which are characterized by black color, medium or large size and longitudinal shape; Candida tropicalis and Saccharomyces cerevisiae have very similar mean Bag of Words, which confirms high morphological similarity described in [28], i.e., size 3.0-8.0 × 5.0-10 μm, oval shape, elongated, and occurring singly or in small groups).

Analysis of deep Fisher Vector and SVM classifier

In this section, we first analyze the power of deep Fisher Vector representation using the t-SNE algorithm [29] by projecting it on a 2D surface. Then, we analyze classifier certainty based on the scores obtained for various patches. Projection of high-dimensional deep Fisher Vector to 2D using the t-SNE algorithm is presented in Fig 9. One can observe that classes are generally well separated in the case of AlexNet and ResNet18 architectures. Nevertheless, species of the same genus are not more coherent than the other species, in contrast to what we expected. Moreover, one can observe that InceptionV3 fails domain adaptation in the case of microbiological images, which explains the results in Table 1. Direct explanation of this behavior is hard, due to the neural network’s black-box character. However, it is consistent with the results reported in [30-32], which suggest that ResNet representations are more robust than those obtained from InceptionV3.
Fig 9

Projection of high-dimensional deep Fisher Vector to 2D using the t-SNE algorithm.

The second task in this section was to analyze classifier certainty. For this purpose, we investigate the distance of patches’ representations from the classifier hyperplane, which roughly corresponds to how sure the classifier is of its decisions. Most left and right patches in Fig 10 are correctly classified with high probability, while the ones in the middle are ambiguous. The most representative fungal Malassezia furfur (MF) cells have oval, longitudinal shape, and often occur in the budding form, in which the daughter cells are as wide as the parent cells. While in the case of Saccharomyces cerevisae (SC), fungal cells characterize with round shapes, more significant in relation to Candida albicans (CA), which are arranged individually or in small groups.
Fig 10

Classifier certainty for deep Fisher Vector.

From left to right, one can observe the most negative and most positive patches according to the classifier of the particular species.

Classifier certainty for deep Fisher Vector.

From left to right, one can observe the most negative and most positive patches according to the classifier of the particular species.

Scan-based classification

To analyze classification score for the whole scan (instead of just patches, like in previous sections), we predict classification for all foreground patches of one scan and aggregate them to obtain the most frequently predicted species. As presented in Table 3, deep Fisher Vector performs better than the other methods, also in this case, obtaining 15.6% better accuracy than the best baseline method (ResNet18).
Table 3

Test accuracy of scan-based classification obtained by aggregating patch-based classification and determining the most frequent prediction.

MethodCACGCLCNCPCTMFSBSCTotal
AlexNet95.5 ± 5.095.0 ± 5.070.0 ± 10.057.0 ± 6.995.0 ± 5.045.0 ± 10.070.0 ± 10.075.0 ± 5.085.0 ± 5.077.3 ± 4.2
DenseNet16980.0 ± 10.085.0 ± 5.045.0 ± 5.057.0 ± 21.270.0 ± 10.0100.0 ± 0.085.0 ± 15.095.0 ± 5.075.0 ± 10.077.6 ± 6.6
InceptionV350.0 ± 10.050.0 ± 10.080.0 ± 0.078.6 ± 14.355.0 ± 10.060.0 ± 10.050.0 ± 10.085.0 ± 5.085.0 ± 5.065.9 ± 4.9
ResNet18100.0 ± 0.075.0 ± 5.0100.0 ± 0.078.6 ± 6.950.0 ± 10.070.0 ± 0.070.0 ± 10.095.0 ± 5.080.0 ± 10.078.3 ± 5.4
ResNet50100.0 ± 0.085.0 ± 15.0100.0 ± 0.057.0 ± 6.950.0 ± 10.045.0 ± 15.080.0 ± 10.095.0 ± 5.075.0 ± 5.078.1 ± 8.3
AlexNet BoW RF100.0 ± 0.075.0 ± 25.0100.0 ± 0.075.0 ± 15.0100.0 ± 0.090.0 ± 10.090.0 ± 10.0100.0 ± 0.0100.0 ± 0.092.2 ± 4.4
InceptionV3 BoW RF80.0 ± 20.075.0 ± 25.0100.0 ± 0.045.0 ± 15.040.0 ± 0.060.0 ± 30.00.0 ± 0.070.0 ± 10.045.0 ± 15.057.2 ± 2.8
ResNet18 BoW RF100.0 ± 0.070.0 ± 10.0100.0 ± 0.045.0 ± 15.0100.0 ± 0.0100.0 ± 0.080.0 ± 10.090.0 ± 10.0100.0 ± 0.087.2 ± 1.7
AlexNet BoW SVM100.0 ± 0.065.0 ± 25.0100.0 ± 0.070.0 ± 10.0100.0 ± 0.0100.0 ± 0.085.0 ± 5.0100.0 ± 0.0100.0 ± 0.091.1 ± 2.2
InceptionV3 BoW SVM70.0 ± 10.055.0 ± 15.0100.0 ± 0.040.0 ± 20.045.0 ± 5.055.0 ± 15.00.0 ± 0.055.0 ± 15.040.0 ± 20.051.1 ± 2.3
ResNet18 BoW SVM100.0 ± 0.060.0 ± 20.0100.0 ± 0.065.0 ± 0.0100.0 ± 0.090.0 ± 10.060.0 ± 0.095.0 ± 5.0100.0 ± 0.085.6 ± 2.2
AlexNet FV RF100.0 ± 0.065.0 ± 35.0100.0 ± 0.055.0 ± 5.0100.0 ± 0.090.0 ± 10.095.0 ± 5.0100.0 ± 0.0100.0 ± 0.089.4 ± 2.2
InceptionV3 FV RF65.0 ± 5.095.0 ± 5.0100.0 ± 0.050.0 ± 10.030.0 ± 10.075.0 ± 25.05.0 ± 5.045.0 ± 25.045.0 ± 35.056.7 ± 3.3
ResNet18 FV RF95.0 ± 5.060.0 ± 0.0100.0 ± 0.065.0 ± 5.0100.0 ± 0.0100.0 ± 0.095.0 ± 5.095.0 ± 5.095.0 ± 5.089.4 ± 1.7
AlexNet FV SVM100.0 ± 0.075.0 ± 25.0100.0 ± 0.075.0 ± 15.0100.0 ± 0.0100.0 ± 0.095.0 ± 5.0100.0 ± 0.0100.0 ± 0.093.9 ± 3.9
InceptionV3 FV SVM75.0 ± 25.060.0 ± 0.0100.0 ± 0.055.0 ± 5.045.0 ± 15.085.0 ± 5.05.0 ± 5.025.0 ± 15.045.0 ± 25.055.0 ± 5.6
ResNet18 FV SVM100.0 ± 0.060.0 ± 0.0100.0 ± 0.045.0 ± 15.095.0 ± 5.0100.0 ± 0.095.0 ± 5.0100.0 ± 0.0100.0 ± 0.088.3 ± 2.7

Conclusions and future work

In this paper, we apply deep neural networks and bag-of-words approaches to classify microscopic images of various fungi species. According to our experiments, the combination of features from deep neural networks with Fisher Vector works better than fine-tuning the classifier’s block of the well-known network architectures and has the potential to be successfully used by microbiologists in their daily practice. A large part of this paper is dedicated to the explainability of deep bag-of-words approaches to increase the trust in deep neural networks. For this purpose, we introduce an in-depth visual description of the properties pre-defined by the microbiologists. We hope that it will help to understand similarities and differences between fungi species better. In our experiment, we assumed that images are obtained from the same laboratory and with the same scanner (details are presented in the Materials). However, in our opinion, this method could be easily extended to more diversified datasets by using additional preprocessing steps, which unify the input data. Due to the lack of data for such experiments, we did not cover this issue in the current article; however, it is planned for future research. Moreover, we would like to extend the DIFaS database so that it contains more preparations for all species, also gathered from other laboratories and scanners. Finally, we plan to prepare scans containing more than one species, as the automatic classification of such images would help to exclude the culture phase from the microbiological pipeline. (PDF) Click here for additional data file. (PDF) Click here for additional data file.

The number of foreground patches overlapped by less than 50% for the images from S1 Fig.

(PDF) Click here for additional data file.

The number of background patches overlapped by less than 50% for the images from S1 Fig.

(PDF) Click here for additional data file.

The number of foreground patches overlapped by less than 50% for the images from S2 Fig.

(PDF) Click here for additional data file.

The number of background patches overlapped by less than 50% for the images from S2 Fig.

(PDF) Click here for additional data file.

Images of the first preparations from the DIFaS database and their foreground-background masks.

(TIF) Click here for additional data file.

Images of the second preparations from the DIFaS database and their foreground-background masks.

(TIF) Click here for additional data file.
  14 in total

1.  SIFT flow: dense correspondence across scenes and its applications.

Authors:  Ce Liu; Jenny Yuen; Antonio Torralba
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2011-05       Impact factor: 6.226

Review 2.  Candida glabrata, Candida parapsilosis and Candida tropicalis: biology, epidemiology, pathogenicity and antifungal resistance.

Authors:  Sónia Silva; Melyssa Negri; Mariana Henriques; Rosário Oliveira; David W Williams; Joana Azeredo
Journal:  FEMS Microbiol Rev       Date:  2011-06-06       Impact factor: 16.408

3.  Detection and identification of fungal pathogens by PCR and by ITS2 and 5.8S ribosomal DNA typing in ocular infections.

Authors:  C Ferrer; F Colom; S Frasés; E Mulet; J L Abad; J L Alió
Journal:  J Clin Microbiol       Date:  2001-08       Impact factor: 5.948

Review 4.  Candida parapsilosis, an emerging fungal pathogen.

Authors:  David Trofa; Attila Gácser; Joshua D Nosanchuk
Journal:  Clin Microbiol Rev       Date:  2008-10       Impact factor: 26.132

5.  Serologic evidence for reactivation of cryptococcosis in solid-organ transplant recipients.

Authors:  D C Saha; D L Goldman; X Shao; A Casadevall; S Husain; A P Limaye; M Lyon; J Somani; K Pursell; T L Pruett; N Singh
Journal:  Clin Vaccine Immunol       Date:  2007-10-24

Review 6.  Fungal infections in solid organ transplantation.

Authors:  Fernanda P Silveira; Shahid Husain
Journal:  Med Mycol       Date:  2007-06       Impact factor: 4.076

Review 7.  Candida glabrata: a review of its features and resistance.

Authors:  C F Rodrigues; S Silva; M Henriques
Journal:  Eur J Clin Microbiol Infect Dis       Date:  2013-11-19       Impact factor: 3.267

8.  Sensitization of Candida albicans to terbinafine by berberine and berberrubine.

Authors:  Pikling Lam; Stanton Hon Lung Kok; Kenneth Ka Ho Lee; Kim Hung Lam; Desmond Kwok Po Hau; Wai Yeung Wong; Zhaoxiang Bian; Roberto Gambari; Chung Hin Chui
Journal:  Biomed Rep       Date:  2016-02-23

9.  Deep learning approach to bacterial colony classification.

Authors:  Bartosz Zieliński; Anna Plichta; Krzysztof Misztal; Przemysław Spurek; Monika Brzychczy-Włoch; Dorota Ochońska
Journal:  PLoS One       Date:  2017-09-14       Impact factor: 3.240

Review 10.  Fungal Identification Using Molecular Tools: A Primer for the Natural Products Research Community.

Authors:  Huzefa A Raja; Andrew N Miller; Cedric J Pearce; Nicholas H Oberlies
Journal:  J Nat Prod       Date:  2017-02-15       Impact factor: 4.050

View more
  5 in total

1.  Species determination using AI machine-learning algorithms: Hebeloma as a case study.

Authors:  Peter Bartlett; Ursula Eberhardt; Nicole Schütz; Henry J Beker
Journal:  IMA Fungus       Date:  2022-06-30       Impact factor: 8.044

Review 2.  Machine Learning and Deep Learning Based Computational Approaches in Automatic Microorganisms Image Recognition: Methodologies, Challenges, and Developments.

Authors:  Priya Rani; Shallu Kotwal; Jatinder Manhas; Vinod Sharma; Sparsh Sharma
Journal:  Arch Comput Methods Eng       Date:  2021-08-31       Impact factor: 8.171

3.  Medical Image Recognition Technology in the Effect of Substituting Soybean Meal for Fish Meal on the Diversity of Intestinal Microflora in Channa argus.

Authors:  Aixia Huang; Lihui Sun; Feng Lin; Jianlin Guo; Jianhu Jiang; Binqian Shen; Jianming Chen
Journal:  J Healthc Eng       Date:  2021-11-25       Impact factor: 2.682

4.  Detection of Cytopathic Effects Induced by Influenza, Parainfluenza, and Enterovirus Using Deep Convolution Neural Network.

Authors:  Jen-Jee Chen; Po-Han Lin; Yi-Ying Lin; Kun-Yi Pu; Chu-Feng Wang; Shang-Yi Lin; Tzung-Shi Chen
Journal:  Biomedicines       Date:  2021-12-30

5.  Automatic Fungi Recognition: Deep Learning Meets Mycology.

Authors:  Lukáš Picek; Milan Šulc; Jiří Matas; Jacob Heilmann-Clausen; Thomas S Jeppesen; Emil Lind
Journal:  Sensors (Basel)       Date:  2022-01-14       Impact factor: 3.576

  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.