Literature DB >> 28798957

Holographic deep learning for rapid optical screening of anthrax spores.

YoungJu Jo1, Sangjin Park2,3, JaeHwang Jung1, Jonghee Yoon1, Hosung Joo4, Min-Hyeok Kim5, Suk-Jo Kang5, Myung Chul Choi6, Sang Yup Lee2, YongKeun Park1,7.   

Abstract

Establishing early warning systems for anthrax attacks is crucial in biodefense. Despite numerous studies for decades, the limited sensitivity of conventional biochemical methods essentially requires preprocessing steps and thus has limitations to be used in realistic settings of biological warfare. We present an optical method for rapid and label-free screening of Bacillus anthracis spores through the synergistic application of holographic microscopy and deep learning. A deep convolutional neural network is designed to classify holographic images of unlabeled living cells. After training, the network outperforms previous techniques in all accuracy measures, achieving single-spore sensitivity and subgenus specificity. The unique "representation learning" capability of deep learning enables direct training from raw images instead of manually extracted features. The method automatically recognizes key biological traits encoded in the images and exploits them as fingerprints. This remarkable learning ability makes the proposed method readily applicable to classifying various single cells in addition to B. anthracis, as demonstrated for the diagnosis of Listeria monocytogenes, without any modification. We believe that our strategy will make holographic microscopy more accessible to medical doctors and biomedical scientists for easy, rapid, and accurate point-of-care diagnosis of pathogens.

Entities:  

Mesh:

Year:  2017        PMID: 28798957      PMCID: PMC5544395          DOI: 10.1126/sciadv.1700606

Source DB:  PubMed          Journal:  Sci Adv        ISSN: 2375-2548            Impact factor:   14.136


INTRODUCTION

Bacillus anthracis, a gram-positive spore-forming bacterium causing the disease anthrax, is one of the most destructive biological weapons, which is prone to be abused for bioterrorism (). It is thus crucial to rapidly detect and identify anthrax spores for biodefense (). Various biological, chemical, and optical fingerprinting methods have been studied to accelerate diagnosis of B. anthracis (–). Conventional culture-based methods take days and are often inaccurate. Polymerase chain reaction–based methods provide species-level specificity but still take hours and require heavy instrumentation with skilled personnel to operate the system (). Photoluminescence and surface-enhanced Raman scattering methods take only minutes but require labeling with exogenous agents and cannot discriminate B. anthracis from other Bacillus species that are ubiquitous in nature (). Most of these methods are limited by detection sensitivities that require minimum sample sizes of at least thousands of bacterial cells; thus, their applications in practical settings such as aerosolized spores require sample amplification processes that significantly limit the detection speed. Recent developments of optical methods based on holographic microscopy combined with machine learning, which enables rapid and label-free identification of single cells, can be an important step to address the anthrax issue (–). These techniques were pioneered by Javidi’s group (–) and developed further by several groups (–). Holographic microscopy (), or quantitative phase imaging (QPI) in a broader sense, measures optical field images (that is, nanometer-scale distortions of wavefronts passing through a sample) using laser-based interferometry. In addition to the amplitude images available from conventional intensity-based microscopy techniques, holographic microscopy quantitatively measures the optical phase delay maps dictated by the refractive index (RI) distribution of a sample (). Because the endogenous RI distribution in a cell is strongly related to the structural and biochemical characteristics () of the target classes (for example, species or cell types), the measured field images of single cells and the corresponding class labels are passed to data-driven machine learning algorithms for systematic discovery of class-specific fingerprints encoded in the images. These approaches can be combined with flow cytometry and/or bioaerosol collection systems to achieve ultrafast identification of unlabeled cells and pathogens (, ). However, none of these methods achieved subgenus specificity required for discriminating B. anthracis from other Bacillus species ubiquitous in nature. Here, we present a next-generation holographic screening method by adopting “deep learning,” a state-of-the-art machine learning technique based on deep multilayered neural networks (, ), to holographic microscopy. We designed a deep convolutional neural network (CNN), HoloConvNet, specialized in the classification of holographic images of living cells. After training with quantitative phase images of individual Bacillus spores, the network identified new anthrax spores with single-spore sensitivity and subgenus specificity. Its remarkable learning ability enables direct training from raw images by automatically recognizing key biological traits encoded in the images, including the dry mass of individual bacteria, and presents outstanding accuracy that outperforms previous approaches in all accuracy measures. As demonstrated below, this method is readily applicable to classification of various single cells, in addition to B. anthracis, without any modification.

RESULTS

The overall framework of our method is shown in Fig. 1. We used QPI unit (QPIU), a cost-effective palm-sized module that converts a conventional microscope into a holographic microscope (), for phase imaging of individual Bacillus spores in an isolated biosafety level 3 (BSL-3) laboratory at the Agency for Defense Development, Korea. It is attached to the output port of an existing bright-field microscope to form a common-path interferometry for optical field imaging (Fig. 1, A to C). After imaging B. anthracis and four different Bacillus species with various levels of phylogenetic relatedness (see Materials and Methods), we trained our deep neural network named HoloConvNet as a species classifier using the phase images of individual spores and the corresponding species labels (training set). The learnable parameters of the deep neural network were iteratively adjusted by the error backpropagation algorithm (Fig. 1D) (, ). The performance of the trained HoloConvNet was tested by taking new images (test set), which were never seen before by the network, as the input to the network (Fig. 1E). The machine-predicted species labels were compared with the true classes to estimate identification accuracy.
Fig. 1

Holographic deep learning framework for screening of anthrax spores.

(A) Schematic diagram of QPIU for holographic imaging of individual Bacillus spores. (B) Interferogram formed by spatial modulation. It encodes quantitative phase images of individual spores, as shown in (C). (D) The measured phase images from multiple Bacillus species are used to train a deep neural network using the error backpropagation algorithm. (E) The trained network accurately predicts the corresponding species when independently measured phase images are shown.

Holographic deep learning framework for screening of anthrax spores.

(A) Schematic diagram of QPIU for holographic imaging of individual Bacillus spores. (B) Interferogram formed by spatial modulation. It encodes quantitative phase images of individual spores, as shown in (C). (D) The measured phase images from multiple Bacillus species are used to train a deep neural network using the error backpropagation algorithm. (E) The trained network accurately predicts the corresponding species when independently measured phase images are shown. The following five Bacillus species were used in this study. (i) B. anthracis Sterne is an attenuated strain of B. anthracis (). The Sterne strain has the pXO1 plasmid encoding anthrax toxins. However, it lacks the pXO2 plasmid that encodes the polysaccharide capsule, whose role is to defend against phagocytosis by the immune cells in the vegetative state. Because we were interested in the sporulated state of B. anthracis, it was appropriate to choose the Sterne strain as a representative strain instead of fully virulent strains (for example, the Ames strain) for safety reasons. Although B. anthracis was the key target species in the present study, we selected other Bacillus species with various levels of relatedness to this species to assess the subgenus specificity of the proposed method (). (ii) B. thuringiensis and (iii) B. cereus are the closest neighbors of B. anthracis (). These three species form the so-called B. cereus group. Despite their genetic similarity, their pathogenic properties differ by species: the causative agent of anthrax, the source of powerful biological insecticides, and the cause of several food-borne illnesses, respectively. (iv) B. atrophaeus and (v) B. subtilis are the common simulants of B. anthracis (). These two species are closely related to each other; however, they have some distance from the species mentioned above (). These species have been the most prominent surrogate organisms for B. anthracis in biodefense programs. We also note that B. subtilis is one of the best-studied model microbes, which is considered the gram-positive counterpart of Escherichia coli. The detailed data on Bacillus spores are presented in table S1. The quantitative nature of holographic microscopy captures subcellular phase delay distribution that could be exploited by machine learning algorithms to extract fingerprint information (, ). On the other hand, conventional techniques (for example, phase contrast microscopy) provide rough morphological information only (fig. S1). Simple morphological parameters such as spore size (fig. S2), which is consistent with previous reports (, ), are not enough for species discrimination due to high genetic similarities and large cell-to-cell variations (, ). The endogenous RI distribution of Bacillus spores, which dictates the sample-induced phase delay imaged by QPIU, is strongly related to specific characteristics of each species (, ). However, because this relation is often indirect, it should be approximated using supervised learning. The precision of this function approximation obviously dominates the performance of the trained classifiers. Deep neural networks are universal approximators for virtually any arbitrarily nonlinear functions (), whereas conventional machine learning techniques mostly rely on linear or only slightly nonlinear decision boundaries (). The network architecture of HoloConvNet is illustrated in Fig. 2 (and table S2). A phase image of a single spore is processed by multiple layers of convolution, nonlinearity, and pooling operations and then finally receives scored class labels through fully connected layers. The network makes its prediction by selecting the final-layer neuron with the strongest activation. The key functional block of this process is a convolutional layer followed by nonlinearitywhere x and y are input and output vectors, respectively, and w and b are synaptic weights and biases, respectively. Equation 1 emulates integration of synaptic inputs by a biological neuron (, ) that fires only when the net input exceeds a certain threshold [more precisely, a population of neurons with an output firing rate modeled by a rectified linear unit (ReLU)]. Note that the entire processing by the network from images to class labels is a nonlinear mapping that corresponds to the approximating function explained above.
Fig. 2

Architecture of HoloConvNet.

When a phase image of an individual spore is taken as the input, the network first processes the images through three rounds of convolution, ReLU nonlinearity, and max pooling layers. Then, two fully connected (and ReLU) layers follow: (i) the last hidden layer under dropout regularization and (ii) the output layer with the class scores. These scores are used to calculate the loss function and to make species predictions in the training and test stages, respectively. Only 10 two-dimensional activation maps per layer are presented with layer-wise scaling for visualization (see table S2 for detailed architecture).

Architecture of HoloConvNet.

When a phase image of an individual spore is taken as the input, the network first processes the images through three rounds of convolution, ReLU nonlinearity, and max pooling layers. Then, two fully connected (and ReLU) layers follow: (i) the last hidden layer under dropout regularization and (ii) the output layer with the class scores. These scores are used to calculate the loss function and to make species predictions in the training and test stages, respectively. Only 10 two-dimensional activation maps per layer are presented with layer-wise scaling for visualization (see table S2 for detailed architecture). Training a deep neural network is essentially a large-scale nonlinear optimization of the synaptic weights (and biases) that govern the network behavior. The large number of the learnable parameters makes training process extremely difficult. However, CNNs such as HoloConvNet have markedly smaller number of parameters (, ) by using localized and shared receptive field structures inspired by physiological visual processing. Thus, the network can be trained using the error backpropagation algorithm that minimizes the mismatch between the machine-predicted and true labels (see Materials and Methods). HoloConvNet efficiently converges to a hierarchical representation of the images that gradually transforms the data space in which the classes are easily separable. This property is called the “representation learning” capability of deep learning () and enables direct training from raw images. The performance of HoloConvNet is shown in Fig. 3 (and fig. S3 with different visualization). A well-trained neural network reflects the general relations between the input and output data so that it accurately predicts the class labels of new images (generalization property). First, the multiclass identification performance of the network for the five Bacillus species (B. anthracis, B. thuringiensis, B. cereus, B. atrophaeus, and B. subtilis), trained with five class labels representing individual species, is shown in Fig. 3A. HoloConvNet identifies B. anthracis spores from the other four species with high sensitivity and specificity (table S3), despite relatively less accurate classification between the other species that is irrelevant for the screening of anthrax spores.
Fig. 3

Performance of HoloConvNet.

(A to C) The test images are used to measure the performance of (A) multiclass classification of the five Bacillus species (B. anthracis, B. thuringiensis, B. cereus, B. atrophaeus, and B. subtilis), (B) binary classification of B. anthracis and the other four species (B. thuringiensis, B. cereus, B. atrophaeus, and B. subtilis), and (C) binary classification of B. anthracis and the two nonmember species of the B. cereus group (B. atrophaeus and B. subtilis). (D) The performance of the proposed method is compared to previous techniques (see the main text). Holographic microscopy and deep learning significantly improve the performance in all cases. (E to G) t-SNE visualization of the CNN codes at the last hidden layer, corresponding to the classification schemes of (A) to (C), which shows the representation learning capability of HoloConvNet. The error bars in (A) to (D) indicate the SD calculated from 10 classification models with different random initializations.

Performance of HoloConvNet.

(A to C) The test images are used to measure the performance of (A) multiclass classification of the five Bacillus species (B. anthracis, B. thuringiensis, B. cereus, B. atrophaeus, and B. subtilis), (B) binary classification of B. anthracis and the other four species (B. thuringiensis, B. cereus, B. atrophaeus, and B. subtilis), and (C) binary classification of B. anthracis and the two nonmember species of the B. cereus group (B. atrophaeus and B. subtilis). (D) The performance of the proposed method is compared to previous techniques (see the main text). Holographic microscopy and deep learning significantly improve the performance in all cases. (E to G) t-SNE visualization of the CNN codes at the last hidden layer, corresponding to the classification schemes of (A) to (C), which shows the representation learning capability of HoloConvNet. The error bars in (A) to (D) indicate the SD calculated from 10 classification models with different random initializations. Because diagnosing anthrax spores from other species is our prime objective, the network was next trained with binary class labels (anthrax versus non-anthrax). With this method, the performance could be enhanced (Fig. 3B and table S4) by letting the optimization process focus on the characteristics distinguishing B. anthracis from others (B. thuringiensis, B. cereus, B. atrophaeus, and B. subtilis). When the problem was relaxed by excluding the two B. cereus group species (B. thuringiensis and B. subtilis), HoloConvNet achieved a remarkable accuracy of 96.3% (Fig. 3C and table S5). These results suggest the potential of deep learning–based holographic screening of anthrax spores in realistic settings. In Fig. 3D, the performance of our method (“Holography + Deep”) was compared with the performance of several previous techniques: holographic microscopy with conventional machine learning (“Holography + Conventional”) (), conventional microscopy with deep learning (“Conventional + Deep”) (training HoloConvNet with binary morphology images; see Materials and Methods), and conventional microscopy with conventional machine learning (“Conventional + Conventional”) (linear discriminant analysis with the morphological parameters in fig. S2). More extensive comparisons to various conventional classifiers, such as support vector machines, are presented in tables S6 and S7. HoloConvNet outperformed the previous methods in all performance measures with a substantial margin. Representation learning by HoloConvNet, the fundamental improvement developed in this study, was further examined (Figs. 3, E to G, and 4). The network transforms the images into a representation in which the data points are linearly separable because a single layer of neurons is a linear classifier (). We applied t-distributed stochastic neighbor embedding (t-SNE), a high-dimensional data visualization technique (), to the activation of individual neurons in the last hidden layer (Fig. 3, E to G). The good separation observed indicates the great ability of HoloConvNet to learn the optimal representation of phase images without any predesigned features required by conventional machine learning techniques. The different degrees of separation in the three cases explain the different identification performance. Additionally, the relative distances between the species clusters shown in Fig. 3E are consistent with the phylogenetic relationship explained above; it should be noted here that the relationship was independently discovered by HoloConvNet through training.
Fig. 4

Representation learning by HoloConvNet: Dry mass as a key biological trait.

The interspecies difference in cellular dry mass is automatically recognized and used for screening of anthrax spores. (A to C) The activation of the “anthrax neuron” at the output layer shows a strong correlation with dry mass. a.u., arbitrary units. (D) Dry mass of individual Bacillus spores calculated from the quantitative phase images. (E) Computationally disabling the dry mass information significantly impairs the performance of HoloConvNet. Dry mass alone is not enough for full performance as well. Data in (D) are presented as box-and-whisker plots displaying median and interquartile ranges. The error bars in (E) indicate the SD calculated from 10 classification models with different random initializations.

Representation learning by HoloConvNet: Dry mass as a key biological trait.

The interspecies difference in cellular dry mass is automatically recognized and used for screening of anthrax spores. (A to C) The activation of the “anthrax neuron” at the output layer shows a strong correlation with dry mass. a.u., arbitrary units. (D) Dry mass of individual Bacillus spores calculated from the quantitative phase images. (E) Computationally disabling the dry mass information significantly impairs the performance of HoloConvNet. Dry mass alone is not enough for full performance as well. Data in (D) are presented as box-and-whisker plots displaying median and interquartile ranges. The error bars in (E) indicate the SD calculated from 10 classification models with different random initializations. The outstanding performance of the proposed method raises a question: What are the key biological traits that are measured and exploited for the identification of anthrax spores? We speculated that cellular dry mass (), the mass of nonaqueous cellular components, is one of the most important traits. This hypothesis is based on the domain knowledge that there exists an additional outermost structure, called the exosporium, in the B. cereus group spores but not in the remaining two species (). It was reasoned that structural distinction might result in an interspecies difference of dry mass that is inherently measured by holographic imaging with femtogram-level sensitivity (see Materials and Methods). A strong positive correlation was found between dry mass and activation of the “anthrax neuron” at the output layer (Fig. 4, A to C). This observation makes sense if the mean dry mass of B. anthracis is the heaviest among the five species, which turns out to be true (Fig. 4D). As expected, B. anthracis is slightly heavier than the other two B. cereus group species, and the remaining two species lacking exosporium have considerably lighter dry mass. The subtle difference within the B. cereus group might be due to species-dependent compositions and nanostructures of exosporium (), although their contribution to dry mass should be confirmed by additional investigations. It was noted that the same order relation of dry mass was observed in all independent measurements (fig. S4), and the overall range of measured dry mass is consistent with previous studies (, ). To confirm the causality between dry mass and species prediction by HoloConvNet, we used a computational disabling strategy. Detrimental effects on the performance were observed by computationally normalizing the phase images to remove the dry mass information. As shown in Fig. 4E, the network trained and tested with normalized phase images shows a significantly impaired performance, supporting the key role of dry mass. However, it does not mean that dry mass is the sole information extracted by the network; the performance of a single-feature linear discriminant classifier based solely on dry mass was also significantly worse. This suggests that other traits such as spatial distribution of subcellular components in the spores play roles in screening. From these observations, it can be concluded that the interspecies difference of dry mass is recognized and exploited through representation learning by HoloConvNet. Here, it should be emphasized that we never taught the network on how to calculate dry mass from phase images. On the other hand, a conventional machine learning algorithm cannot make use of dry mass unless it is manually selected by a researcher. Finally, the generality of our method expected from the outstanding learning abilities was investigated. As a proof-of-concept example, HoloConvNet was trained for diagnosing the pathogen Listeria monocytogenes, the causative agent of listeriosis that is often fatal to neonates and the elderly (), from five different Listeria species (see table S8 for data specification). The diagnostic accuracy was surprisingly high, showing higher than 85% (fig. S5). The architecture and learning rules were identical to those used for the diagnosis of Bacillus species. It is also noted that L. monocytogenes is not the species with the heaviest dry mass in this case (fig. S6). Accordingly, the other classifiers based only or mostly on dry mass show markedly lower performance compared to the equivalent comparison in the Bacillus problem (fig. S5, B and C). The superior performance of HoloConvNet suggests that, for the Listeria problem, the deep neural network discovers and exploits key biological traits other than dry mass. These results suggest that the holographic deep learning framework reported here has immediate and wide applicability, in contrast to problem-specific conventional machine learning approaches.

DISCUSSION

We proposed and experimentally demonstrated a novel method for screening of anthrax spores by combining holographic microscopy and deep learning for the first time. The new strategy enables rapid label-free identification of individual anthrax spores with subgenus specificity, extending our previous intergenus bacterial fingerprinting method based on conventional machine learning (). In addition to the superior performance due to the extreme flexibility of deep neural networks, the transition from classical machine learning to deep learning fundamentally transforms holographic single-cell identification techniques by acquiring the representation learning capability. HoloConvNet automatically recognizes and then uses key biological characteristics that are species-dependent (for example, dry mass in the anthrax problem) from raw images. Additionally, the present method can be readily extended to other single-cell classification problems, such as the diagnosis of L. monocytogenes demonstrated in this study, without any modification. Thus, our method eliminates the need to manually design and optimize features based on trial and error for individual problems. The next steps beyond this proof-of-concept study to achieve practical ultrafast screening of anthrax spores are straightforward. Above all, the proposed method should be combined with flow cytometry (, ) and bioaerosol collection () systems to fully exploit the single-spore and label-free nature of the method. Then, a large amount of holographic imaging data from the resultant high-throughput device would be used to train HoloConvNet for more species and strains under various environmental conditions to assure stable field performance. The performance could be further improved by adopting multimodal QPI [for example, spectral (), polarimetric (), or tomographic () images as the stacked input to the network] to increase the amount of raw information investigated by the network. Despite the fast and label-free nature of holographic microscopy, the limited chemical specificity has left this tool overshadowed by fluorescence microscopy. Specific domain knowledge (for example, homogeneity of hemoglobin concentration in red blood cells and high RI of lipid droplets in eukaryotic cells) has been required for effective use of the technique. The method proposed in this paper solves this difficulty by using the powerful learning abilities of deep neural networks. As we demonstrated in this study, intelligent holographic microscopy can now actively recognize and exploit the class-specific fingerprints, encoded in the raw images of various biological samples, without any previous knowledge. We believe that our strategy will make holographic microscopy more accessible to medical doctors and biomedical scientists for easy, rapid, and accurate diagnosis of pathogens and facilitate exciting new applications.

MATERIALS AND METHODS

Experimental design

Preparation of Bacillus spores

B. anthracis Sterne (pXO1+ and pXO2−) was obtained from the Centers for Disease Control and Prevention, Korea (KCDC). B. thuringiensis BGSC 4AJ1 was obtained from the Bacillus Genetic Stock Center (BGSC). B. cereus ATCC 4342 was obtained from the American Type Culture Collection (ATCC). B. atrophaeus KCCM 11314 was obtained from the Korean Culture Center for Microorganisms (KCCM). B. subtilis 168 was obtained from the Korean Collection for Type Cultures (KCTC). All experiments involving B. anthracis were approved by the institutional biosafety committee and conducted in a BSL-3 laboratory following the regulations in the Republic of Korea. Bacterial cells from frozen glycerol stocks were streaked onto Luria-Bertani (LB) agar plates and incubated at 30°C overnight. The next day, a single colony was inoculated into 5 ml of LB broth in a 50-ml CELLSTAR CELLreactor tube (Greiner Bio-One) and incubated at 30°C with shaking (200 rpm) for 8 hours. Then, 250 μl of the culture broth was transferred to 25 ml of GYS (glucose yeast salt) sporulation medium () in a 125-ml polycarbonate Erlenmeyer flask with a vent cap (Corning) and incubated at 30°C with shaking (200 rpm) for 48 hours. After sporulation was completed, spores were harvested by centrifugation (5420g, 4°C) and washed four times with phosphate-buffered saline (PBS) (Life Technologies). Finally, the spores were suspended in 5 ml of PBS and stored at 4°C until use. Note that we prepared all the species with the same procedure. A small volume (approximately 10 μl) of the bacterial solution was placed in an imaging chamber composed of standard cover glasses (C024501, Matsunami Glass) and (optional) spacers with a thickness of 20 to 30 μm. Imaging was performed at room temperature after the spores settled down to the bottom and spread into a single layer. Although the optimal conditions for cultivation and sporulation somewhat vary, all the species were prepared using the same protocol to guarantee that the screening system recognized only the species-dependent characteristics. Additionally, all the procedures for spore preparation and imaging were performed multiple times to ensure the independence of the training and test sets.

Holographic imaging

Because all anthrax experiments had to be conducted in a separate BSL-3 facility at the Agency for Defense Development, we used a compact and portable QPIU recently developed in our group (), as the holographic imaging modality. It consists of two polarizers (LPVISE100-A, Thorlabs Inc.) and a Rochon prism (#68-824, Edmund Optics Inc.) inside an aluminum tube mounted in front of a charge-coupled device (CCD) camera (FL3-U3-88S2C-C, Point Grey). Inserting the unit into the output port of a conventional bright-field microscope (B-382PLi-ALC, Optika) converts it into a holographic microscope. The light source for illumination was a diode laser (CPS532, λ = 532 nm, 4.5 mW; Thorlabs Inc.), and the total magnification was ×100 determined by an objective lens (M-148; numerical aperture, 1.25; oil immersion; Optika). Acquisition time per interferogram was less than 20 ms, which could be reduced by many orders with high-intensity light sources and more sensitive cameras. QPIU, shown in Fig. 1A, is a spatially modulated self-reference interferometry. When the light passing through the sample encounters the unit, it becomes linearly polarized by the front polarizer. Then, the following Rochon polarizing prism divides the beam into two duplicated beams with slightly different propagation directions. Finally, the orthogonal polarization states of the divided beams become parallel by the rear polarizer. Thus, the two beams of parallel polarization generate an interference pattern at the overlapped region on the CCD plane. The linear polarizers before and after the prism are adjusted so that the interferogram has a high visibility (Fig. 1B). The quantitative phase information is retrieved (Fig. 1C) from the measured interferogram using a standard field retrieval algorithm (). The details on the principle of QPIU can be found elsewhere ().

Convolutional neural networks

The network architecture of HoloConvNet shown in Fig. 2 is categorized as a CNN, which is the most popular deep learning framework for image classification and still has high interest in research (, ). Because the input for a deep learning algorithm is a raw image, instead of extracted features, the large number of learnable parameters (see Eq. 1 in the main text and filter dimensions shown in table S2) makes the training process extremely difficult. Using a CNN can resolve this issue by reducing the dimensionality through several constraints on the synaptic weights. Inspired by the classical work on early visual cortex (), the synaptic weights of a CNN are localized and shared to mimic the receptive fields of biological vision. The resulting network exploits local spatial correlations to be robust under natural distortions, just as in physiological visual processing, with a significantly smaller number of parameters to be optimized. Although dimensionality reduction in conventional machine learning is done by manually designed feature extraction using problem-dependent domain expertise (, , –), biologically inspired intuition about the general properties of images mentioned above can have the same role in deep learning. This point is important because it enables direct training from raw images.

The Listeria experiments

The six major bacterial species of the genus Listeria, namely, L. monocytogenes (10403S), L. grayi (ATCC 19120), L. innocua (ATCC 33090), L. ivanovii (ATCC 19119), L. seeligeri (ATCC 35967), and L. welshimeri (ATCC 35897), were cultured in brain-heart infusion medium without antibiotics. After culturing overnight in a 37°C shaking incubator, the vegetative bacterial cells were washed and diluted with PBS based on the cultured concentration estimated by optical density measurements at 600 nm. The bacterial solution was placed and imaged in imaging chambers described above for the Bacillus experiments. The holographic imaging of the prepared samples was done with a Mach-Zehnder interferometer () with varying illumination angles to exploit the high-resolution synthetic aperture imaging technique (). Optical field reconstruction and image processing protocols were identical to those of the Bacillus experiments. The same network architecture and learning rule for training the original HoloConvNet (for Bacillus spores) were used to train the network for Listeria. The only preprocessing was to adjust the size of the input images to match the input dimension of HoloConvNet.

Statistical analysis

Image analysis

All image analysis procedures were done with MatLab (R2014b, MathWorks Inc.). The reconstructed phase images containing multiple spores were segmented by phase thresholding to be separated into images of single spores. The isolated spores were computationally aligned at the centers of square backgrounds for further analysis. The segmented regions were considered as the morphologies of individual spores that could be measured with conventional microscopy techniques such as phase contrast microscopy. The representative morphological parameters plotted in fig. S2 were quantified with the regionprops function of MatLab. For implementing the conventional classifiers, the built-in functions in the Statistics and Machine Learning Toolbox of MatLab were used. Calculation of the single-spore dry mass from phase images exploited the well-known proportionality between the optical phase delay and cellular dry mass (). The total dry mass (m) can be calculated from the phase delay map () as followswhere λ is the illumination wavelength, S is the projection area of the cell surface, and α is the RI increment for nonaqueous molecules. Because the RI increment is known to be 0.18 to 0.21 ml/g for typical biological cells (), we used α = 0.2 ml/g, and the results were consistent with those measured by other techniques (, , ). Note that we never explicitly taught the network about this relation.

Deep learning

HoloConvNet is a CNN designed for the classification of holographic images of single cells. We implemented HoloConvNet using the MatConvNet () framework (version 1.0, beta 20) because of its simplicity and compatibility with our experimental data primarily processed with MatLab. The final network architecture shown in Fig. 2 and table S2 was carefully chosen after comparing several variations. Motivated by the recent trend for “small receptive fields and deep layers,” the sizes of the receptive fields of the convolutional layers were chosen to be small (3 × 3), and thus, the total number of learnable parameters was relatively manageable (approximately 0.1 million). Because the architecture is substantially deep, we used the ReLU nonlinearity as the neuron model to avoid the vanishing gradient problem (). Note that we only used phase images and not noisy amplitude images (due to the transparency and small sizes of the single spores) as the inputs to the network. In addition to the traditional “weight decay” regularization (), several recent techniques were used to reduce overfitting. We used the “dropout” technique, a regularization method based on efficient ensemble learning (), for the last hidden layer with a dropout rate of 0.5. For further regularization and accelerated training speed, “batch normalization” was done at every interface between a convolutional layer and the following ReLU layer (). Because of the large number of learnable parameters, we used “data augmentation” that enlarged the training data set () by a factor of 128. This is done by generating the new labeled images by rotating the original images by random angles sampled from a zero-mean Gaussian distribution with an SD of 10° and by flipping the images with a probability of 0.5. During the training stage, the learnable parameters were updated toward the direction minimizing the loss function using the concept of error backpropagation (, ). We used cross-entropy loss based on softmax function, which quantifies the mismatch between the machine-predicted and true labels, as the objective function to be minimized. By calculating the partial derivatives of the loss function with respect to the elements of the synaptic weight tensors using the chain rule, we could update the parameters in a stochastic gradient descent (SGD) scheme (, ). The learning rule was conventional SGD assisted with a momentum of 0.5 (note that using recent learning rules instead could further improve the performance), and the training batch size was 1024. The weights were initialized from a zero-mean Gaussian distribution with layer-wise scaling based on the input sizes (). The biases were initialized with the constant 0. We used an equal learning rate for all layers, which was attenuated by a factor of 5 per five epochs (). The hyperparameters were selected by cross-validation; the grid searching process with the initial learning rate and weight decay regularization strength resulted in values of 0.05 and 0.0005, respectively. We used one graphics processing unit (GPU) (GeForce GTX 680, NVIDIA) and CUDA Toolkit 7.5 (NVIDIA), which increased the training speed typically by 5- to 10-fold. We note that it is possible with more computing resources to train multiple network models with different random initializations to compose a committee machine to further enhance the performance (). Finally, the identification performance was estimated using separate test images that were never shown during the training stage. The error bars in Figs. 3 and 4 represent the SD calculated from 10 classification models with different random initializations. The typical training time for 30 epochs (which were used for species prediction) was 25 min (multiclass identification with five Bacillus species, GTX 680). The typical time required for species prediction was less than 1 ms per cell (batch size of 445 cells). We found an approximately twofold increase of computing speed for GTX 980 Ti (NVIDIA) and expect more than a threefold increase when using the state-of-the-art GPUs such as GTX 1080 Ti (NVIDIA). The visualization of HoloConvNet codes was performed by the unsupervised dimensionality reduction technique t-SNE, which embeds high-dimensional data in a low-dimensional space while preserving the pairwise distances of the data points, implemented in MatLab (). The activation strengths of individual neurons at the last hidden layer by the test images were used as the raw variables. The parameters for the stochastic optimization for t-SNE were as follows: The perplexity was 30, and the dimension for initial principal components analysis was 30.
  36 in total

1.  Biological warfare detection.

Authors:  D R Walt; D R Franz
Journal:  Anal Chem       Date:  2000-12-01       Impact factor: 6.986

2.  Polarization holographic microscopy for extracting spatio-temporally resolved Jones matrix.

Authors:  Youngchan Kim; Joonwoo Jeong; Jaeduck Jang; Mahn Won Kim; Yongkeun Park
Journal:  Opt Express       Date:  2012-04-23       Impact factor: 3.894

3.  Three-dimensional identification of stem cells by computational holographic imaging.

Authors:  Inkyu Moon; Bahram Javidi
Journal:  J R Soc Interface       Date:  2007-04-22       Impact factor: 4.118

4.  Three-dimensional imaging and recognition of microorganism using single-exposure on-line (SEOL) digital holography.

Authors:  Bahram Javidi; Inkyu Moon; Seokwon Yeom; Edward Carapezza
Journal:  Opt Express       Date:  2005-06-13       Impact factor: 3.894

5.  Synthetic Fourier transform light scattering.

Authors:  Kyeoreh Lee; Hyeon-Don Kim; Kyoohyun Kim; Youngchan Kim; Timothy R Hillman; Bumki Min; Yongkeun Park
Journal:  Opt Express       Date:  2013-09-23       Impact factor: 3.894

6.  Difference between the spore sizes of Bacillus anthracis and other Bacillus species.

Authors:  M Carrera; R O Zandomeni; J Fitzgibbon; J-L Sagripanti
Journal:  J Appl Microbiol       Date:  2007-02       Impact factor: 3.772

7.  Rapid detection of an anthrax biomarker by surface-enhanced Raman spectroscopy.

Authors:  Xiaoyu Zhang; Matthew A Young; Olga Lyandres; Richard P Van Duyne
Journal:  J Am Chem Soc       Date:  2005-03-30       Impact factor: 15.419

8.  Using buoyant mass to measure the growth of single cells.

Authors:  Michel Godin; Francisco Feijó Delgado; Sungmin Son; William H Grover; Andrea K Bryan; Amit Tzur; Paul Jorgensen; Kris Payer; Alan D Grossman; Marc W Kirschner; Scott R Manalis
Journal:  Nat Methods       Date:  2010-04-11       Impact factor: 28.547

9.  Wet and dry density of Bacillus anthracis and other Bacillus species.

Authors:  M Carrera; R O Zandomeni; J-L Sagripanti
Journal:  J Appl Microbiol       Date:  2008-02-20       Impact factor: 3.772

10.  Automated Detection of P. falciparum Using Machine Learning Algorithms with Quantitative Phase Images of Unstained Cells.

Authors:  Han Sang Park; Matthew T Rinehart; Katelyn A Walzer; Jen-Tsan Ashley Chi; Adam Wax
Journal:  PLoS One       Date:  2016-09-16       Impact factor: 3.240

View more
  26 in total

1.  Reagent-Free and Rapid Assessment of T Cell Activation State Using Diffraction Phase Microscopy and Deep Learning.

Authors:  Sukrut Hemant Karandikar; Chi Zhang; Akilan Meiyappan; Ishan Barman; Christine Finck; Pramod Kumar Srivastava; Rishikesh Pandey
Journal:  Anal Chem       Date:  2019-02-22       Impact factor: 6.986

2.  Holotomography: Refractive Index as an Intrinsic Imaging Contrast for 3-D Label-Free Live Cell Imaging.

Authors:  Doyeon Kim; Sangyun Lee; Moosung Lee; Juntaek Oh; Su-A Yang; YongKeun Park
Journal:  Adv Exp Med Biol       Date:  2021       Impact factor: 2.622

3.  Noninvasive detection of macrophage activation with single-cell resolution through machine learning.

Authors:  Nicolas Pavillon; Alison J Hobro; Shizuo Akira; Nicholas I Smith
Journal:  Proc Natl Acad Sci U S A       Date:  2018-03-06       Impact factor: 11.205

4.  Quantitative assessment of cancer cell morphology and motility using telecentric digital holographic microscopy and machine learning.

Authors:  Van K Lam; Thanh C Nguyen; Byung M Chung; George Nehmetallah; Christopher B Raub
Journal:  Cytometry A       Date:  2017-12-28       Impact factor: 4.355

5.  Understanding the Research Landscape of Deep Learning in Biomedical Science: Scientometric Analysis.

Authors:  Seojin Nam; Donghun Kim; Woojin Jung; Yongjun Zhu
Journal:  J Med Internet Res       Date:  2022-04-22       Impact factor: 7.076

6.  Identification of non-activated lymphocytes using three-dimensional refractive index tomography and machine learning.

Authors:  Jonghee Yoon; YoungJu Jo; Min-Hyeok Kim; Kyoohyun Kim; SangYun Lee; Suk-Jo Kang; YongKeun Park
Journal:  Sci Rep       Date:  2017-07-27       Impact factor: 4.379

7.  PhaseStain: the digital staining of label-free quantitative phase microscopy images using deep learning.

Authors:  Yair Rivenson; Tairan Liu; Zhensong Wei; Yibo Zhang; Kevin de Haan; Aydogan Ozcan
Journal:  Light Sci Appl       Date:  2019-02-06       Impact factor: 17.782

8.  Phase recovery and holographic image reconstruction using deep learning in neural networks.

Authors:  Yair Rivenson; Yibo Zhang; Harun Günaydın; Da Teng; Aydogan Ozcan
Journal:  Light Sci Appl       Date:  2018-02-23       Impact factor: 17.782

Review 9.  Combining Three-Dimensional Quantitative Phase Imaging and Fluorescence Microscopy for the Study of Cell Pathophysiology.

Authors:  Young Seo Kim; SangYun Lee; JaeHwang Jung; Seungwoo Shin; He-Gwon Choi; Guang-Ho Cha; Weisun Park; Sumin Lee; YongKeun Park
Journal:  Yale J Biol Med       Date:  2018-09-21

10.  DeepScreen: An Accurate, Rapid, and Anti-Interference Screening Approach for Nanoformulated Medication by Deep Learning.

Authors:  Yanjing Zhu; Ruiqi Huang; Rui Zhu; Wei Xu; Rongrong Zhu; Liming Cheng
Journal:  Adv Sci (Weinh)       Date:  2018-07-23       Impact factor: 16.806

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.