| Literature DB >> 31480740 |
Qian Liu1, Anna Junker2, Kazuhiro Murakami3, Pingzhao Hu4,5.
Abstract
High-content and high-throughput digital microscopes have generated large image sets in biological experiments and clinical practice. Automatic image analysis techniques, such as cell counting, are in high demand. Here, cell counting was treated as a regression problem using image features (phenotypes) extracted by deep learning models. Three deep convolutional neural network models were developed to regress image features to their cell counts in an end-to-end way. Theoretically, ensembling imaging phenotypes should have better representative ability than a single type of imaging phenotype. We implemented this idea by integrating two types of imaging phenotypes (dot density map and foreground mask) extracted by two autoencoders and regressing the ensembled imaging phenotypes to cell counts afterwards. Two publicly available datasets with synthetic microscopic images were used to train and test the proposed models. Root mean square error, mean absolute error, mean absolute percent error, and Pearson correlation were applied to evaluate the models' performance. The well-trained models were also applied to predict the cancer cell counts of real microscopic images acquired in a biological experiment to evaluate the roles of two colorectal-cancer-related genes. The proposed model by ensembling deep imaging features showed better performance in terms of smaller errors and larger correlations than those based on a single type of imaging feature. Overall, all models' predictions showed a high correlation with the true cell counts. The ensembling-based model integrated high-level imaging phenotypes to improve the estimation of cell counts from high-content and high-throughput microscopic images.Entities:
Keywords: autoencoder; automatic cell counting; deep learning; ensembling feature; microscopic imaging; transfer learning
Mesh:
Year: 2019 PMID: 31480740 PMCID: PMC6770845 DOI: 10.3390/cells8091019
Source DB: PubMed Journal: Cells ISSN: 2073-4409 Impact factor: 6.600
The two synthetic datasets.
| Datasets | Dataset Profiles | |||
|---|---|---|---|---|
| Size of Images | Resolution | Channel | ||
| Dot density | Raw data | 200 | 256 × 256 | 3 |
| Preprocessed data | 200 | 256 × 256 | 1 | |
| Foreground | Raw data | 1200 | 696 × 520 | 1 |
| Preprocessed data | 1200 | 256 × 256 | 1 | |
| Real data | Raw data | 385 * | 512 × 512 | 1 |
| Preprocessed data | 385 | 256 × 256 | 1 | |
* 164 of AKTP organoids (organoids with mutations in four cancer driver genes, APC∆716 mutation/KRasG12D mutation/TgfbrII knock-out/P53R270H mutation), 107 of AKTP-P2rX7 (AKTP organoids with P2rX7 gene overexpression) knock down organoids, and 114 of AKTP- Nt5e (AKTP organoids with Nt5e gene overexpression) knock down organoids.
Figure 1The workflow of the four deep convolutional neural network (DCNN) models used in this study. Gray arrows are trainable weights in the models. Colored arrows represent the weights inherited from previous well-trained feature extraction models. (A) Dot-density-map-based regression DCNN (DRDCNN). It contained two parts: feature extraction autoencoder (dashed box) and regression parts. The blue dashed arrows represent the well-trained weights, which can be loaded into the end-to-end counting regression model. (B) Ensembling-based regression DCNN (ERDCNN). Colored arrows represent the weights inherited from the two previous well-trained feature extraction models. (C) Foreground-mask-based regression DCNN (FRDCNN). It contained feature extraction autoencoder and regression parts as well. The orange arrows represent the well-trained weights which can be transferred into the ensemble regression model. (D) Density-only DCNN (DDCNN). This model first trained an autoencoder to extract a dot density map from a microscopic image, then summed the pixel values up in the dot density map to get the estimated cell count without any regression process.
The three regression DCNN models.
| DRDCNN | FRDCNN | ERDCNN | ||||||
|---|---|---|---|---|---|---|---|---|
| Autoencoder | Regression | Autoencoder | Regression | Autoencoder | Regression | |||
| Dataset 1 1 | Training set ( | Raw image |
|
|
|
|
|
|
| Feature label |
|
|
|
|
|
| ||
| Count label |
|
|
|
|
|
| ||
| Testing set ( | Raw image |
|
|
|
|
|
| |
| Feature label |
|
|
|
|
|
| ||
| Count label |
|
|
|
|
|
| ||
| Dataset 2 2 | Training set ( | Raw image |
|
|
|
|
|
|
| Feature label |
|
|
|
|
|
| ||
| Count label |
|
|
|
|
|
| ||
| Testing set ( | Raw image |
|
|
|
|
|
| |
| Feature label |
|
|
|
|
|
| ||
| Count label |
|
|
|
|
|
| ||
1 The dataset with 200 synthetic microscopic images and their corresponding dot density maps as well as the cell counts. Training set with 150 samples and testing set with 50 samples were randomly split. 2 The dataset with 1200 synthetic microscopic images and their corresponding foreground masks as well as the cell counts. Training set with 1150 samples and testing set with 50 samples were randomly split. 3 indicates that the set was used to train the certain part of the model. 4 indicates that the set was not used to train the certain part of the model.
Figure 2Features predicted by autoencoders of DRDCNN and FRDCNN. (A): Predicted dot density map by the autoencoder part of DRDCNN. The predicted one (in the middle) was a little brighter than the true dot density map but had the full ability to represent the center of each cell in the raw image. (B): Predicted foreground mask by the autoencoder part of FRDCNN. The predicted foreground mask was smoother than the true mask, but this small drawback did not influence its representative for the foreground of the raw image.
Performance of the four models on the 100 test images.
| DDCNN | DRDCNN | FRDCNN | ERDCNN | |
|---|---|---|---|---|
|
| 98.24 | 57.68 | 56.07 | 49.25 |
|
| 78.65 | 37.49 | 39.48 | 31.49 |
|
| 1.32 | 0.33 | 0.35 | 0.28 |
|
| 0.34 | 0.81 | 0.74 | 0.85 |
Figure 3Cell counts predicted by the four models vs. their true cell counts based on the 100 test images. Predictions made by the ERDCNN model were closer to the true cell counts; especially for those images with smaller cell counts (e.g., the number of cells is smaller than 100).
Results of the t-tests evaluating the predicted cell counts under the three different conditions using the different DCNN models.
| DDCNN | DRDCNN | FRDCNN | ERDCNN | ||
|---|---|---|---|---|---|
| AKTP-P2rx7 vs. AKTP | 0.0009 | 0.84 | 0.002 | 0.0002 | |
| Mean (x,y) | 5.13, 6.66 | 16.70, 16.93 | 44.89, 53.49 | 24.26, 29.69 | |
| AKTP-Nt5e vs. AKTP |
| 0.27 | 0.52 | 0.94 | |
| Mean (x,y) | 6.67, 5.99 | 17.81, 16.93 | 55.73, 53.49 | 29.80, 29.69 | |
Figure 4Density plots of the predicted cell counts for the three experimental groups from the four models. (A) Each plot compared the predictions of the three experimental groups using the same model. (B) Each plot compared the predictions of the four models for the same experimental group.