| Literature DB >> 31426441 |
Darius Dirvanauskas1, Rytis Maskeliūnas1, Vidas Raudonis2, Robertas Damaševičius3,4, Rafal Scherer5.
Abstract
We propose a method for generating the synthetic images of human embryo cells that could later be used for classification, analysis, and training, thus resulting in the creation of new synthetic image datasets for research areas lacking real-world data. Our focus was not only to generate the generic image of a cell such, but to make sure that it has all necessary attributes of a real cell image to provide a fully realistic synthetic version. We use human embryo images obtained during cell development processes for training a deep neural network (DNN). The proposed algorithm used generative adversarial network (GAN) to generate one-, two-, and four-cell stage images. We achieved a misclassification rate of 12.3% for the generated images, while the expert evaluation showed the true recognition rate (TRR) of 80.00% (for four-cell images), 86.8% (for two-cell images), and 96.2% (for one-cell images). Texture-based comparison using the Haralick features showed that there is no statistically (using the Student's t-test) significant (p < 0.01) differences between the real and synthetic embryo images except for the sum of variance (for one-cell and four-cell images), and variance and sum of average (for two-cell images) features. The obtained synthetic images can be later adapted to facilitate the development, training, and evaluation of new algorithms for embryo image processing tasks.Entities:
Keywords: deep learning; generative adversarial network; neural network; synthetic images
Year: 2019 PMID: 31426441 PMCID: PMC6720205 DOI: 10.3390/s19163578
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1A pipeline of a generative adversarial network (GAN).
Figure 2Architecture of the discriminator network.
Figure 3Architecture of the generator network.
Figure 4Sample images from Esco embryo image dataset.
Figure 5Example of final generated embryo cell images using the proposed algorithm. From left to right: One, two, and four cells. Images were filtered from “salt and pepper” noise using a median filter. See Table 1 for raw outputs.
Evolution of embryo cell images during GAN training.
| Epoch | 0 | 25,000 | 50,000 | 75,000 | 100,000 | 125,000 | 150,000 | 175,000 | 200,000 |
|---|---|---|---|---|---|---|---|---|---|
| one-cell |
|
|
|
|
|
|
|
|
|
| two-cells |
|
|
|
|
|
|
|
|
|
| four-cells |
|
|
|
|
|
|
|
|
|
Figure 6Loss log graph for the generator network.
Figure 7Loss log graph for the discriminator network.
Evaluation of the generated embryo image cells by human experts.
| Class | Test Images (Generated) | Good Images | Accuracy by Expert Selection |
|---|---|---|---|
| one-cell | 500 | 481 | 96.2% |
| two-cells | 500 | 434 | 86.8% |
| four-cells | 500 | 400 | 80.00% |
Figure 8Comparison of histograms of real vs. generated images (real images—blue, generated images—red): (a) one-cell images, (b) two-cell images, (c) four-cell images.
Evaluation of generated artificial embryo cell images using histogram comparison criteria.
| One-cell | Two-cells | Four-cells | |
|---|---|---|---|
| Correlation | 0.995 | 0.990 | 0.986 |
| Chi-square | 0.236 | 0.455 | 0.442 |
| Intersection | 1.883 | 1.849 | 1.873 |
| Bhattacharyya | 0.147 | 0.210 | 0.208 |
Results of principal component analysis (PCA) of Haralick features from original and synthetic embryo images.
| One-cell | Two-cells | Four-cells | |
|---|---|---|---|
| Explained variance of PC1 (original) | 99.90% | 99.94% | 99.86% |
| Explained variance of PC1 (synthetic) | 98.95% | 99.60% | 99.73% |
| Correlation between values of PC1 (original) and PC1 (synthetic) | 1.0000 | 0.9997 | 0.9999 |
Misclassification rate.
| Model | Images Analyzed | Overall Misclassification Rate |
|---|---|---|
| DGN [ | Digits 0 to 9 and combination (based on SVHN, and NORB sets) ( | 36.02% |
| Skip Deep Generative Model [ | Digits 0 to 9 and combination (based on SVHN, and NORB sets) ( | 16.61% |
| GAN (feature matching) [ | LSVRC2012 dataset with 1,000 categories ( | 8.11% |
| ALI [ | 25 types of shapes ( | 7.42% |
| WGAN [ | 6 classes of brain MR images ( | 50% |
| WGAN-GP [ | 8 classes of florescent microscopy images ( | 6.9% |
| HEMIGEN (our approach) | Once cell, Two cells, Four cells ( | 12.3% |