Automated image processing approaches are indispensable for many biomedical experiments and help to cope with the increasing amount of microscopy image data in a fast and reproducible way. Especially state-of-the-art deep learning-based approaches most often require large amounts of annotated training data to produce accurate and generalist outputs, but they are often compromised by the general lack of those annotated data sets. In this work, we propose how conditional generative adversarial networks can be utilized to generate realistic image data for 3D fluorescence microscopy from annotation masks of 3D cellular structures. In combination with mask simulation approaches, we demonstrate the generation of fully-annotated 3D microscopy data sets that we make publicly available for training or benchmarking. An additional positional conditioning of the cellular structures enables the reconstruction of position-dependent intensity characteristics and allows to generate image data of different quality levels. A patch-wise working principle and a subsequent full-size reassemble strategy is used to generate image data of arbitrary size and different organisms. We present this as a proof-of-concept for the automated generation of fully-annotated training data sets requiring only a minimum of manual interaction to alleviate the need of manual annotations.
Automated image processing approaches are indispensable for many biomedical experiments and help to cope with the increasing amount of microscopy image data in a fast and reproducible way. Especially state-of-the-art deep learning-based approaches most often require large amounts of annotated training data to produce accurate and generalist outputs, but they are often compromised by the general lack of those annotated data sets. In this work, we propose how conditional generative adversarial networks can be utilized to generate realistic image data for 3D fluorescence microscopy from annotation masks of 3D cellular structures. In combination with mask simulation approaches, we demonstrate the generation of fully-annotated 3D microscopy data sets that we make publicly available for training or benchmarking. An additional positional conditioning of the cellular structures enables the reconstruction of position-dependent intensity characteristics and allows to generate image data of different quality levels. A patch-wise working principle and a subsequent full-size reassemble strategy is used to generate image data of arbitrary size and different organisms. We present this as a proof-of-concept for the automated generation of fully-annotated training data sets requiring only a minimum of manual interaction to alleviate the need of manual annotations.
Current developments of fluorescence microscopy imaging techniques allow to acquire vast amounts of image data, capturing different cellular structures for various kinds of biological experiments in 3D and over time [1, 2]. The growing amount of data that needs to be analyzed, demands an automation of image processing tasks [3]. To this end, automated approaches became widely used and especially machine learning and deep learning-based approaches offer powerful tools for detection, segmentation, and tracking of different cellular structures [4, 5]. However, learning-based approaches need large annotated data sets to become robust and generalist. On the other hand, the creation of manually annotated data sets is very time-consuming and tedious, causing those data sets to be rarely available. Although there are many classical and machine learning-assisted annotation tools accessible [6-9], which reduce the annotation time for biological experts, especially the dense annotation of 3D data remains difficult. Current approaches propose to reduce annotation efforts and increase generalizability of machine learning-based approaches by collecting a manifold of annotated image data from slightly different domains, creating a highly diverse training data set [10]. This supports the claim that a large amount of annotated training data is a key factor in creating robust approaches.To reduce the need of manual annotations, simulation tools have been proposed, which rely on known imaging and microscopy parameters [11-13]. Those techniques can be used to automatically create annotated data samples, but they rely on precise prior knowledge about each specific experimental setup. With the recent success of generative adversarial networks (GANs) [14] and their progressive developments, simulation approaches became more realistic, less parameter dependent and, thus, more applicable. Techniques have been proposed to directly synthesize 2D cellular structures [15-17] or simulate 3D cellular structures [18], which in turn can be used to synthesize realistic 3D images with existing classical approaches [19] or GAN-based approaches [20].In this paper we demonstrate how a conditional generative adversarial network [21] can be used to transform large realistic 3D annotation masks of multi-cellular structures into realistic 3D microscopy image data in a patch-based manner, by employing a patch merging strategy to obtain seamless results. Furthermore, the proposed conditioning leverages the synthesis of different image quality levels to further increase diversity of simulated image data. We introduce different approaches for the simulation of these 3D annotation masks, which can be chosen with respect to the complexity of cellular structures and the availability of manually obtained annotation masks. As a contribution to the community, two fully-annotated 3D synthetic data sets are made publicly available to potentially serve as training or benchmark data sets for 3D detection and segmentation approaches for fluorescently labeled nuclei and membranes.
2 Generation of cellular 3D structures
The generation of realistic annotation masks of cellular structures is crucial for the synthesis of realistic image data, since unrealistic and overly artificial structures can impede structural correlation between annotation masks and synthetic images, as demonstrated in [15]. If features of cellular structures in the annotation masks deviate too much from those in the image data , the generator network G can no longer find a suitable mapping and starts to deform annotated structures to resemble those in the image domain, which would cause the annotated structures to no longer match the structures in the synthetic image data. Therefore, it is crucial to use simulation techniques that reproduce structures as realistically as possible.Realistic annotation masks can be obtained in several ways, depending on the availability of suitable segmentation approaches, time for manual expert annotations and complexity of cellular structures. Besides manual annotation or automated segmentation, we experimented with a few modelling approaches that can be used for the simulation of both organism shape and cell structures. The discussed simulation is divided into three parts (Fig 1): first, the generation of the foreground region outlining an organism shape, second, determining positions of cellular structures within the foreground region and finally, modelling of cellular structures at each given position. Ultimately, cells are either represented as single nuclei or as cellular membranes, i.e., as a structural mesh within the foreground region.
Fig 1
Qualitative 2D illustration of the 3D mask simulation pipeline.
The simulation process generates an annotation mask in three refinement steps, starting with organism shape generation, adding cell positions within the foreground region and final nuclei or membranes structure generation at each respective positions.
Qualitative 2D illustration of the 3D mask simulation pipeline.
The simulation process generates an annotation mask in three refinement steps, starting with organism shape generation, adding cell positions within the foreground region and final nuclei or membranes structure generation at each respective positions.
2.1 Geometrical modelling
Modelling cellular structures with geometrical functions is the least complex simulation technique discussed in this paper and, consequently, offers a limited set of shapes that can be represented. Nevertheless, it offers a straightforward possibility to model cellular structures.For cell positioning we follow a layer-wise assembly, starting to add layers of cells at the outer boundary of the organism and we progressively add additional layers inwards, if reasonable. The additional layer locations are determined by shrinking the organism shape and randomly placing new cells along the determined boundary, with random small offsets to add further realism. Cell distribution and density in each layer and layer thickness are adjusted using prior knowledge about cell sizes to avoid unnatural shape variations in the 3D space. For the simulation of cellular membranes, each foreground voxel x is assigned to the closest cell position x, creating a Voronoi-like tessellation of the foreground region. Since the resulting segments show unnaturally straight edges, an additional weighting is applied to the distance calculation, which is adapted from [22] and bends the planes between different segments, resulting in the distance metric
with weight γ being fixed for cell position x. The final tessellation is used to either create annotation masks for cell membranes by considering planes between segments or instance segmentations of whole cells by considering entire segments.
2.2 Statistical shape models
If a set of annotations is already available, characteristic shape parameters can be extracted and used to generate additional data. Statistical shape models offer a way to determine distinctive shape modes and encode them in an accessible way [23] to generate shapes by changing only a desirable small number of parameters. As a prerequisite, each voxelized shape needs to be sampled at predefined angular rays starting at the organism centroid or cell centroid and it needs to be converted into a list of 3D boundary coordinates p = (x1, y1, z1, …, x, y, z). From coordinate lists of all shapes, the mean shape and the covariance matrix are computed. To identify distinctive modes, an eigenvalue decomposition of the covariance matrix Σ is computed, resulting in eigenvectors and corresponding eigenvalues λ, while modes are ordered so that λ1 ≥ λ2 ≥ … ≥ λ. Accordingly, the first eigenvectors encode most of the shape variance and account for the most distinctive modes. New shapes p are generated by altering the mean shape by a linear combination of the first n modes, while n ≤ N can be chosen as a required level of detail:
Influence of each mode is defined by a weight vector b, which is randomly sampled from but restricted to the limit b ∈ [−3λ, 3λ] to only allow the generation of shapes that lie within the determined variance range. The obtained list of boundary points is transformed to a voxelized mask representation using a Delaunay triangulation. Possible shape variations learned from a public data set [24] are shown in Fig 2.
Fig 2
Overview of possible spherical shapes initialized with statistical shape models.
Different shape variations are visualized, with statistics learned from a public data set [24].
Overview of possible spherical shapes initialized with statistical shape models.
Different shape variations are visualized, with statistics learned from a public data set [24].
2.3 Spherical harmonics
Spherical harmonics (SH) offer another way of encoding shapes in a compact format [25], facilitating the generation of new shapes by changing a small number of parameters. This concept is related to the Fourier transform, but is inherently suitable for spherical shapes. The generation of shapes with spherical harmonics is accomplished by defining each shape as a composition of a predefined set of R weighted orthonormal basis functions, which can be identified by an order l describing the level of high frequency shape components and a degree m describing a particular variant of an order. The resulting set of functions is organized in a pyramid scheme with l ≥ 0 and −l ≤ m ≤ l and each spherical harmonics basis function can be computed by
with describing the Legendre polynomials of degree m and order l and parameters θ and ϕ denoting the spherical angular sampling coordinates. An individual 3D shape S is defined as a weighted linear combination of the spherical harmonic basis functions
with weight c specifying the influence of the corresponding spherical harmonic basis functions Y. Since the first basis function (l = m = 0) already represents a perfect sphere scalable by a single parameter c1, spherical harmonics are well-suited to describe and model cellular structures [26-28]. Note that SH representations are limited to star-convex shapes, which we, however, found sufficient for simulation of nuclei and various early stage embryo shapes. Since the set of basis functions is already formulated, the generation of shapes reduces to choosing the number R of desired spherical harmonics and determining corresponding weights c1,…,. R is chosen by specifying the desired level of detail, i.e., the order of high frequency shape components l, since R = (l + 1)2 when using all available degrees m. Weights c1,…, are designed to have an exponentially decreasing magnitude to approximate smooth shapes, while further diversity is added by initializing each weight by a random value drawn from a standard normal distribution . This results in
with r defining the approximate radius of the spherical shape, w(l, m) constituting the random initialization assigned to the coefficient for degree m and order l and γ controlling the smoothness of the resulting shape. Examples for different γ are shown in Fig 3.
Fig 3
Overview of possible spherical shapes initialized with spherical harmonics.
Shapes are shown for different values of γ and colors indicate a positive (red) or negative (blue) deviation from a perfect sphere.
Overview of possible spherical shapes initialized with spherical harmonics.
Shapes are shown for different values of γ and colors indicate a positive (red) or negative (blue) deviation from a perfect sphere.
3 Synthesis of 3D microscopy data
The proposed synthesis pipeline is designed to transform binary annotations of cellular structures in 3D image data into realistic microscopy image data. Generating realistic image data from annotation masks offers several advantages, since it allows to control the outline and geometry of structures that should be generated and it allows to generate image data sets that inherently include corresponding annotations, independent from image quality or degradation.For the synthesis of realistic 3D microscopy image data, a generative adversarial network (GAN) [21] is used, which allows for an unsupervised translation between mask annotations and microscopy image data . This concept includes two individual networks (Fig 4): a generative network G, which is trained to learn a mapping G(m) from a mask into an image , and a discriminator network D, which is trained to distinguish between real and generated image data. Consequently, the training procedure is an adversarial assembly, where the generator wants to outperform the discriminator by generating realistic images, while the discriminator aims to be as good as possible in distinguishing real from fake images. In principle, the generator is trained to simulate the image formation process of a microscope without the necessity of prior knowledge about the acquisition process parameters.
Fig 4
Schematic of the conditional generative adversarial network.
The generator transforms annotation masks m into realistic images , which are assessed by the discriminator using realistic image data . The positional conditioning is used to generate and assess positional image characteristics. Note that, for simplicity, visualizations are provided in 2D, despite processing being done in 3D.
Schematic of the conditional generative adversarial network.
The generator transforms annotation masks m into realistic images , which are assessed by the discriminator using realistic image data . The positional conditioning is used to generate and assess positional image characteristics. Note that, for simplicity, visualizations are provided in 2D, despite processing being done in 3D.For the generator network a 3D U-Net [29] is used and the pixel-shuffle technique [30] is incorporated for higher quality upsampling in the decoding path. The generator training objective consists of two separate losses, including the adversarial loss as formulated in [31] to assess the generated image quality and an identity loss to support the generation of realistic intensity characteristics and impose a structural correlation between the input and output. A Patch-GAN is used as discriminator network and trained with adversarial losses for real and fake images, as proposed in [31].Due to large data sizes of 3D microscopy image data and memory limitations of current graphics processing units (GPUs), processing 3D images as a whole is not feasible, demanding a patch-based processing. Processing patches, however, causes two sources of errors, including tiling artifacts when reassembling the full-size image and loss of global positional information for structures within each patch, which is very crucial in generating realistic image degradation and positional intensity characteristics. Firstly, tiling artifacts are avoided by overlapping neighbouring patches by a fixed margin d and averaging the overlap regions based on a concentric weight map, decreasing the influence of voxels based on their distance to the patch center, as done in [26]. Additionally, to avoid potential patch border artifacts, output patches are cropped by a fixed margin d before full-size assembly. Secondly, global positional information of patches are maintained by providing an additional conditioning as input, encoding the position of each voxel in relation to the specimen’s boundary, which is motivated by the occurrence of depth-dependent intensity decay in microscopy imaging. We choose a hyperbolic tangent function to further indicate if a voxel lies within a specimen (foreground) or in a background region to enable the generation of more detailed noise features. To consider different influences of distances within the foreground and background regions, respectively, the positional function is adjustable for both regions individually, leading to the following formulation:
where x is the 3D position vector of a voxel and dist(x) measures the smallest distance to the specimen’s boundary. Scaling parameters α and β are used to adjust the saturation of the hyperbolic tangent to realistically simulate the limited penetration depth of the imaging system. Applying this function to each voxel, an encoded distance map is created, which matches the spatial dimension of the corresponding annotation mask and is used as conditional input for both, the generator and discriminator network (Fig 4).During training with small data sets, we observed the generator to preferably generate specific intensity and noise patterns, resulting in a slight overfitting to the training samples. In order to counteract this problem, augmentation strategies could be employed to further increase diversity of the image training data, which, however, needs to be configured carefully to create as much diversity as possible, but to include as little alterations as possible to still enable a smooth and realistic generator training. We follow the concept proposed in [32], stating that overfitting to small data sets can be avoided by utilizing an adaptive discriminator augmentation (ADA), i.e., augmentation of real and generated images before they are being assessed by the discriminator. This helps to include as much diversity as possible by simultaneously controlling the impact of those alterations. To establish this control, a set of augmentations is predefined and each individual transformation is sequentially applied to an image with probability p. To automatically identify the optimal value for p, an overfitting measure is determined, which calculates the fraction of real image samples that are correctly classified by the discriminator, leading to the following heuristic:
As proposed in [32], a target value of r = 0.6 is chosen, indicating that the augmentation probability p should increase for larger values and decrease for smaller values by a step size of δ every e epochs. To enhance robustness of the training procedure, we use rectified Adam [33] as an optimizer for all networks.
3.1 Putting it all together
Our data generation pipeline consists of two major parts (Fig 5). First, the simulation or acquisition of annotations of cellular structures, which either represent nuclei or cellular membranes. The generation of annotations has different stages, starting with organism shape generation, cell positioning and final generation of cellular structures, while each simulation stage can be implemented by manual interaction or automated detection and segmentation approaches. Second, the final annotations are used to generate image data with the proposed GAN training and they are further used to generate corresponding instance segmentations, allowing to create fully-annotated data sets for further utilization.
Fig 5
Qualitative 2D overview of our 3D data generation pipeline.
It comprises the acquisition or simulation of annotations and the image synthesis. Annotations can be obtained from manual, automated, and simulation approaches, and final cellular annotations are used to generate corresponding instance segmentations.
Qualitative 2D overview of our 3D data generation pipeline.
It comprises the acquisition or simulation of annotations and the image synthesis. Annotations can be obtained from manual, automated, and simulation approaches, and final cellular annotations are used to generate corresponding instance segmentations.
4 Experiments and quality assessment
Two publicly available data sets are used for the evaluation of our approach:Data set
The first data set includes 125 3D image stacks showing fluorescently labeled cell membranes in six different Arabidopsis thaliana shoot apical meristem that were imaged with confocal microscopy at a resolution of 0.22 × 0.22 × 0.26 μm3 per voxel using a 63×/1.0 N.A. water immersion objective [24]. Additionally, 3D instance segmentations are available for all image stacks, which were automatically obtained and partly manually corrected by the authors. All images have a spatial size ranging between 326 × 367 × 107 and 512 × 512 × 396 voxel.Data set
The second data set consists of a total of 394 3D image stacks showing fluorescently labeled cell membranes and nuclei in two different Danio rerio (DR1 and DR2) [34]. Image stacks were captured with multi-photon microscopy at a resolution of 1.37 × 1.37 × 1.37 μm3 per voxel [34]. No ground truth annotations are available and all images have a spatial size of 512 × 512 × z voxel, while z is ranging from 104 to 120.Ideally, a sufficiently large fully-annotated data set is available for the training of the proposed image synthesis pipeline, as cellular structures have to be as similar as possible to the structures in the real data. However, due to the lack of fully-annotated 3D image data sets, a more realistic scenario would consider automatically-annotated image stacks or desirably even fully-simulated annotations, where no tedious and time-consuming manual annotation is required. To evaluate different aspects of our simulation pipeline and influences of different annotation acquisition techniques, we conduct multiple experiments ranging from an ideal case scenario to more practical scenarios, as described in the following.For all experiments in this section real image data is used as the target domain and annotation masks from different sources are used as source domain, i.e., as input to the generative adversarial network. Training is performed for 5000 epochs using a patch size of 128 × 128 × 64 voxel and distance scalings of α = β = 100 (Eq 6). During each epoch only one randomly located patch is extracted from each of the images to increase data diversity during training. Furthermore, the set of random augmentations available for the adaptive discriminator augmentation includes linear intensity scaling in the range [0.6, 1.2], additive Gaussian noise sampled from , voxel shuffling in a randomly located region of size 25 × 25 × 25 voxel, randomly located inpaintings of size 15 × 15 × 15 voxel and linear intensity reduction along a random dimension. Augmentation updates are performed every e = 1 epochs with a step size of δ = 0.05 to allow adjusting the augmentation parameters in a reasonable number of epochs. For resembling of the full-size image, an overlap and crop of d = d = (30, 30, 15) are used.
4.1 Image synthesis with manually annotated data
As a first experiment, the 3D instance segmentation masks of the data set are used as input for the generative adversarial network, which serves as a baseline to assess the capability of the proposed synthesis pipeline. The six different specimens of this data set are divided into training and test sets, using plants 1, 2, 4 and 13 for training and plants 15 and 18 for testing.Qualitative results are illustrated in Fig 6 (left column pair), comparing different views of real and synthetic image data. As further qualitative measures the intensity profile for a real and the corresponding synthetic center xz- and yz-slice is generated, integrated over the z dimension and plotted along the x and y dimension, respectively (Fig 6, Membrane). Additionally, intensity spectra are created for a xy-slice of the real and synthetic image (Fig 6, Membrane). Both illustrations demonstrate the similarity between real and synthetic image data in the spatial and in the spectral domain. A large portion of the images contains noise, which is non-deterministic and can vary between images without impairing the image quality. Furthermore, especially since the proposed concept operates in an unsupervised fashion, the trained networks tend to optimize towards a consensus of intensity patterns within the training data set, causing the resulting images to contain less prominent or differently located intensity peaks by still realistically reconstructing the given structures. These aspects cause a visual discrepancy between high-frequency areas in the spectra and peaks in the intensity profiles.
Fig 6
Synthetic image data.
Different views of real image data (blue columns) in comparison to synthetic image data (red columns) generated by our GAN approach. Examples are shown for different experiments using manually corrected masks from the data set [24] (Membrane), and automatically annotated masks for membranes (Membrane) and nuclei (Nuclei) from the data set [34]. Additionally, the spectra and intensity profiles of different slices are shown as qualitative metrics.
Synthetic image data.
Different views of real image data (blue columns) in comparison to synthetic image data (red columns) generated by our GAN approach. Examples are shown for different experiments using manually corrected masks from the data set [24] (Membrane), and automatically annotated masks for membranes (Membrane) and nuclei (Nuclei) from the data set [34]. Additionally, the spectra and intensity profiles of different slices are shown as qualitative metrics.As quantitative quality metrics the normalized root mean square error (NRMSE), structural similarity index measure (SSIM) and the zero mean normalized cross-correlation (ZNCC) are used (Table 1, Membrane). By considering that a large portion of the image is background or random noise and only a small portion of the image contain cellular structures, these scores prove the synthetic data to be realistic, supporting the impression from the qualitative assessment.
Table 1
Quantitative assessment of image quality.
Quality scores obtained for the different synthetic data sets.
NRMSE
SSIM
ZNCC
Membranemanual
0.130
0.658
0.739
Membraneautomatic
0.123
0.584
0.847
Nucleiautomatic
0.119
0.617
0.751
Nucleisimulated
0.129
0.706
0.632
Quantitative assessment of image quality.
Quality scores obtained for the different synthetic data sets.
4.2 Image synthesis with automatically obtained data
In a second experiment, annotations of cellular structures are automatically obtained for the data set. For cellular membranes, we used the approach described in [35], which is a 3D extension of the Cellpose approach [10] using a 3D convolutional neural network to represent cell shapes as generalist gradient maps. An iterative post-processing technique is used to reconstruct the individual cell instances. Training of the neural network is performed for 1000 epochs using the manually annotated training data set and the model is applied to images of the image data. For cell nuclei we used the approach proposed in [36], which makes use of a Laplacian of Gaussian (LoG) scale-space responding to a predefined range of cell sizes. The shape-sensitive output is used to determine cell locations and to facilitate a watershed-based segmentation of cell nuclei. Parameters for this segmentation approach are manually tweaked to produce good results.Since there is no ground truth segmentation data available for the data set, segmentation approaches serve as a straightforward way to create annotations that represent the shapes and distribution of cellular structures as realistically as possible. Note that no further post-processing or manual corrections are performed to represent a realistic scenario with minimum manual interaction. The first specimen of the data set (DR1) was used for training of the GAN and the second specimen (DR2) was used for testing.Qualitative results for the synthetic membrane data and further qualitative measures, including the intensity profile of center xz- and yz-slices and intensity spectra of xy-slices are illustrated in Fig 6 (Membrane). Results for the synthetic nuclei data are illustrated in Fig 6 (Nuclei).By keeping in mind, that the precision of the automatically obtained annotation masks can not be assessed, again NRMSE, SSIM and ZNCC are used as quantitative quality metrics (Table 1, Membrane and Nuclei), which are similar to the scores obtained in the previous experiment . Qualitative and quantitative measures and figures again demonstrate success in generation of realistic 3D microscopy image data from unrefined, automatically generated annotation masks.
4.3 Image synthesis on simulated data
For training of deep learning-based synthesis approaches it is important to use annotations of cellular structures that are as similar as possible to structures in the real image data to prevent the generation of image artifacts and structural misalignment [15]. In this experiment, we further assessed if the application of a trained model requires the same level of structural correlation. To this end, we simulated annotation masks for both, cellular membranes and cell nuclei from very sparse annotations and used the models from and to synthesize real microscopy image data.As a first scenario, only automatically obtained point annotations from the data set as described in are considered and, subsequently, cell nuclei are modelled at each position utilizing spherical harmonics. Spherical harmonics coefficients are randomly constructed using a smoothness factor γ = 5 and approximate radii pixel. For each of the 200 raw image stacks of the test set (DR2), one corresponding mask is simulated and the synthesized image data is generated using the trained nuclei model from (Fig 7, top). Since positional correspondences are maintained and only structural characteristics differ between real and synthetic image data, again NRMSE, SSIM and ZNCC are used as quantitative quality metrics (Table 1, Nuclei). The obtained scores are similar to the scores obtained from the previous experiments and allow us to again conclude success in the generation of realistic 3D image data.
Fig 7
Fully synthetic image data.
Different views of synthetic nuclei image data (top) and synthetic membrane image data (bottom) generated by our GAN approach using the simulated annotation masks.
Fully synthetic image data.
Different views of synthetic nuclei image data (top) and synthetic membrane image data (bottom) generated by our GAN approach using the simulated annotation masks.As a second scenario, statistical shape models are utilized to generate organism shapes similar to the specimen of the data set and, subsequently, cellular structures are simulated within the foreground region by using geometrical modelling. Corresponding synthesized image data is obtained with the trained model from (Fig 7, bottom). Accurate shape model statistics are derived from the foreground segmentation of the corresponding manually obtained annotation masks, although foreground segmentation has also been shown to be automatable, e.g., by automated 3D interpolation techniques using very sparse manually drawn 2D outlines [37]. This scenario renders pixel-level metrics to be no longer feasible due to the missing direct structural correspondences between real and synthetic image samples. However, since the data is ultimately generated to alleviate the need of manually annotated data for learning-based approaches, we assess how accurate a 3D segmentation approach [38] performs on the test data set, when trained on synthetic image data. This multi-class segmentation approach predicts probability maps of background, membrane structures, and centroids, which helps to demonstrate how the segmentation of different structures is affected by the image synthesis. Finally, a watershed-based post-processing technique is used to reconstruct individual cell instances. Therefore, 125 simulated masks are generated and split into 82 for training and 43 for testing to match the quantity of image data from the real data set (Fig 7, right). Subsequently, the segmentation approach is trained on the synthetic data for 1000 epochs using a patch size of 128 × 128 × 64 voxel and it is applied to both, the test data of the synthetic data set (Syn2Syn) and the test data of the data set (Syn2Real), including plants 15 and 18. Additionally, the segmentation approach is trained on the train split of the original data set and applied to the real (Real2Real) and the synthetic (Real2Syn) test split. Furthermore, a mixed data set is constructed by combining both real and synthetic data sets to further increase data diversity. The model trained on the combined training set is tested on real (Mix2Real) and synthetic (Mix2Syn) test data. The probability map of each class, including background, cellular membrane and cell centroids, is thresholded by an individual value t = 0.1, t = 0.4 and t = 0.2 and, afterwards, an average intersection over union (IoU) is computed as a metric for each of the three classes (Fig 8, top). To obtain representative results, we allow predictions to be within a small range around the ground truth masks, which accommodates the small sizes of cellular structures and centroids. The allowed misplacement distances are 1 voxel for the background and cellular membrane classes and 5 voxel for the centroid class. Additionally, instance segmentations are obtained from the multi-class approach and another robust instance segmentation approach [35] and resulting average intersection over union (IoU) scores are plotted in Fig 8 (bottom). Overall high segmentation scores support that the generated data is realistic. However, due to differences in structure and intensity characteristics, the transfer training setups (Real2Syn and Syn2Real) show a decrease in segmentation scores, which is more severe when training on synthetic and testing on real data. Since centroids represent the class with the smallest object sizes, the highest difference in obtained segmentation scores are observed for this class. Nevertheless, the obtained segmentation scores when training on synthetic and testing on real data motivate the utilization of synthetic microscopy image data as a training data set for segmentation approaches, especially considering that the synthetic training data can be obtained with a minimum of manual interaction. Mixed training setups (Mix2Real and Mix2Syn) further motivate the incorporation of synthetic image data to diversify the training data set and improve results. Due to the controlled generation of synthetic image data, the resulting fully-annotated data sets don’t contain large segmentation inaccuracies, which accounts for the score difference between the pure training cases (Real2Real and Syn2Syn).
Fig 8
Segmentation scores obtained for different data setups.
Multi-class segmentation scores obtained with the approach from [38] (top) and instance segmentation scores obtained with the approaches from [38, 35] (bottom) trained on real data (Real2Real, Real2Syn) and synthetic data (Syn2Real, Syn2Syn) and a mixed data set containing real and synthetic data (Mix2Real, Mix2Syn). Trained models are applied to real and synthetic data, respectively.
Segmentation scores obtained for different data setups.
Multi-class segmentation scores obtained with the approach from [38] (top) and instance segmentation scores obtained with the approaches from [38, 35] (bottom) trained on real data (Real2Real, Real2Syn) and synthetic data (Syn2Real, Syn2Syn) and a mixed data set containing real and synthetic data (Mix2Real, Mix2Syn). Trained models are applied to real and synthetic data, respectively.
4.4 Image synthesis on different quality levels
The proposed positional conditioning of the generative adversarial network allows to control the quality level of the generated data by altering the positional foreground parameter α (Eq 6). This property is investigated by using different scalings α ∈ {10, 50, 100, 500, 1000} to generate different quality levels of the test set (plants 15 and 18) using the generative network from experiment . Image slices of three different quality levels are depicted in Fig 9 (top), alongside the integrated intensity profile of the center xz-slices (bottom), with both of these qualitative illustrations showing the decaying intensity for poor quality levels.
Fig 9
Qualitative assessment of altered image quality.
2D slices of 3D synthetic image data generated by our GAN approach using the same manually corrected mask from [24], with different foreground distance scalings α (top). For the center xz-slice, intensities are integrated over the z dimension and plotted along the x dimension (bottom).
Qualitative assessment of altered image quality.
2D slices of 3D synthetic image data generated by our GAN approach using the same manually corrected mask from [24], with different foreground distance scalings α (top). For the center xz-slice, intensities are integrated over the z dimension and plotted along the x dimension (bottom).To evaluate the quality of the generated data, we make use of the structural correspondence to the real image data and compute NRMSE, SSIM, ZNCC and PSNR as metrics (Table 2). The obtained scores slightly deteriorate for both, worse and improved image quality, which is caused by the changes in structural quality and noise content. Since these changes are to be expected and the scores remain in the same value range, the claim of generating realistic image data holds.
Table 2
Quantitative assessment of altered image quality.
Quality scores obtained for synthetic image data generated on different quality levels by varying α from Eq 6, including the normalized root mean square error (NRMSE), structural similarity index measure (SSIM), zero mean normalized cross-correlation (ZNCC) and peak signal-to-noise ratio (PSNR).
α = 10
α = 50
α = 100
α = 500
α = 1000
NRMSE
0.134
0.130
0.130
0.136
0.140
SSIM
0.640
0.654
0.658
0.660
0.643
ZNCC
0.704
0.735
0.739
0.721
0.703
PSNR
17.485
17.802
17.768
17.376
17.079
Quantitative assessment of altered image quality.
Quality scores obtained for synthetic image data generated on different quality levels by varying α from Eq 6, including the normalized root mean square error (NRMSE), structural similarity index measure (SSIM), zero mean normalized cross-correlation (ZNCC) and peak signal-to-noise ratio (PSNR).Additionally, we obtain multi-class segmentations for the real and the synthetic test image data on all quality levels using the approach proposed in [38] and consult these accuracies as a further quality metric. Again, class probability maps are thresholded by an individual value t = 0.1, t = 0.4 and t = 0.2 and, afterwards, an aggregated intersection over union (IoU) is computed as a metric for each of the three classes (Fig 10, top). Due to the small sizes of membrane structures and centroids, small displacements would have a large, hardly interpretable impact on the obtained segmentation accuracies and, therefore, both classes are allowed to lie within a small distance to the ground truth of 1 voxel for membrane structures and 5 voxel for centroids. Instance segmentation results for both segmentation approaches [35, 38] are evaluated by computing average intersections over union (IoU) scores (Fig 10, bottom).
Fig 10
Segmentation scores obtained on altered image quality.
Multi-class segmentation scores [15] (top) and instance segmentation scores [15, 35] (bottom) obtained for synthetic image data generated on different quality levels by varying α from Eq 6. Dashed lines represent results obtained on the original test data.
Segmentation scores obtained on altered image quality.
Multi-class segmentation scores [15] (top) and instance segmentation scores [15, 35] (bottom) obtained for synthetic image data generated on different quality levels by varying α from Eq 6. Dashed lines represent results obtained on the original test data.The baseline (Fig 10, dashed line) is computed from segmentation scores obtained on the real test set. Segmentation scores for cellular membranes increase with increasing image quality until nearly reaching the baseline score, demonstrating the progressively increasing quality of cellular structures. Background accuracy decreases with increasing image quality, which is caused by the diminishing noise content within the foreground region, causing the segmentation approach to confuse regions within cells with background regions. Additionally, we observed harsh and unnatural intensity transitions for the highest quality level α = 1000, which causes further segmentation inaccuracies and points out an upper boundary for a natural increase of the image quality. Although these facts produce segmentation artifacts, overall background segmentation scores stay close to the baseline scores and promote the claim of generating increasing image quality. This problem also affects the centroid segmentation and results in decreased segmentation accuracy for the best quality level. Due to the small sizes of centroid segmentations, accuracy scores are very sensitive to small imprecisions and, thus, cause a discrepancy to the baseline scores. The same score progression is shown for instance segmentations obtained with two segmentation approaches, with a minimum decrease of around 0.1 IoU from the baseline results. Nevertheless, the increasing progression of scores according to the increasing quality level demonstrates the utility of the synthesized image data.
5 Conclusion and availability
In this work we demonstrated how a conditional generative adversarial network can be utilized to synthesize realistic 3D fluorescence microscopy image data from binary annotations of cellular structures in a patch-based manner, allowing for a controlled generation of fully-annotated 3D image data sets. Due to the patch-based approach, data of arbitrary size can be generated, enabling an image-based simulation of entire organisms. Different techniques for the generation of annotations of cellular structures were shown in several experiments with varied necessity of manual interaction. The conducted experiments demonstrated the generation of realistic 3D image data from an ideal scenario with manually obtained annotations being available, to a practical scenario with annotations simulated from very sparse detections or segmentations. Using predefined cellular structures as a source domain allows to generate fully-annotated data sets, which, we believe, leverages the usage of learning-based approaches for many biomedical experiments. The proposed conditioning of the generative adversarial network establishes the ability to generate data on different quality levels, which further strengthens control over the generated image data for increased data diversity or detailed benchmarking. In future work we plan to extend our experiments to additional organisms, striving for a larger collection of publicly available fully-annotated 3D microscopy data sets and we plan to further extend the simulation to time series data.In order to make the generated data accessible to the community for training, we publish two data sets comprising nuclei (DOI: 10.17605/OSF.IO/9RG2D) and cellular membrane (DOI: 10.17605/OSF.IO/5EFM9) data generated from automatically obtained annotations with the corresponding instance segmentation data (Fig 11). Furthermore, we publish three different quality levels of the test set (plants 15 and 18, Fig 12) with instance segmentations being available from [24], to provide image data with increasing quality for potential benchmarkings. We decided to publish synthetic image data scaled with α ∈ 10, 100, 500 (DOI: 10.17605/OSF.IO/E6N7B), as those levels proved to be of increasingly better quality for segmentation approaches (Fig 10). We also publish the code repository used to generate 3D annotations and synthetic image data, which is available at https://github.com/stegmaierj/CellSynthesis (DOI: 10.5281/zenodo.5118755).
Fig 11
Public two channel data set.
Samples of the first published data set, including generated 3D image data of nuclei and cellular membrane (top) and the corresponding automatically obtained instance segmentations for nuclei (middle) and membranes (bottom).
Fig 12
Public altered quality data set.
Samples of the second published data set, including 3 different quality levels of the same cellular membrane structures. Corresponding instance segmentations (bottom) are available from [24].
Public two channel data set.
Samples of the first published data set, including generated 3D image data of nuclei and cellular membrane (top) and the corresponding automatically obtained instance segmentations for nuclei (middle) and membranes (bottom).
Public altered quality data set.
Samples of the second published data set, including 3 different quality levels of the same cellular membrane structures. Corresponding instance segmentations (bottom) are available from [24].
Transfer Alert
This paper was transferred from another journal. As a result, its full editorial history (including decision letters, peer reviews and author responses) may not be present.1 Oct 2021PONE-D-21-255013D fluorescence microscopy data synthesis for segmentation and benchmarkingPLOS ONEDear Dr. Eschweiler,Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.Please consider the revisions suggested by the reviewers.Please submit your revised manuscript by Nov 15 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.Please include the following items when submitting your revised manuscript:A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.We look forward to receiving your revised manuscript.Kind regards,Kristen C. Maitland, Ph.D.Academic EditorPLOS ONEJournal Requirements:When submitting your revision, we need you to address these additional requirements.1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found athttps://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf andhttps://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdfPlease review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.[Note: HTML markup is below. Please do not edit.]Reviewers' comments:Reviewer's Responses to QuestionsComments to the Author1. Is the manuscript technically sound, and do the data support the conclusions?The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.Reviewer #1: YesReviewer #2: Yes**********2. Has the statistical analysis been performed appropriately and rigorously?Reviewer #1: YesReviewer #2: Yes**********3. Have the authors made all data underlying the findings in their manuscript fully available?The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.Reviewer #1: YesReviewer #2: Yes**********4. Is the manuscript presented in an intelligible fashion and written in standard English?PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.Reviewer #1: YesReviewer #2: No**********5. Review Comments to the AuthorPlease use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)Reviewer #1: The authors present work addressing the need for automated generation of training data sets for deep-learning based segmentation of cell membranes and nuclei. Different methods for generating the synthetic training sets are evaluated qualitatively and quantitatively, as well as how effective they are when used to train a segmentation network that is applied to real data. This type of methods analysis is important for biologists to appreciate and understand the nuance and challenges still existing with deep-learning segmentation tools. It also highlights the importance of data quality and provides a way to asses optimal data quality for future acquisitions. The annotated data sets made available may be very useful to research groups that are still in the process of manually annotating their own data and don't have access to a useful already trained network.I recommend this work be published with a few suggestions for the author's consideration:My biggest suggestion is to consolidate the figures. I think they could be combined for a total of 10 figures at most that tell the story in a more succinct way.Line 5: the comma after "analyzed" should be removedLine 7: a comma could be included after "segmentation"When discussing spherical harmonics, it may be useful for a general audience to compare to Fourier Transform, a more familiar weighted sum of basis functions. Though, the presentation given is already done well.Line 202: The "in conclusion" seems unnecessaryLine 209: It would be good to mention the accuracy measures that will be used here.Figure 5 legend: I would add commas so the text reads "Annotations can be obtained from manual, automated, and simulation approaches, and final cellular annotations are used to...."Figure 4: If there are these visual examples provided of the spherical harmonic models, it would be nice to see the statistical shape models also.Line 216: It would be useful to mention the imaging lens numerical aperture and physical pixel sizes here from ref 32. for quick understanding of the spatial resolution of the data sets.Line 222: It would be useful to mention this data set was acquired with a multi-photon fluorescence microscope (ref 33).Line 250: By integrating the z-dimension for the intensity profiles you lose the ability to asses the depth-dependent performance of the GAN. It would be interesting to have at least two -perhaps an "upper" and "lower" profile if not a variety of individual xy planes at different z positions.Line 252: The similarity between the real and synthetic xz intensity spectra could be qualified with the visually obvious discrepancies that happen in the high-frequency areas. What accounts for those discrepancies could be discussed.Line 262 and 263: If a few words describing the approaches in [34] and [35] could be used here it would improve the readability.Line 304: "allow to again conclude the generation of realistic 3D image data" is missing a subject, consider "allow us to again conclude..."Line 316: Please describe the approach in [37] for the audience's quick reference.Line 358: It would be interesting to also quantify the different "quality levels" of the synthesized data sets in terms of signal-to-noise ratio, contrast ratio, or some other normal image quality metric.Reviewer #2: The paper is overall well presented but there are a couple confusing aspects to the presentation.There is no top down explanation of the logic of the various experimental mixes of images and masks which makes things more challenging than necessary to understandThere are a few grammatically problematic sentences e.g. "The obtained scores are similar to the scores obtained from the previous experiments and allow [us] to again conclude [success in] the generation of realistic 3D image data."It would be helpful to situate this work with respect to other CNN based training data simulation efforts e.g. NucleAIzer Hollandi et al Cell Systems 2020**********6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.If you choose “no”, your identity will remain anonymous but your review may still be made public.Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.Reviewer #1: NoReviewer #2: Yes: Anthony Santella[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.3 Nov 2021We thank you for considering our manuscript for publication as an article in PLOS ONE and we are excited to revise our manuscript based on the very helpful comments and suggestions from both reviewers.As one of the major changes made to our manuscript, we swapped sections 2 and 3 for improved readability and better coherence of the whole story. Further changes and detailed responses to the reviewers’ suggestions are listed in the following. We are looking forward to your assessment of the improvements made to the manuscript and hope all requests have been fulfilled.Reviewer #1:My biggest suggestion is to consolidate the figures. I think they could be combined for a total of 10 figures at most that tell the story in a more succinct way.- We totally agree that there have been to many figures impairing the reader’s experience. The total number of figures has been reduced by combining or removing previous figures, although we ended up at a quantity slightly above 10.Line 5: the comma after "analyzed" should be removed- Removed the comma, thanks for pointing this out.Line 7: a comma could be included after "segmentation"- Added the comma, thanks for the hint.When discussing spherical harmonics, it may be useful for a general audience to compare to Fou-rier Transform, a more familiar weighted sum of basis functions. Though, the presentation given is already done well.- We agree that the concept of the Fourier transform is very likely more familiar to the reader, and we added a hint to the similarities between both concepts to the spherical harmonic section. Thanks for pointing this out. However, we feel that a full comparison between both approaches would lead to an overly complex and lengthy explanation, that might lead to confusion, as Fourier transform is not used later on.Line 202: The "in conclusion" seems unnecessary- This phrase was removed.Line 209: It would be good to mention the accuracy measures that will be used here.- This should refer to the general quality of the generated structures. However, we think that this sentence is rather confusing and does not provide reasonable information, which is why we re-moved it.Figure 5 legend: I would add commas so the text reads "Annotations can be obtained from man-ual, automated, and simulation approaches, and final cellular annotations are used to...."- Commas were added to the legend.Figure 4: If there are these visual examples provided of the spherical harmonic models, it would be nice to see the statistical shape models also.- For consistency we added a few examples of shapes generated with statistical shape models.Line 216: It would be useful to mention the imaging lens numerical aperture and physical pixel sizes here from ref 32. for quick understanding of the spatial resolution of the data sets.- Good point, we missed this information entirely. The missing information was added to the data set description.Line 222: It would be useful to mention this data set was acquired with a multi-photon fluores-cence microscope (ref 33).- Microscopy technique and imaging resolution have been added to the data set description.Line 250: By integrating the z-dimension for the intensity profiles you lose the ability to assess the depth-dependent performance of the GAN. It would be interesting to have at least two -perhaps an “upper” and “lower” profile if not a variety of individual xy planes at different z positions.- Assessing the depth-dependent performance was the intention of these plots. To get a better overview of the distribution of intensities, we added further plots showing the distribution over YZ. To our understanding, these plots help to assess how well the cumulative intensity for differ-ent thicknesses of the specimen can be reconstructed. We think, XY profiles would be harder to interpret and would only show depth-dependent performance when considering quite a few z po-sitions, which would start to be cluttering. If you have any tips on how we could represent this in-formation in a more compact format, we would be happy to extend the figure.Line 252: The similarity between the real and synthetic xz intensity spectra could be qualified with the visually obvious discrepancies that happen in the high-frequency areas. What accounts for those discrepancies could be discussed.- One of our possible interpretations is that those discrepancies are caused by non-deterministic components like noise, and differences in signal intensities. We added further interpretations and explanations to Section 4.Line 262 and 263: If a few words describing the approaches in [34] and [35] could be used here it would improve the readability.- A short description of both approaches has been added, which hopefully helps to get a better un-derstanding of the experimental setup.Line 304: “allow to again conclude the generation of realistic 3D image data” is missing a subject, consider “allow us to again conclude…”- Thanks for pointing this out, we corrected the sentence.Line 316: Please describe the approach in [37] for the audience’s quick reference.- We added further explanations of the approach’s concept for a quick reference.Line 358: It would be interesting to also quantify the different “quality levels” of the synthesized data sets in terms of signal-to-noise ratio, contrast ratio, or some other normal image quality metric.- The PSNR was added as further metric. Due to the different value ranges of PSNR and the other metrics, we replaced the plot with a table.Reviewer #2:There is no top down explanation of the logic of the various experimental mixes of images and masks which makes things more challenging than necessary to understand.- A general explanation of the choice of our experiments was indeed missing and would help to clar-ify the whole evaluation section. A few sentences explaining our intention of the experiments was added to the beginning of Section 4.There are a few grammatically problematic sentences e.g. "The obtained scores are similar to the scores obtained from the previous experiments and allow [us] to again conclude [success in] the generation of realistic 3D image data."- Thanks for pointing this out, we corrected the sentence.It would be helpful to situate this work with respect to other CNN based training data simulation efforts e.g. NucleAIzer Hollandi et al Cell Systems 2020- We think that a direct comparison to this approach would not be straightforwardly possible, as the generation of images in 2D and patch-based 3D faces significantly different challenges. However, we agree that referring to this approach would complement the state-of-the-art literature, which is why we added a citation of this approach to the introduction section.Submitted filename: Response to Reviewers.pdfClick here for additional data file.11 Nov 20213D fluorescence microscopy data synthesis for segmentation and benchmarkingPONE-D-21-25501R1Dear Dr. Eschweiler,We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.Kind regards,Kristen C. Maitland, Ph.D.Academic EditorPLOS ONEAdditional Editor Comments (optional):Reviewers' comments:23 Nov 2021PONE-D-21-25501R13D fluorescence microscopy data synthesis for segmentation and benchmarkingDear Dr. Eschweiler:I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.If we can help with anything else, please email us at plosone@plos.org.Thank you for submitting your work to PLOS ONE and supporting open access.Kind regards,PLOS ONE Editorial Office Staffon behalf ofDr. Kristen C. MaitlandAcademic EditorPLOS ONE
Authors: Erik Meijering; Anne E Carpenter; Hanchuan Peng; Fred A Hamprecht; Jean-Christophe Olivo-Marin Journal: Nat Biotechnol Date: 2016-12-07 Impact factor: 54.908
Authors: Simon Eck; Stefan Wörz; Katharina Müller-Ott; Matthias Hahn; Andreas Biesdorf; Gunnar Schotta; Karsten Rippe; Karl Rohr Journal: Med Image Anal Date: 2016-03-17 Impact factor: 8.545
Authors: Petr Strnad; Stefan Gunther; Judith Reichmann; Uros Krzic; Balint Balazs; Gustavo de Medeiros; Nils Norlin; Takashi Hiiragi; Lars Hufnagel; Jan Ellenberg Journal: Nat Methods Date: 2015-12-14 Impact factor: 28.547
Authors: Emmanuel Faure; Thierry Savy; Barbara Rizzi; Camilo Melani; Olga Stašová; Dimitri Fabrèges; Róbert Špir; Mark Hammons; Róbert Čúnderlík; Gaëlle Recher; Benoît Lombardot; Louise Duloquin; Ingrid Colin; Jozef Kollár; Sophie Desnoulez; Pierre Affaticati; Benoît Maury; Adeline Boyreau; Jean-Yves Nief; Pascal Calvat; Philippe Vernier; Monique Frain; Georges Lutfalla; Yannick Kergosien; Pierre Suret; Mariana Remešíková; René Doursat; Alessandro Sarti; Karol Mikula; Nadine Peyriéras; Paul Bourgine Journal: Nat Commun Date: 2016-02-25 Impact factor: 14.919
Authors: Martin Weigert; Kaushikaram Subramanian; Sebastian T Bundschuh; Eugene W Myers; Moritz Kreysing Journal: PLoS Comput Biol Date: 2018-04-13 Impact factor: 4.475
Authors: Johannes Stegmaier; Jens C Otte; Andrei Kobitski; Andreas Bartschat; Ariel Garcia; G Ulrich Nienhaus; Uwe Strähle; Ralf Mikut Journal: PLoS One Date: 2014-02-27 Impact factor: 3.240
Authors: Claire McQuin; Allen Goodman; Vasiliy Chernyshev; Lee Kamentsky; Beth A Cimini; Kyle W Karhohs; Minh Doan; Liya Ding; Susanne M Rafelski; Derek Thirstrup; Winfried Wiegraebe; Shantanu Singh; Tim Becker; Juan C Caicedo; Anne E Carpenter Journal: PLoS Biol Date: 2018-07-03 Impact factor: 8.029