Literature DB >> 32885818

Synthetic minority oversampling of vital statistics data with generative adversarial networks.

Aki Koivu1, Mikko Sairanen2, Antti Airola1, Tapio Pahikkala1.   

Abstract

OBJECTIVE: Minority oversampling is a standard approach used for adjusting the ratio between the classes on imbalanced data. However, established methods often provide modest improvements in classification performance when applied to data with extremely imbalanced class distribution and to mixed-type data. This is usual for vital statistics data, in which the outcome incidence dictates the amount of positive observations. In this article, we developed a novel neural network-based oversampling method called actGAN (activation-specific generative adversarial network) that can derive useful synthetic observations in terms of increasing prediction performance in this context.
MATERIALS AND METHODS: From vital statistics data, the outcome of early stillbirth was chosen to be predicted based on demographics, pregnancy history, and infections. The data contained 363 560 live births and 139 early stillbirths, resulting in class imbalance of 99.96% and 0.04%. The hyperparameters of actGAN and a baseline method SMOTE-NC (Synthetic Minority Over-sampling Technique-Nominal Continuous) were tuned with Bayesian optimization, and both were compared against a cost-sensitive learning-only approach.
RESULTS: While SMOTE-NC provided mixed results, actGAN was able to improve true positive rate at a clinically significant false positive rate and area under the curve from the receiver-operating characteristic curve consistently. DISCUSSION: Including an activation-specific output layer to a generator network of actGAN enables the addition of information about the underlying data structure, which overperforms the nominal mechanism of SMOTE-NC.
CONCLUSIONS: actGAN provides an improvement to the prediction performance for our learning task. Our developed method could be applied to other mixed-type data prediction tasks that are known to be afflicted by class imbalance and limited data availability.
© The Author(s) 2020. Published by Oxford University Press on behalf of the American Medical Informatics Association.

Entities:  

Keywords:  artificial intelligence; deep learning; machine learning; stillbirth; vital statistics

Mesh:

Year:  2020        PMID: 32885818      PMCID: PMC7750982          DOI: 10.1093/jamia/ocaa127

Source DB:  PubMed          Journal:  J Am Med Inform Assoc        ISSN: 1067-5027            Impact factor:   4.497


INTRODUCTION

Background and significance

Real-life vital statistics data commonly suffer from class imbalance, in which 1 or more of the predicted classes are highly underrepresented. This occurs naturally, as the amount of positive observations is controlled by the prevalence of a disease. This has been shown to have a negative effect on the learning process of probabilistic models. Enough data to model from is a recurrent problem, and several methods have been proposed over the years to address it, ranging from method-specific to more universal. Simple methods such as minority oversampling and majority undersampling are still commonplace, as they provide straightforward and universal solution. Another common practice is to calculate class weights from the learning data and apply them during model training. This is called cost-sensitive learning, in which the model is made aware of the class unbalance in the training data by first assigning higher weight value for the minority and lower for the majority, then incorporating those weights to the training process. For neural networks, one could use for example a weighted loss function such as the weighted cross-entropy with backpropagation. Cost-sensitive learning and sampling methods both try to solve the class imbalance problem, but it is not yet clear which is better in general. Weiss et al compared cost-sensitive learning with sampling methods and found out that there was no clear winner when the testing was done using multiple different datasets with differing percentage of minority observations. Oversampling the minority class with duplicate observations was deemed useful with some datasets and useless with others. Synthetic data generation refers to a case of oversampling where instead of duplicating existing observations, completely new ones are created. This is achieved by successfully modelling the distributions of each variable of the training data and then sampling from the created joint multivariate probability distribution. In the clinical domain, established methods for data generation, such as Synthetic Minority Over-sampling Technique (SMOTE), have been successfully implemented. Modern methods such as the generative adversarial networks (GANs) have also made significant improvements to medical image generation tasks. GAN methods for tabular data have also been proposed, in which a long short-term memory recurrent neural network was used as the generator. The success of a generator method is usually tied to the amount of feasible data available. Proper domain expertise can also be utilized in a limited data setting. Real-life data can have constrains that are obvious to domain experts, but they can be missed by the modeling method if the used data does not represent this sufficiently. The problem is amplified with data that has a substantial class imbalance. To be able to generate valuable synthetic mixed-type observations in terms of improving prediction performance from highly imbalanced vital statistics data is therefore not an elementary task. Discovering a robust and generalizable method for this problem would be significant because deriving more value from this type of data could have a positive impact on public health in general. Stillbirth is a serious pregnancy-related condition defined as a deadborn outcome of a delivery after 20 weeks of gestation. Stillbirth can be further categorized as early and late stillbirth according to the gestational age thresholds. Maternal characteristics contributing to elevated risk for stillbirth have been identified as being high body mass index, advanced maternal age, maternal smoking, parity, ethnicity, education, and various preexisting conditions or comorbidities. Currently, several screening and monitoring mechanisms have been proposed; however a common golden standard has not been agreed upon, so most nations report, rather than screen. One of these mechanisms is predicting risk from maternal characteristics. Recently, we proposed a novel model that assesses the risk of early stillbirth from maternal characteristics. Also, models predicting risk of stillbirth in different gestational age intervals have been developed. Performance of the models utilizing maternal characteristics has been modest, achieving area under the curve (AUC) from a receiver-operating characteristic (ROC) between 0.6 and 0.7, while models that had biophysical measurements as added features such as fetal presentation achieved an improved AUC of 0.8.

OBJECTIVE

The aim of this research is to develop a novel GAN-based data generation method suitable for generating synthetic cases from vital statistics data. Tabular mixed-type data with a significant class imbalance problem pose challenges for modelling and limit the amount of applicable methods. Our goal is to introduce a new method capable of generating more value in this context. Prediction performance is the key optimizable parameter we want to focus on. The research question was the following: can we derive more predictive power from the existing data with generative methods while not changing the model that is responsible for the outcome prediction? Our new method could be applied to similar prediction tasks that involve highly imbalanced data. For demonstrating performance over established methods such as SMOTE, the prediction of early stillbirth pregnancies was chosen for our use case, as it provided the desired restrictions data-wise.

MATERIALS AND METHODS

Study data

The dataset was provided by the New York City Department of Health and Mental Hygiene and contained reported pregnancies in the New York City area from 2014 to 2016. The source of the data was collected birth and death certificates, and the data were deidentified of any variables that could be linked to a specific person. Use of the dataset for this research was granted an institutional review board approval by the ethics committee of the Hospital District of Southwest Finland. The selected maternal characteristics feature variables were based on literature and are listed in Table 1.
Table 1.

Feature variables

Demographics
Agediscrete
 Race (white, black, American Indian, Alaskan native, Asian, or Pacific Islander)nominal
 Marital statusbinary
 Education (8th grade or less to doctorate)nominal
 Number of previous terminationsnominal
 Special supplemental nutrition programbinary
 Smoking before pregnancynominal
 BMIcontinuous
 Heightcontinuous
 Paritynominal
Pregnancy history
 Prepregnancy diabetesbinary
 Gestational diabetesbinary
 Prepregnancy hypertensionbinary
 Gestational hypertensionbinary
 Hypertension eclampsiabinary
 Previous preterm birthsbinary
 Infertility treatmentbinary
 Infertility drugsbinary
 Assisted reproductive technologybinary
 Previous cesarean sectionsbinary
Infections
 Gonorrheabinary
 Syphilisbinary
 Chlamydiabinary
 Hepatitis Bbinary
 Hepatitis Cbinary

BMI: body mass index.

Feature variables BMI: body mass index. The dataset contains 364 124 pregnancies in total. In order to predict early stillbirth from a representative population, sample selection was conducted using the following exclusion criteria: Age of the mother was 18 years of age or older. Cases of maternal morbidity were excluded. All the feature variables and predicted class were present, and no value imputation was required. Pregnancies that concluded in fetal death because of external causes were excluded. These were identified with the International Classification of Diseases–Tenth Revision code values containing U, V, W, X, or Y characters. Reported gestational age in pregnancies that concluded in live birth was 21 weeks or older. Because the earliest known preterm baby is 21 weeks and 6 days, it is probable that these pregnancies were erroneously inputted during data collection. Postnatal death cases were excluded. Multiple birth pregnancies were excluded. This decreased the amount of pregnancies to 363 560 live births and 139 early stillbirths. The prevalence for early stillbirth in this dataset is therefore 0.04%, or 1 in 2500. Because of this, the data are highly imbalanced. Data were randomly split into 2 equal-sized sets, 1 for hyperparameter optimization and 1 for model evaluation. This procedure was class-stratified so that the class imbalance would be present in both datasets. Feature variables were preprocessed based on their type. Nominal features were one-hot encoded in order to accommodate modelling with neural networks. Continuous variables were standardized by zero-mean normalization and unit-variance normalization, and the parameters were calculated using the training dataset only and then applied to both datasets.

Generator methods

SMOTE-Nominal Continuous

SMOTE is a well-established oversampling method based on k-nearest neighbors. Given minority class data, the method utilizes Euclidean distances between data points to generate new ones. It features 2 hyperparameters: the number of closest neighbors to be included in calculation is controlled with and , which determines the number of generated observations. The impact of SMOTE and its variants on prediction performance has been studied, Blagus et al demonstrated that SMOTE had no significant effect on performance when used with microarray data, and Van Hulse et al showed that random undersampling of the majority class overperformed SMOTE. However, a variant of SMOTE called SMOTE-Nominal Continuous (SMOTE-NC) addresses an issue that is not addressed by other conventional oversampling methods, generating mixed-type data. Our study dataset contains nominal feature variables mixed with continuous features. SMOTE-NC takes this into consideration by having a separate logic for nominal features. First, the median of standard deviations of all continuous features of the minority class are calculated. When new data points are about to be created, if the nominal features differ when compared with the nearest neighbors, this median is added to the Euclidean distances. The mechanism penalizes differences in nominal features more effectively, which should in theory result in more accurate synthetic nominal features.

Activation-specific GAN

Generative adversarial networks use a competitive learning setting for 2 networks: a generator and a discriminator   is trained to map random noise derived from a distribution to data points in the training dataset, so its objective is to create synthetic observations that fool . is trained to classify the input as fake or real. This means either from the target distribution or not. The 2-objective training loss can be formalized as where is the training data distribution and is the model distribution, defined by , when is the random noise input sampled from a distribution Training this type of network has been shown to be fragile and unstable. Nonconvergence of the 2 models, mode collapse, and diminishing gradient of the generator plagued the original GAN. In 2017, Wasserstein GAN (WGAN) was proposed as a more stable method that also provided more meaningful learning curves in terms of model fitting performance. In WGAN, the discriminator network is replaced by a critic network that scores observations as being real or fake by learning a -Lipschitz function to compute Wasserstein distance. In training, decreasing this distance is the objective loss function, so as it decreases, the resemblance of the generator output and the training data increases. The loss function used with the critic is now defined as where sup is supremum and is the Lipschitz constant for function , and is made to satisfy , or -Lipschitz continuity. The method was a clear improvement over GAN, but in order to preserve the -Lipschitz continuity, the iterated weights of the model were limited to a small value range so that the -Lipschitz function’s lower and upper bounds could be obtained in a feasible manner. This was called weight clipping, which the model is highly sensitive to. Later in 2017, Gulrajani et al improved on this method and proposed WGAN with gradient penalty (WGAN-GP). In WGAN-GP, gradient penalty is used instead of weight clipping to impose the -Lipschitz continuity. A differentiable function is 1-lipschitz if and only if it has gradients with a norm of at most 1 everywhere. The loss function was reworked to penalize if the gradient norm moved away from 1, so it was defined as where is sampled from and with uniformly sampled between 0 and 1 so that and is the penalty coefficient. This change in WGAN-GP stabilized the model training even further because there was no more need to parameterize weight clipping; however, it introduced a new level of complexity to the model. Another popular variation of GAN called deep convolutional GAN has also been proposed. The benefit of using convolution is data aggregation to a smaller space, which is something we do not want to do with mixed-type data, so WGAN-GP was chosen to be the starting point of our research. In order to create synthetic positives that follow the variable-specific constrains of tabular mixed-type data, WGAN-GP needed to be altered to accommodate this. On the one hand, image data, the most common application of GANs, contains pixels that are represented as continuous values with no pixel-specific constraints. On the other hand, mixed-type data can contain nominal and ordinal variables that are affected by rules, such as non-negative integer values only. One-hot encoding also produces multidimensional representations of these integer values that have conditional properties like a stochastic vector; the representation should add up to 1. All these rules are learned by the generator model of a GAN when given enough data, but in a minority oversampling situation in which training data can be limited, this can become unfeasible. To solve this problem, we propose an output layer activation-specific GAN (actGAN). Based on WGAN-GP architecture with Wasserstein loss and gradient penalty, actGAN was developed to specifically handle mixed-type data in a minority learning setting. Output layer neurons’ activation functions of the generator model were selected based on the variable type they were generating. Continuous and discrete variables were generated using ReLU function  which demands that only non-negative values are created for them. Binary variables were created with a logistic function which is suitable for creating a binary result. One-hot encoding representations of nominal features were generated with softmax  for and . The internal normalization procedure of softmax ensures that the sum of the elements of the output vector is 1. This customized output layer enabled us to provide prior knowledge of the generated variables, which enabled actGAN to learn more sufficiently from a smaller number of observations. The hidden layer activations were chosen to be scaled exponential linear unit (SELU) functions accompanied with LeCun normal weight initialization. The design of the generator is depicted in Figure 1.
Figure 1.

Schematic view of the training cycle of the actGAN (activation-specific generative adversarial network) generator and critic models. Random noise z is given to the generator model as input, and the model produces vector representation of an observation based on the architecture defined in the output layer. The output layer of the generator is comprised of nominal (N), continuous (C), binary (B), and discrete (D) nodes. The critic is given real positive observations and fake generated observations as input, and it outputs a score of realness by approximating the Wasserstein distance. These 2 models are trained in adversarial manner. After training, the critic is discarded, and the generator is used to create fake positive observations.

Schematic view of the training cycle of the actGAN (activation-specific generative adversarial network) generator and critic models. Random noise z is given to the generator model as input, and the model produces vector representation of an observation based on the architecture defined in the output layer. The output layer of the generator is comprised of nominal (N), continuous (C), binary (B), and discrete (D) nodes. The critic is given real positive observations and fake generated observations as input, and it outputs a score of realness by approximating the Wasserstein distance. These 2 models are trained in adversarial manner. After training, the critic is discarded, and the generator is used to create fake positive observations. The critic network architecture was kept simple on purpose, because discrimination as a task is simpler compared with generation. Two hidden layers of width 128 and 64 were used with Leaky ReLU activations that were parameterized with 0.2 slope coefficient, and He normal initialization of weights. The output would be linear activation of 1 node that was required by WGAN functionality. After training, the critic network would be discarded, and the finalized generator network would be used with random noise input to generate synthetic observations. The designs of the generator and critic models are depicted in Figure 1.

Classifiers

In order to compare data generation methods in terms of added prediction performance, 2 classifier algorithms were chosen that would utilize the generated data for prediction. They were the frequently used logistic regression (LR) and the neural network with a SELU network, which represents state of the art of feed-forward networks. Hyperparameters for them were also tuned during hyperparameter optimization, with the aim of maximizing prediction performance. For fitting the LR, limited-memory version of Broyden-Fletcher-Goldfarb-Shanno algorithm was used with the L2 norm penalty. For SELU network, mini-batch adaptive moment estimation (Adam) gradient descent with 0.001 learning rate, 0.9 decay rate, and 0.999 decay rate were used for updating weights with backpropagation. Weight initialization was anchored with a random seed, and was used in all the experiments. The SELU network requires the use of LeCun normal weight initialization and alpha node dropout; the amount of dropout after every hidden layer was set to 15%, which was deemed appropriate for preventing overfitting. Cost-sensitive learning with class weights was deemed necessary for any feasible classifier model fit because of the magnitude of class imbalance in the data. Class weights were calculated from the training dataset using the following equation: where is the number of observations, is the number of different classes, and is the frequency of classes in data labels .

Experimental overview

The study experiments were divided into 2 phases: hyperparameter optimization and model evaluation. Each of them had their own subset dataset. The model’s hyperparameters were optimized in a 2-fold cross-validation procedure with the training dataset, and then finalized models were trained with the same data. After this, they were evaluated with the separate evaluation dataset. This guaranteed that there was no chance to overfit the model to the evaluation data, as both model fitting and hyperparameter selection were done on the separate training dataset. The data flow of our experiments is depicted in Figure 2. The performance of a generator or a classifier model is highly affected by used hyperparameters, so the hyperparameter spaces of LR, SELU network, SMOTE-NC, and actGAN were investigated iteratively. Hyperparameters for classifiers were first investigated independently, and the best found sets were used while optimizing hyperparameters for generator models. Bayesian optimization was used to calculate optimal hyperparameters, and the process can be represented as a formula: where is the AUC score from an ROC curve to be maximized, and are the optimal hyperparameters calculated from training dataset. Twofold class-stratified cross-validation was used for the optimization process, and the reported AUC was averaged over the folds. For measuring the effect of hyperparameters in terms of AUC, Pearson’s correlation coefficient was used with the effect interpretation table to determine linear correlation and its significance. The number of optimization iterations after 10 initial random iterations was chosen to be 500 because it was estimated that it would be enough for Bayesian optimization.
Figure 2.

Data flow diagram of the experiments. Study data would be randomly divided in half, in a class-stratified way. Training data would be used to tune hyperparameters and train the final models, while evaluation data would be used to test their performance. CV: cross-validation.

Data flow diagram of the experiments. Study data would be randomly divided in half, in a class-stratified way. Training data would be used to tune hyperparameters and train the final models, while evaluation data would be used to test their performance. CV: cross-validation. Tunable hyperparameters of LR were regularization parameter C and the number of iterations. For the SELU network, tunable hyperparameters were the number of hidden layers, nodes in hidden layers, number of epochs, and batch size. Number of synthetic observations and were iterated for SMOTE-NC. For actGAN, the length of latent dimension, number of epochs, batch size, number of synthetic observations, hidden layers, and nodes in the generator network were iterated over. actGAN’ s penalty coefficient was set to and Adam optimizer parameters of learning rate and decay rates and for both networks were set to 0.0001, 0, and 0.9, respectively, according to the original publication of WGAN-GP. Our method is WGAN-based, which is stated to be not sensitive to generator model architecture or chosen hyperparameters. We wanted to verify this by including the network architecture parameters in our optimization experiment. Hyperparameters and their selected ranges for the optimization process are listed in Supplementary Table 1. For designing an experiment that would test the data generator model’s generalizability and prediction performance, the final models were selected based on 2-fold cross-validation on the training dataset. These models were then evaluated on the independent model evaluation dataset. The used metric was the AUC from an ROC curve, and true positive rate (TPR) at a clinically significant false positive rate (FPR), which would be TPR at 1%, 3%, and 5% FPR for early stillbirth, based on the real-world incidence. After hyperparameter optimization, LR and SELU network classifiers would serve as benchmarks, and all the possible combinations of classifier and generator models were experimented with in model evaluation. The full list of used software libraries and hardware are described in the Supplementary Appendix.

RESULTS

Model evaluation

After Bayesian hyperparameter optimization, the best sets of hyperparameters for the model and training datasets were used to train the final models. The optimization results of the classifiers and generators are listed in the Supplementary Appendix. The material also contains the learning curves of both final actGAN models, the first optimized with LR and the second with the SELU network. Six different prediction configurations were used to predict the evaluation dataset, and the results are presented in Table 2.
Table 2.

Model evaluation results

NameTPR at 1% FPRTPR at 3% FPRTPR at 5% FPRAUC (95% CI)
LR9%16%20%0.688 (0.620-0.756)
SELU network7%16%20%0.659 (0.590-0.728)
SMOTE-NC and LR9%16%20%0.688 (0.620-0.756)
SMOTE-NC and SELU6%17%23%0.663 (0.594-0.733)
actGAN and LR9%16%24%0.637 (0.562-0.712)
actGAN and SELU9% 23% 27% 0.704 (0.635-0.772)

actGAN: activation-specific generative adversarial network; AUC: area under the curve; CI: confidence interval; FPR: false positive rate; LR: logistic regression; SELU: scaled exponential linear unit; SMOTE-NC: Synthetic Minority Over-sampling Technique-Nominal Continuous; TPR: true positive rate.

Model evaluation results actGAN: activation-specific generative adversarial network; AUC: area under the curve; CI: confidence interval; FPR: false positive rate; LR: logistic regression; SELU: scaled exponential linear unit; SMOTE-NC: Synthetic Minority Over-sampling Technique-Nominal Continuous; TPR: true positive rate. When comparing benchmark models, LR was able to achieve better AUC and improve TPR at FPR by over the SELU network. The usage of SMOTE-NC was minimized during SMOTE-NC and LR optimization, so its performance is identical to LR. SMOTE-NC and SELU was able to improve TPR at and FPR over the SELU network; however, AUC and TPR at FPR decreased. actGAN and LR retained TPR at and FPR while improving TPR at FPR but producing the worst AUC. This was caused by a drop in performance in higher FPRs, shown in Figure 3. actGAN and SELU improved every metric over the SELU network and achieved the best performance of the experiment. Variable importance results of the tested classifiers are listed in Supplementary Table 2.
Figure 3.

Receiver-operating characteristic curves of predicting the evaluation dataset with all the experimented configurations of models. The clinically significant false positive range of 1%, 3%, and 5% is presented in the magnified section, which illustrates the noteworthy performance of activation-specific generative adversarial network (actGAN) and scaled exponential linear units (SELU) in 3% and 5% false positive rate. The best area under the curve (AUC) of 0.704 was achieved by actGAN and SELU network. The usage of Synthetic Minority Over-sampling Technique-Nominal Continuous (SMOTE-NC) was minimized in SMOTE-NC and logistic regression (LR), resulting in overlapping receiver-operating characteristic curves with LR and same AUC of 0.688. The third-best AUC of 0.663 was achieved by SMOTE-NC and the SELU network. The worst AUCs of 0.659 and 0.637 were obtained by the SELU network and actGAN, and LR.

actGAN: activation-specific generative adversarial network; AUC:; CI: confidence interval; FPR: false positive rate; LR: logistic regression; SELU: scaled exponential linear unit; SMOTE-NC:; TPR: true positive rate.

Receiver-operating characteristic curves of predicting the evaluation dataset with all the experimented configurations of models. The clinically significant false positive range of 1%, 3%, and 5% is presented in the magnified section, which illustrates the noteworthy performance of activation-specific generative adversarial network (actGAN) and scaled exponential linear units (SELU) in 3% and 5% false positive rate. The best area under the curve (AUC) of 0.704 was achieved by actGAN and SELU network. The usage of Synthetic Minority Over-sampling Technique-Nominal Continuous (SMOTE-NC) was minimized in SMOTE-NC and logistic regression (LR), resulting in overlapping receiver-operating characteristic curves with LR and same AUC of 0.688. The third-best AUC of 0.663 was achieved by SMOTE-NC and the SELU network. The worst AUCs of 0.659 and 0.637 were obtained by the SELU network and actGAN, and LR. actGAN: activation-specific generative adversarial network; AUC:; CI: confidence interval; FPR: false positive rate; LR: logistic regression; SELU: scaled exponential linear unit; SMOTE-NC:; TPR: true positive rate.

DISCUSSION

Applying minority oversampling to vital statistics data can be challenging due to the limited amount of available positive case data, and mixed-type predictor variables. It limits the amount of feasible oversampling methods, and even when applied, the increase in predictive performance might not be substantial enough to justify the usage. The best AUC of the 6 tested methods was achieved by actGAN and SELU, but confidence intervals in Table 2 reveal a notable overlap between all methods, so the statistical significance of the AUC cannot be stated. However, while AUC measures performance across the whole curve, the clinically significant performance in screening for rare disorders is achieved by either increasing TPR within the feasible FPR range of 1% to 5% or decreasing FPR while maintaining the same TPR. SMOTE-NC provided mixed results in model evaluation, implying that the method could improve performance in specific FPRs while decreasing it at others. However, actGAN and SELU was able to derive observations from this data that provided consistent improvement to TPR at FPRs of 3% and 5%. The method was able to produce similar TPR with 3% FPR when compared with competing methods that used 5% FPR. This improvement of 2% to the FPR has a substantial clinical impact in a screening environment. For example, in our study data of 363 560 live births, the improvement from 5% FPR to 3% FPR would mean 7271 less false positive cases over 3 years of screening. The total amount of hidden nodes in actGAN was optimized to 700 with LR and 944 with the SELU network, which would indicate that a more complex classifier paired with a more complex generator can create additional value from the same data in terms of prediction performance. An activation-specific output layer of the generator network enabled us to provide additional information about the underlying data structure of the task, which overperformed the nominal mechanism of SMOTE-NC. While actGAN performed better when compared with base classifiers and SMOTE-NC, the generalizability of our proposed method should be investigated further. The method is not restricted to stillbirth prediction or even the clinical domain, and any prediction task that utilizes mixed-type data could in theory benefit from actGAN. External validation of stillbirth prediction and experimentation of several datasets from various domains would be the topic for future work. Predicting an outcome of low incidence is bound to produce high FPRs. In clinical risk modeling of a particular outcome, instead of maximizing AUC, the TPR at a clinically significant FPR indicates the performance a model would have in routine use. Developing actGAN further to take this performance metric into consideration during training would be the secondary topic for future work. This work provides evidence that proposed synthetic data generation tools can significantly improve maternal characteristics–based risk prediction in rare conditions. Baseline performance received with the multivariate LR model in our experiment is in line with current published tools, in which predicting stillbirth from maternal characteristics resulted in an AUC of 0.658 and 21.1% TPR at 5% FPR. However, synthetic data generation improved on this result in our experiments. We believe that synthetic data generation of rare conditions combined with screening variables that are targeted toward detecting those conditions would be essential for improving clinical risk prediction in the future. Our study data did not contain first‐trimester biophysical and biochemical measurements for stillbirth screening, prediction variables that have been suggested to improve detection. Properly generating synthetic data of these continuous variables could be the key of improving stillbirth screening in general. At the population level, the improvement in prediction could change management of pregnancy in a number of women, resulting in fewer stillbirth outcomes.

CONCLUSION

Our demonstrated actGAN improved the clinically significant performance of the early stillbirth risk modelling. Activation-specific architecture could be designed to fit other clinical risk modelling tasks predicting outcomes from mixed-type data. This would improve development and validation of such clinical prediction tools. The benefits would be evident in models that rely on class-imbalanced population-level data, in which synthetic data generation of rare conditions would be valuable.

AUTHOR CONTRIBUTIONS

AK and MS conceived of the presented idea. AK developed the theory and performed the computations. AK and MS discussed the results and contributed to the first version of the manuscript. For the first revision of the manuscript, AA and TP reviewed the technical aspects and contributed to the grammar corrections, while MS contributed to the clinical aspects of the revision. AK reviewed the overall structure, wrote the necessary changes, and facilitated the revision process. For the second revision of the manuscript, all authors discussed the corrections and AK implemented the necessary changes.

SUPPLEMENTARY MATERIAL

Supplementary material is available at Journal of the American Medical Informatics Association online.

CONFLICT OF INTEREST STATEMENT

None declared. Click here for additional data file.
  15 in total

1.  Digital selection and analogue amplification coexist in a cortex-inspired silicon circuit.

Authors:  R H Hahnloser; R Sarpeshkar; M A Mahowald; R J Douglas; H S Seung
Journal:  Nature       Date:  2000-06-22       Impact factor: 49.962

Review 2.  Reducing stillbirths: screening and monitoring during pregnancy and labour.

Authors:  Rachel A Haws; Mohammad Yawar Yakoob; Tanya Soomro; Esme V Menezes; Gary L Darmstadt; Zulfiqar A Bhutta
Journal:  BMC Pregnancy Childbirth       Date:  2009-05-07       Impact factor: 3.007

3.  Statistics corner: A guide to appropriate use of correlation coefficient in medical research.

Authors:  M M Mukaka
Journal:  Malawi Med J       Date:  2012-09       Impact factor: 0.875

4.  Prediction of stillbirth from maternal demographic and pregnancy characteristics.

Authors:  G Yerlikaya; R Akolekar; K McPherson; A Syngelaki; K H Nicolaides
Journal:  Ultrasound Obstet Gynecol       Date:  2016-10-05       Impact factor: 7.299

5.  Prediction of stillbirth from biochemical and biophysical markers at 11-13 weeks.

Authors:  S Mastrodima; R Akolekar; G Yerlikaya; T Tzelepis; K H Nicolaides
Journal:  Ultrasound Obstet Gynecol       Date:  2016-11       Impact factor: 7.299

6.  The predictive power of diagnostic tests and the effect of prevalence of illness.

Authors:  R J Baldessarini; S Finklestein; G W Arana
Journal:  Arch Gen Psychiatry       Date:  1983-05

7.  Risk factors for antepartum and intrapartum stillbirth.

Authors:  R E Little; C R Weinberg
Journal:  Am J Epidemiol       Date:  1993-06-01       Impact factor: 4.897

8.  A stillbirth calculator: Development and internal validation of a clinical prediction model to quantify stillbirth risk.

Authors:  Amanda S Trudell; Methodius G Tuuli; Graham A Colditz; George A Macones; Anthony O Odibo
Journal:  PLoS One       Date:  2017-03-07       Impact factor: 3.240

9.  Predicting stillbirth in a low resource setting.

Authors:  Gbenga A Kayode; Diederick E Grobbee; Mary Amoakoh-Coleman; Ibrahim Taiwo Adeleke; Evelyn Ansah; Joris A H de Groot; Kerstin Klipstein-Grobusch
Journal:  BMC Pregnancy Childbirth       Date:  2016-09-20       Impact factor: 3.007

10.  Predicting risk of stillbirth and preterm pregnancies with machine learning.

Authors:  Aki Koivu; Mikko Sairanen
Journal:  Health Inf Sci Syst       Date:  2020-03-25
View more
  5 in total

1.  Predicting the WHO/ISUP Grade of Clear Cell Renal Cell Carcinoma Through CT-Based Tumoral and Peritumoral Radiomics.

Authors:  Yanqing Ma; Zheng Guan; Hong Liang; Hanbo Cao
Journal:  Front Oncol       Date:  2022-02-14       Impact factor: 6.244

2.  Prediction of lung metastases in thyroid cancer using machine learning based on SEER database.

Authors:  Wenfei Liu; Shoufei Wang; Ziheng Ye; Peipei Xu; Xiaotian Xia; Minggao Guo
Journal:  Cancer Med       Date:  2022-02-22       Impact factor: 4.711

3.  The CT delta-radiomics based machine learning approach in evaluating multiple primary lung adenocarcinoma.

Authors:  Yanqing Ma; Jie Li; Xiren Xu; Yang Zhang; Yi Lin
Journal:  BMC Cancer       Date:  2022-09-03       Impact factor: 4.638

4.  Comparison of endocrine therapy and chemotherapy as different systemic treatment modes for metastatic luminal HER2-negative breast cancer patients -A retrospective study.

Authors:  Qiuyue Liu; Juan Qiu; Qianrun Lu; Yujin Ma; Shu Fang; Bing Bu; Lihua Song
Journal:  Front Oncol       Date:  2022-07-26       Impact factor: 5.738

5.  SMOTE-NC and gradient boosting imputation based random forest classifier for predicting severity level of covid-19 patients with blood samples.

Authors:  Elif Ceren Gök; Mehmet Onur Olgun
Journal:  Neural Comput Appl       Date:  2021-06-11       Impact factor: 5.606

  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.