| Literature DB >> 30716085 |
Ahmed A Metwally1,2,3, Philip S Yu2, Derek Reiman1, Yang Dai1, Patricia W Finn3,4, David L Perkins1,3,5.
Abstract
Food allergy is usually difficult to diagnose in early life, and the inability to diagnose patients with atopic diseases at an early age may lead to severe complications. Numerous studies have suggested an association between the infant gut microbiome and development of allergy. In this work, we investigated the capacity of Long Short-Term Memory (LSTM) networks to predict food allergies in early life (0-3 years) from subjects' longitudinal gut microbiome profiles. Using the DIABIMMUNE dataset, we show an increase in predictive power using our model compared to Hidden Markov Model, Multi-Layer Perceptron Neural Network, Support Vector Machine, Random Forest, and LASSO regression. We further evaluated whether the training of LSTM networks benefits from reduced representations of microbial features. We considered sparse autoencoder for extraction of potential latent representations in addition to standard feature selection procedures based on Minimum Redundancy Maximum Relevance (mRMR) and variance prior to the training of LSTM networks. The comprehensive evaluation reveals that LSTM networks with the mRMR selected features achieve significantly better performance compared to the other tested machine learning models.Entities:
Mesh:
Year: 2019 PMID: 30716085 PMCID: PMC6361419 DOI: 10.1371/journal.pcbi.1006693
Source DB: PubMed Journal: PLoS Comput Biol ISSN: 1553-734X Impact factor: 4.475
Fig 1The proposed deep learning framework, where n denotes the number of timepoints from each subject, which is not the same for all subjects.
Fig 2Architecture of the autoencoder used in the proposed model.
It has three hidden layers with 60, 25, and 60 neurons in that order. The number of neurons for the output layer equals the number of raw features (215 in our case).
Fig 3Summary of the DIABIMMUNE dataset.
(A) Number of subjects allergic to milk, egg, and peanut within the DIABIMMUNE cohort after filtering out missing data. The food-allergy group is the summation of milk, egg, and peanut allergy. (B) Time distribution of 731 samples from 195 subjects (281 samples from 71 Finnish, 197 samples from 70 Estonian, and 253 samples from 54 Russian) stool samples sequenced using MGS from the DIABIMMUNE project. The collected samples have various forms of inconsistencies, such as different numbers of samples per subject.
Losses (training, validation, and testing) of the best trained model for several autoencoder architectures.
The 60x25x60 architecture is the chosen model because it achieves small loss while using the least number of parameters.
| Architecture | Training | Validation | Testing |
|---|---|---|---|
| 150x50x150 | 16.95 | 17.26 | 16.90 |
| 150x25x150 | 16.90 | 18.13 | 16.59 |
| 150x15150 | 16.87 | 17.31 | 17.74 |
| 150x10x150 | Model does not converge | ||
| 125x50x125 | 17.01 | 16.74 | 17.35 |
| 125x25x125 | 16.98 | 17.52 | 17.31 |
| 125x15x125 | 16.91 | 16.54 | 17.21 |
| 125x10x125 | Model does not converge | ||
| 100x50x100 | 16.83 | 19.00 | 16.09 |
| 100x25x100 | 16.95 | 17.58 | 160.9 |
| 100x15x100 | Model does not converge | ||
| 100x10x100 | Model does not converge | ||
| 75x50x75 | 16.07 | 17.96 | 16.30 |
| 75x25x75 | 17.27 | 14.56 | 112.0 |
| 75x15x75 | Model does not converge | ||
| 75x10x75 | Model does not converge | ||
| 60x50x60 | 16.80 | 17.55 | 17.06 |
| 60x25x60 | 16.18 | 16.79 | 17.04 |
| 60x15x60 | 16.75 | 16.83 | 50.90 |
| 60x10x60 | 16.60 | 35.70 | 51.99 |
| 50x50x50 | 15.87 | 17.34 | 35.05 |
| 50x25x50 | 17.09 | 16.02 | 34.00 |
| 50x15x50 | 17.09 | 16.08 | 207.0 |
| 50x10x50 | 17.19 | 16.55 | 263.1 |
Fig 4(A) Trajectory of training and validation loss function (Loss(2)) of the autoencoder with 60x25x60 architecture. The lowest loss on validation set was 16.79 and when the test set applied to this best model, the loss was 17.04. (B) Trajectory of training and validation loss function (Loss(1)) of the LSTM. The lowest loss on validation set was 0.91 and when the test set applied to this best model, the loss was 1.02
Evaluation of auROC and MCC for the proposed LSTM model versus baseline models.
In this table, we evaluated six classifiers; LSTM, HMM, MLPNN, RF, SVM, and LASSO. For each classifier, we evaluated four types of input features; latent features which extracted from the trained autoencoder, 25 features selected by mRMR method, 25 most variable features, and 215 raw taxonomic profile features. The auROC and MCC results shown below are the average of auROC and MCC measured on the test set. The experiments were repeated 10 times and samples were shuffled after each 10-fold cross-validation to test the robustness of each classifier. P-values were calculated using Mann-Whitney U test between LSTM-mRMR-25 versus each corresponding method.
| Model | Features | auROC [mean (sd)] | auROC [p-value] | MCC [mean (sd)] | MCC [p-value] |
|---|---|---|---|---|---|
| 0.67 (0.07) | 4.20E-01 | 0.29 (0.22) | 3.47E-03 | ||
| 0.69 (0.09) | 1 | 0.40 (0.19) | 1 | ||
| 0.61 (0.12) | 7.00E-02 | 0.23 (0.20) | 9.51E-04 | ||
| 0.65 (0.10) | 1.70E-01 | 0.37 (0.18) | 5.10E-02 | ||
| 0.52 (0.06) | 5.99E-22 | 0.14 (0.20) | 2.25E-12 | ||
| 0.55 (0.08) | 8.92E-16 | 0.21 (0.19) | 3.67E-07 | ||
| 0.52 (0.07) | 4.68E-22 | 0.16 (0.15) | 1.18E-10 | ||
| 0.50 (0.07) | 5.63E-25 | 0.06 (0.21) | 7.66E-17 | ||
| 0.54 (0.08) | 5.50E-04 | 0.11 (0.13) | 5.93E-06 | ||
| 0.57 (0.06) | 1.95E-03 | 0.18 (0.11) | 5.95E-03 | ||
| 0.54 (0.03) | 6.66E-05 | 0.05 (0.07) | 1.13E-05 | ||
| 0.63 (0.05) | 4.93E-02 | 0.12 (0.12) | 1.13E-03 | ||
| 0.44 (0.14) | 3.00E-22 | 0.05 (0.25) | 1.99E-15 | ||
| 0.63 (0.13) | 6.73E-02 | 0.23 (0.23) | 6.00E-03 | ||
| 0.57 (0.12) | 1.08E-07 | 0.08 (0.23) | 2.31E-15 | ||
| 0.59 (0.13) | 2.00E-03 | 0.14 (0.23) | 2.71E-11 | ||
| 0.52 (0.15) | 6.49E-12 | 0.06 (0.21) | 9.20E-31 | ||
| 0.57 (0.13) | 1.46E-07 | -0.04 (0.23) | 1.64E-28 | ||
| 0.46 (0.15) | 1.45E-20 | 0.03 (0.22) | 9.20E-31 | ||
| 0.52 (0.14) | 7.21E-13 | -0.04 (0.23) | 1.23E-27 | ||
| 0.51 (0.14) | 1.51E-14 | -0.04 (0.21) | 5.90E-27 | ||
| 0.59 (0.12) | 2.00E-04 | 0.01 (0.23) | 1.45E-26 | ||
| 0.59 (0.13) | 3.20E-05 | -0.17 (0.01) | 2.36E-30 | ||
| 0.57 (0.14) | 1.00E-03 | -0.17 (0.00) | 1.66E-30 |