| Literature DB >> 32929164 |
Bálint Ármin Pataki1,2, Sébastien Matamoros3, Boas C L van der Putten3,4, Daniel Remondini5, Enrico Giampieri6, Derya Aytan-Aktug7, Rene S Hendriksen7, Ole Lund8, István Csabai9,10, Constance Schultsz3,4.
Abstract
It is important that antibiotics prescriptions are based on antimicrobial susceptibility data to ensure effective treatment outcomes. The increasing availability of next-generation sequencing, bacterial whole genome sequencing (WGS) can facilitate a more reliable and faster alternative to traditional phenotyping for the detection and surveillance of AMR. This work proposes a machine learning approach that can predict the minimum inhibitory concentration (MIC) for a given antibiotic, here ciprofloxacin, on the basis of both genome-wide mutation profiles and profiles of acquired antimicrobial resistance genes. We analysed 704 Escherichia coli genomes combined with their respective MIC measurements for ciprofloxacin originating from different countries. The four most important predictors found by the model, mutations in gyrA residues Ser83 and Asp87, a mutation in parC residue Ser80 and presence of the qnrS1 gene, have been experimentally validated before. Using only these four predictors in a linear regression model, 65% and 93% of the test samples' MIC were correctly predicted within a two- and a four-fold dilution range, respectively. The presented work does not treat machine learning as a black box model concept, but also identifies the genomic features that determine susceptibility. The recent progress in WGS technology in combination with machine learning analysis approaches indicates that in the near future WGS of bacteria might become cheaper and faster than a MIC measurement.Entities:
Year: 2020 PMID: 32929164 PMCID: PMC7490380 DOI: 10.1038/s41598-020-71693-5
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1Midpoint-rooted phylogenetic tree of the 704 E. coli samples that had ciprofloxacin MIC measurement. It is clearly visible that the test data is clustered separately from the training data suggesting the generalization power of our model. Nodes with lower than 80% bootstrap support are collapsed.
Number of features, R score, Pearson correlation (R), Major Error (ME), Very Major Error (VME), area under the receiver operating curve (AUC), Accuracy within a two/four-fold dilution (ACC-2, ACC-4) and Mean Absolute Fold Error (MAFE) on the unseen test data. For the AUC, ME, VME the data was binarized using 1 mg/L threshold.
| Model | N_feat | R | R | ME | VME | AUC | ACC-2 | ACC-4 | MAFE |
|---|---|---|---|---|---|---|---|---|---|
| Random forest | 4 | 0.932 | 0.966 | 1 | 0 | 1.000 | 0.658 | 0.944 | 0.883 |
| Random forest | 15 | 0.902 | 0.951 | 5 | 0 | 0.996 | 0.680 | 0.914 | 0.915 |
| Linear regression | 4 | 0.918 | 0.959 | 0 | 2 | 1.000 | 0.650 | 0.929 | 0.954 |
The number of features were selected according to the performance using leave-one-country-out validation on the training data, see Supplementary Fig. S1.
Number of samples.
Calculated on the log2 values.
The lower the better.
Figure 2Prediction on the unseen test set was generated via random forest and linear regression model using the best four predictors. It can be clearly seen that the two models do not differ much in terms of predicted values.
The collected and used data in the analysis grouped by country and MIC values.
| MIC (mg/L) | Denmark | Italy | NA | USA | UK | Vietnam | Total |
|---|---|---|---|---|---|---|---|
| 0.010 | 0 | 0 | 9 | 0 | 0 | 2 | 11 |
| 0.012 | 0 | 0 | 0 | 0 | 0 | 1 | 1 |
| 0.015 | 119 | 13 | 42 | 49 | 92 | 0 | 315 |
| 0.016 | 0 | 0 | 0 | 0 | 0 | 2 | 2 |
| 0.023 | 0 | 0 | 0 | 0 | 0 | 1 | 1 |
| 0.030 | 12 | 0 | 6 | 3 | 4 | 0 | 25 |
| 0.060 | 1 | 0 | 7 | 1 | 0 | 0 | 9 |
| 0.120 | 0 | 0 | 11 | 2 | 0 | 0 | 13 |
| 0.125 | 0 | 0 | 0 | 0 | 0 | 6 | 6 |
| 0.190 | 0 | 0 | 0 | 0 | 0 | 10 | 10 |
| 0.250 | 6 | 0 | 22 | 11 | 3 | 16 | 58 |
| 0.380 | 0 | 0 | 0 | 0 | 0 | 5 | 5 |
| 0.500 | 0 | 0 | 6 | 2 | 0 | 11 | 19 |
| 0.750 | 0 | 0 | 0 | 0 | 0 | 1 | 1 |
| 1.000 | 0 | 0 | 5 | 2 | 0 | 5 | 12 |
| 2.000 | 0 | 0 | 3 | 0 | 0 | 1 | 4 |
| 4.000 | 0 | 0 | 2 | 6 | 0 | 1 | 9 |
| 8.000 | 0 | 0 | 30 | 0 | 1 | 2 | 33 |
| 12.00 | 0 | 0 | 0 | 0 | 0 | 1 | 1 |
| 16.00 | 0 | 0 | 23 | 0 | 0 | 0 | 23 |
| 24.00 | 0 | 0 | 0 | 0 | 0 | 1 | 1 |
| 32.00 | 0 | 0 | 72 | 0 | 0 | 45 | 117 |
| 64.00 | 0 | 0 | 28 | 0 | 0 | 0 | 28 |
| Total | 138 | 13 | 266 | 76 | 100 | 111 | 704 |
Country metadata is not available.
Figure 3Workflow of the study. First, a random forest model was fitted to the training data with leave-one-country-out validation. Feature importances of the fitted models are averaged over all the folds and the four best features are kept. Then the random forest model and a linear regression model were fitted on all the training samples using only the four best features. And model performances are tested using the independent test dataset.