| Literature DB >> 32033607 |
Chung-Yu Chen1, Wei-Chi Lin2, Hsiao-Yu Yang3,4,5,6,7.
Abstract
BACKGROUND: Ventilator-associated pneumonia (VAP) is a significant cause of mortality in the intensive care unit. Early diagnosis of VAP is important to provide appropriate treatment and reduce mortality. Developing a noninvasive and highly accurate diagnostic method is important. The invention of electronic sensors has been applied to analyze the volatile organic compounds in breath to detect VAP using a machine learning technique. However, the process of building an algorithm is usually unclear and prevents physicians from applying the artificial intelligence technique in clinical practice. Clear processes of model building and assessing accuracy are warranted. The objective of this study was to develop a breath test for VAP with a standardized protocol for a machine learning technique.Entities:
Keywords: Breath test; Electronic nose; Machine learning; Ventilator-associated pneumonia; Volatile organic compounds
Mesh:
Year: 2020 PMID: 32033607 PMCID: PMC7006122 DOI: 10.1186/s12931-020-1285-6
Source DB: PubMed Journal: Respir Res ISSN: 1465-9921
Fig. 1Flow diagram of this study. The diagram shows our standardized procedures of data collection, data preparation, model building, model evaluation, and model improvement. When the pathogens are colonized in the lung, they will release volatile organic compounds in the breath. We collected the breath from the endotracheal tube and then analyzed the sensor arrays of an electronic nose. The electric resistance changes of sensors were first normalized and autoscaled. Then, we randomly split subjects into a training set and a testing set. We used eight machine learning algorithms to estimate diagnostic accuracy. The parameters of the algorithms are selected with bootstrapping methods. The optimized models were then applied to the testing set to assess the accuracy of the breath test
Demographic characteristics of the study subjects
| Characteristics | Case group ( | Control group ( | |
|---|---|---|---|
| Age (year), mean (SD) | 71.44 (13.43) | 68.90 (15.55) | 0.53 |
| Male, No. (%) | 21 (63.64) | 13 (50.00) | 0.12 |
| Smoking status | 0.18 | ||
| Current smoker, No. (%) | 4 (12.12) | 4 (15.38) | |
| Former smoker, No. (%) | 12 (36.36) | 4 (15.38) | |
| Nonsmoker, No. (%) | 15 (45.45) | 12 (46.15) | |
| White blood cell (103/μL), mean (SD) | 14.63 (7.87) | 15.03 (8.77) | 0.86 |
| Blood urea nitrogen (mg/dL), mean (SD) | 30.65 (18.74) | 32.54 (18.43) | 0.72 |
| Creatinine (mg/dL), mean (SD) | 1.24 (0.71) | 1.58 (1.01) | 0.15 |
| Aspartate aminotransferase (U/L), mean (SD) | 88.48 (139.48) | 53.71 (139.48) | 0.25 |
| Alanine aminotransferase (U/L), mean (SD) | 55.80 (82.11) | 35.62 (25.82) | 0.22 |
| Number of comorbidities | 3.16 (1.10) | 2.90 (1.18) | 0.43 |
Prediction accuracy of the electronic nose in the test set of machine learning algorithms
| Model and parameters | Accuracy (95% CI) | Sensitivity | Specificity | PPV | NPV | Kappa | AUC (95% CI) |
|---|---|---|---|---|---|---|---|
| 0.77 (0.46–0.95) | 0.71 | 0.83 | 0.83 | 0.71 | 0.54 | 0.80 (0.54–1.00) | |
| Naive Bayes (fL = 0, usekernel = TRUE, adjust = 1) | 0.77 (0.46–0.95) | 0.71 | 0.83 | 0.83 | 0.71 | 0.54 | 0.80 (0.54–1.00) |
| Decision tree (trials = 10, model = rules, window = TRUE) | 0.85 (0.55–0.98) | 0.86 | 0.83 | 0.86 | 0.83 | 0.69 | 0.85 (0.63–1.00) |
| Neural network (size = 3, decay = 1e-04) | 0.85 (0.55–0.98) | 0.86 | 0.83 | 0.86 | 0.83 | 0.69 | 0.85 (0.63–1.00) |
| Support vector machines (linear kernel) (C = 1) | 0.85 (0.55–0.98) | 0.86 | 0.83 | 0.86 | 0.83 | 0.69 | 0.85 (0.63–1.00) |
| Support vector machines (radial kernel) (sigma = 1.432815, C = 1) | 0.77 (0.46–0.95) | 0.71 | 0.83 | 0.83 | 0.71 | 0.54 | 0.85 (0.63–1.00) |
| Support vector machines (polynomial kernel) (degree = 1, scale = 0.1, C = 0.5) | 0.85 (0.55–0.98) | 0.86 | 0.83 | 0.86 | 0.83 | 0.69 | 0.85 (0.63–1.00) |
| Random forest (mtry = 32) | 0.77 (0.46–0.95) | 0.71 | 0.83 | 0.83 | 0.71 | 0.54 | 0.90 (0.74–1.00) |
| Mean value (SD) | 0.81 (0.04) | 0.79 (0.08) | 0.83 (0.00) | 0.85 (0.02) | 0.77 (0.06) | 0.62 (0.08) | 0.85 (0.04) |
PPV positive predictive value; NPV negative predictive value; AUC area under the receiver operating curve
Fig. 2The area under the receiver operating curve (AUC) for ventilator-associated pneumonia in the training set, testing set, and full data set. High AUCs in the three sets show high diagnostic accuracy
Fig. 3The partial area under the receiver operating curve (pAUC). The blue area corresponds to the pAUC region between 80 and 100% for specificity (SP), and the green area corresponds to the pAUC region between 80 and 100% for sensitivity (SE). The corrected pAUCs are printed in the middle of the plot
Fig. 4The area under the receiver operating curve (AUC) for ventilator-associated pneumonia in the testing set, with the 95% confidence interval. We used the random forest algorithm to establish the prediction model and then tested the prediction accuracy on the training set. Using bootstrap resampling for 2000 replicates, the 95% confidence intervals are shown as gray areas around the mean bootstrapped curve
Comparison of strengths and weaknesses of machine learning algorithms in electronic nose studies
| Strengths | Weaknesses | |
|---|---|---|
| • Make no assumption about underlying data distribution | • Does not produce a model, limiting the ability to understand how the features are related to the class | |
| • If there are more samples of one class than other class, the dominant class will control the classification and cause wrong classification | ||
| Naive Bayes | • Requires relatively few examples for training | • Relies on an often-faulty assumption of equally important and independent features |
| • Not ideal for datasets with many numeric features | ||
| Decision tree | • Can be used on small dataset | • It is easy to overfit or underfit the model |
| • Model is easy to interpret | • Small changes in the training data can result in large changes to decision logic | |
| Neural network | • Conceptually similar to human neural function | • Very prone to overfitting training data |
| • Capable of modeling more complex patterns | • Susceptible to multicollinearity | |
| Support vector machines | • High accuracy but not overly influenced by noisy data and not very prone to overfitting | • Finding the best model requires testing of various combinations of kernels and model parameters |
| • Easier for users due to the existence of several well-supported SVM algorithms | ||
| • Most commonly used | ||
| Random forest | • Can handle noisy or missing data | • The model is not easily interpretable |
| • Suitable for class imbalance problems |
Summarized from [27, 53, 54]