| Literature DB >> 25295291 |
Mert Bal1, M Fatih Amasyali2, Hayri Sever3, Guven Kose4, Ayse Demirhan5.
Abstract
The importance of the decision support systems is increasingly supporting the decision making process in cases of uncertainty and the lack of information and they are widely used in various fields like engineering, finance, medicine, and so forth, Medical decision support systems help the healthcare personnel to select optimal method during the treatment of the patients. Decision support systems are intelligent software systems that support decision makers on their decisions. The design of decision support systems consists of four main subjects called inference mechanism, knowledge-base, explanation module, and active memory. Inference mechanism constitutes the basis of decision support systems. There are various methods that can be used in these mechanisms approaches. Some of these methods are decision trees, artificial neural networks, statistical methods, rule-based methods, and so forth. In decision support systems, those methods can be used separately or a hybrid system, and also combination of those methods. In this study, synthetic data with 10, 100, 1000, and 2000 records have been produced to reflect the probabilities on the ALARM network. The accuracy of 11 machine learning methods for the inference mechanism of medical decision support system is compared on various data sets.Entities:
Mesh:
Year: 2014 PMID: 25295291 PMCID: PMC4177776 DOI: 10.1155/2014/137896
Source DB: PubMed Journal: ScientificWorldJournal ISSN: 1537-744X
Figure 1The main structure of decision support system.
Figure 2Structure of a multilayer perceptron [7].
Figure 3A simple Naïve-Bayes structure.
Figure 4ALARM network structure and the variables defined in the network [9].
Figure 5Conditional probability diagram for Alarm.dnet catechol variable.
11 Output variables for one record (100 datasets).
| Variable name (disease) | Accuracy degree | Real situations | Results produced by the software | The comparison of the real situation and the result produced |
|---|---|---|---|---|
| History | 0,9900 | False | False | POSITIVE |
| Pres | 0,9412 | Normal | Zero | NEGATIVE |
| MinVol | 0,9136 | Normal | Zero | NEGATIVE |
| ExpCO2 | 0,9136 | Normal | Zero | NEGATIVE |
| PAP | 0,9000 | Normal | Normal | POSITIVE |
| HRBP | 0,8229 | High | High | POSITIVE |
| HREKG | 0,8229 | High | High | POSITIVE |
| HRSat | 0,8229 | High | High | POSITIVE |
| CVP | 0,7075 | Normal | Normal | POSITIVE |
| PCWP | 0,6970 | Normal | Normal | POSITIVE |
| BP | 0,4052 | Low | Low | POSITIVE |
The class labels for 11 classification datasets.
| Dependent variable (class) | Class labels |
|---|---|
| BP | Normal, low, high |
| CVP | Normal, low, high |
| ExpCO2 | Normal, low, high, zero |
| History | False, true |
| HRBP | Normal, low, high |
| HREKG | Normal, low, high |
| HRSat | Normal, low, high |
| MinVol | Normal, low, high, zero |
| PAP | Normal, low, high |
| PCWP | Normal, low, high |
| Press | Normal, low, high, zero |
Used classification algorithms and abbreviations.
| Algorithm name | Abbreviations |
|---|---|
| Zero rule | ZR |
| Naive-Bayes | NB |
| Multilayer perceptron | MLP |
| Simple logistic | SL |
| Support vector machines | SMO |
| One nearest neighbor | IBK |
| C4.5 decision tree | J48 |
| Rep Tree | RT |
| Boosting | BS |
| Bagging | BG |
| Random forest | RF |
Classification accuracies with datasets having 10 samples (%).
| ZR | NB | MLP | SL | SMO | IBK | BS | BG | J48 | RF | RT | |
|---|---|---|---|---|---|---|---|---|---|---|---|
| BP | 50.00 | 30.00 | 6.00 | 0.00 | 42.00 | 10.00 | 50.00 | 38.00 | 50.00 | 8.00 | 22.00 |
| CVP | 70.00 | 60.00 | 80.00 | 60.00 | 60.00 | 80.00 | 80.00 | 66.00 | 60.00 | 80.00 | 70.00 |
| ExpCO2 | 50.00 | 50.00 | 60.00 | 50.00 | 52.00 | 50.00 | 50.00 | 46.00 | 50.00 | 54.00 | 20.00 |
| History | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 |
| HRBP | 70.00 | 80.00 | 48.00 | 70.00 | 70.00 | 70.00 | 40.00 | 70.00 | 70.00 | 62.00 | 70.00 |
| HREKG | 60.00 | 70.00 | 42.00 | 48.00 | 60.00 | 40.00 | 30.00 | 58.00 | 30.00 | 44.00 | 60.00 |
| HRSat | 60.00 | 50.00 | 70.00 | 30.00 | 32.00 | 50.00 | 70.00 | 38.00 | 30.00 | 62.00 | 30.00 |
| MinVol | 70.00 | 70.00 | 80.00 | 70.00 | 72.00 | 70.00 | 80.00 | 70.00 | 70.00 | 72.00 | 70.00 |
| PAP | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 |
| PCWP | 70.00 | 70.00 | 70.00 | 60.00 | 66.00 | 60.00 | 60.00 | 70.00 | 60.00 | 70.00 | 70.00 |
| Press | 70.00 | 70.00 | 80.00 | 70.00 | 70.00 | 70.00 | 80.00 | 70.00 | 70.00 | 70.00 | 70.00 |
Classification accuracies with datasets having 100 samples (%).
| ZR | NB | MLP | SL | SMO | IBK | BS | BG | J48 | RF | RT | |
|---|---|---|---|---|---|---|---|---|---|---|---|
| BP | 47.00 | 47.60 | 45.40 | 44.00 | 45.00 | 41.60 | 45.80 | 43.20 | 42.60 | 45.40 | 44.60 |
| CVP | 62.00 | 92.00 | 88.40 | 91.40 | 91.40 | 91.20 | 92.00 | 92.00 | 92.00 | 90.00 | 92.00 |
| ExpCO2 | 65.00 | 70.80 | 73.20 | 74.00 | 69.20 | 69.00 | 65.00 | 67.60 | 69.20 | 71.60 | 66.20 |
| History | 98.00 | 98.00 | 100.00 | 98.00 | 98.00 | 98.00 | 100.00 | 98.00 | 98.00 | 98.40 | 98.00 |
| HRBP | 49.00 | 59.40 | 55.40 | 57.80 | 59.80 | 55.60 | 53.40 | 57.80 | 58.40 | 56.80 | 54.40 |
| HREKG | 48.00 | 55.60 | 56.80 | 54.80 | 56.80 | 54.20 | 54.00 | 50.80 | 53.20 | 54.60 | 50.20 |
| HRSat | 49.00 | 61.60 | 61.60 | 62.80 | 63.80 | 59.60 | 57.00 | 61.20 | 62.00 | 60.40 | 56.20 |
| MinVol | 79.00 | 84.00 | 86.20 | 84.60 | 85.20 | 82.00 | 79.00 | 79.00 | 82.00 | 83.40 | 79.00 |
| PAP | 88.00 | 88.00 | 86.80 | 88.00 | 88.00 | 88.00 | 88.00 | 88.00 | 88.00 | 87.40 | 88.00 |
| PCWP | 57.00 | 85.00 | 81.00 | 84.80 | 84.20 | 80.20 | 85.00 | 85.00 | 84.60 | 80.00 | 85.00 |
| Press | 80.00 | 85.00 | 89.80 | 87.20 | 85.20 | 84.00 | 80.40 | 82.20 | 85.00 | 85.60 | 81.40 |
Classification accuracies with datasets having 1000 samples (%).
| ZR | NB | MLP | SL | SMO | IBK | BS | BG | J48 | RF | RT | |
|---|---|---|---|---|---|---|---|---|---|---|---|
| BP | 45.00 | 46.32 | 44.52 | 46.44 | 46.58 | 44.58 | 45.00 | 47.18 | 46.52 | 44.84 | 46.60 |
| CVP | 67.70 | 87.64 | 87.00 | 87.72 | 87.08 | 87.14 | 85.60 | 87.70 | 87.70 | 87.06 | 87.68 |
| ExpCO2 | 66.10 | 79.60 | 78.20 | 79.74 | 79.58 | 78.12 | 69.54 | 79.80 | 79.84 | 78.46 | 79.72 |
| History | 94.20 | 98.30 | 98.30 | 98.30 | 98.30 | 98.20 | 98.24 | 98.30 | 98.30 | 98.30 | 98.30 |
| HRBP | 48.00 | 67.12 | 66.18 | 67.20 | 67.56 | 66.48 | 59.20 | 67.06 | 67.04 | 66.62 | 66.88 |
| HREKG | 47.00 | 66.36 | 65.66 | 66.06 | 67.06 | 65.14 | 59.00 | 65.60 | 65.88 | 65.52 | 65.74 |
| HRSat | 47.40 | 65.68 | 65.02 | 65.88 | 66.84 | 64.60 | 59.00 | 65.14 | 65.20 | 64.68 | 65.10 |
| MinVol | 79.30 | 88.84 | 87.80 | 88.82 | 88.76 | 87.30 | 83.20 | 88.86 | 88.82 | 88.18 | 88.88 |
| PAP | 89.40 | 90.20 | 89.48 | 90.20 | 90.20 | 89.38 | 90.20 | 90.20 | 90.20 | 89.50 | 90.16 |
| PCWP | 65.20 | 86.80 | 86.40 | 86.90 | 86.72 | 86.18 | 83.80 | 86.90 | 86.90 | 86.50 | 86.90 |
| Press | 78.40 | 88.86 | 88.46 | 89.34 | 89.18 | 87.90 | 82.90 | 89.26 | 88.98 | 88.70 | 89.24 |
Classification accuracies with datasets having 2000 samples (%).
| ZR | NB | MLP | SL | SMO | IBK | BS | BG | J48 | RF | RT | |
|---|---|---|---|---|---|---|---|---|---|---|---|
| BP | 45.40 | 47.37 | 45.63 | 46.88 | 46.64 | 46.31 | 45.40 | 46.93 | 47.08 | 46.38 | 46.62 |
| CVP | 69.30 | 88.65 | 88.48 | 88.64 | 88.65 | 88.46 | 86.80 | 88.60 | 88.55 | 88.53 | 88.63 |
| ExpCO2 | 66.65 | 80.15 | 79.19 | 80.25 | 80.23 | 78.96 | 70.24 | 79.98 | 80.03 | 79.42 | 79.99 |
| History | 94.15 | 98.45 | 98.40 | 98.43 | 98.45 | 98.12 | 98.42 | 98.45 | 98.45 | 98.43 | 98.45 |
| HRBP | 49.75 | 66.47 | 66.72 | 66.57 | 66.78 | 66.57 | 58.20 | 67.12 | 67.41 | 66.71 | 67.20 |
| HREKG | 48.80 | 65.26 | 63.92 | 65.00 | 64.85 | 64.79 | 57.25 | 64.91 | 64.60 | 65.10 | 64.52 |
| HRSat | 48.95 | 65.05 | 64.41 | 65.03 | 65.58 | 65.40 | 57.40 | 64.90 | 65.26 | 65.51 | 64.95 |
| MinVol | 77.70 | 88.02 | 87.44 | 87.99 | 87.93 | 86.91 | 82.00 | 88.04 | 88.01 | 87.59 | 88.03 |
| PAP | 88.95 | 89.90 | 89.43 | 89.88 | 89.86 | 89.38 | 89.90 | 89.90 | 89.90 | 89.53 | 89.90 |
| PCWP | 65.45 | 86.95 | 86.53 | 86.95 | 86.95 | 86.67 | 83.55 | 86.95 | 86.95 | 86.67 | 86.95 |
| Press | 78.20 | 89.68 | 89.62 | 90.03 | 90.07 | 89.27 | 82.80 | 90.11 | 90.00 | 89.73 | 90.00 |
Figure 6Classification accuracies changes with datasets' sample size.
Pairwise comparison of accuracies (win/loss over 11 datasets) of all algorithms using 10 cv t-Test.
| ZR | NB | MLP | SL | SMO | IBK | BS | BG | J48 | RF | RT | |
|---|---|---|---|---|---|---|---|---|---|---|---|
| ZR | 0/0 | 0/10 | 0/9 | 0/10 | 0/10 | 0/9 | 0/10 | 0/10 | 0/10 | 0/9 | 0/10 |
| NB | 10/0 | 0/0 | 3/0 | 0/0 | 0/0 | 5/0 | 8/0 | 0/0 | 0/1 | 3/0 | 0/0 |
| MLP | 9/0 | 0/3 | 0/0 | 0/3 | 0/3 | 2/0 | 8/1 | 0/4 | 0/5 | 0/0 | 0/4 |
| SL | 10/0 | 0/0 | 3/0 | 0/0 | 0/0 | 6/0 | 8/0 | 0/0 | 0/1 | 3/0 | 0/0 |
| SMO | 10/0 | 0/0 | 3/0 | 0/0 | 0/0 | 6/0 | 8/0 | 0/0 | 0/0 | 3/0 | 0/0 |
| IBK | 9/0 | 0/5 | 0/2 | 0/6 | 0/6 | 0/0 | 8/2 | 0/6 | 0/7 | 0/2 | 0/6 |
| BS | 10/0 | 0/8 | 1/8 | 0/8 | 0/8 | 2/8 | 0/0 | 0/8 | 0/8 | 1/8 | 0/8 |
| BG | 10/0 | 0/0 | 4/0 | 0/0 | 0/0 | 6/0 | 8/0 | 0/0 | 0/0 | 2/0 | 0/0 |
| J48 | 10/0 | 1/0 | 5/0 | 1/0 | 0/0 | 7/0 | 8/0 | 0/0 | 0/0 | 4/0 | 0/0 |
| RF | 9/0 | 0/3 | 0/0 | 0/3 | 0/3 | 2/0 | 8/1 | 0/2 | 0/4 | 0/0 | 0/2 |
| RT | 10/0 | 0/0 | 4/0 | 0/0 | 0/0 | 6/0 | 8/0 | 0/0 | 0/0 | 2/0 | 0/0 |
The average ranks of the algorithms over 11 datasets “and the sum of win/losses.”
| Algorithm name | ZR | NB | MLP | SL | SMO | IBK | BS | BG | J48 | RF | RT |
|---|---|---|---|---|---|---|---|---|---|---|---|
| Average rank | 11 | 4.27 | 8.27 | 4.64 | 3.73 | 8 | 9.27 | 3.64 | 3.55 | 5.9 | 3.73 |
| The number of wins/losses (over 110) | 0/97 | 29/1 | 19/23 | 30/1 | 30/0 | 17/42 | 14/72 | 30/0 | 36/0 | 19/18 | 30/0 |