| Literature DB >> 34337434 |
Amir Farzad1, T Aaron Gulliver1.
Abstract
Log messages are widely used in cloud servers and other systems. Millions of logs are generated each day which makes them important for anomaly detection. However, they are complex unstructured text messages which makes this task difficult. In this paper, a hybrid log message anomaly detection technique is proposed which employs pruning of positive and negative logs. Reliable positive log messages are first selected using a Gaussian mixture model algorithm. Then reliable negative logs are selected using the K-means, Gaussian mixture model and Dirichlet process Gaussian mixture model methods iteratively. It is shown that the precision for positive and negative logs with pruning is high. Anomaly detection is done using a deep learning long short-term memory network. The proposed model is evaluated using the well-known BGL, Openstack, and Thunderbird data sets. The results obtained indicate that the proposed model performs better than several well-known algorithms.Entities:
Keywords: Anomaly detection; Deep learning; Hybrid learning; Log messages
Year: 2021 PMID: 34337434 PMCID: PMC8310418 DOI: 10.1007/s42979-021-00772-9
Source DB: PubMed Journal: SN Comput Sci ISSN: 2661-8907
Fig. 1An example of a log message consisting of time stamp, verbosity level and raw content
Fig. 2Examples of positive and negative log messages from Openstack
Fig. 3A block of an LSTM network with input, input, output and forget gates [19]
Fig. 4The proposed model architecture
Fig. 5The data preparation for Algorithm 3
The proposed model testing accuracy, precision, recall, F-measure, and average time with (a) GMM for positive log pruning and (b) BGM for positive log pruning
| Data set | Testing accuracy | Label | Precision | Recall | F-measure | Time (s) |
|---|---|---|---|---|---|---|
| (a) | ||||||
| BGL | 99.3%-( | 0 | 93.2%-( | 97.1%-( | 95.6%-( | 3725 |
| 1 | 99.8%-(99.8%)-99.9% | 99.4%-(99.6%)-99.8% | 99.6%-(99.7%)-99.8% | |||
| Openstack | 99.8%-( | 0 | 99.3%-( | 97.9%-( | 99.0%-( | 177 |
| 1 | 99.7%-(99.9%)-100% | 99.9%-(99.9%)-100% | 99.9%-(99.9%)-100% | |||
| Thunderbird | 99.6%-( | 0 | 97.1%-( | 99.6%-( | 98.3%-( | 3550 |
| 1 | 99.9%-(99.9%)-99.9% | 99.5%-(99.8%)-99.9% | 99.7%-(99.9%)-99.9% | |||
| (b) | ||||||
| BGL | 99.4%-( | 0 | 94.0%-( | 97.2%-( | 96.0%-( | 3649 |
| 1 | 99.8%-(99.9%)-99.9% | 99.5%-(99.6%)-99.8% | 99.7%-(99.7%)-99.8% | |||
| Openstack | 99.6%-( | 0 | 96.7%-( | 99.6%-( | 98.3%-( | 134 |
| 1 | 99.9%-(99.9%)-100% | 99.6%-(99.9%)-100% | 99.8%-(99.9%)-100% | |||
| Thunderbird | 99.7%-( | 0 | 97.5%-( | 99.6%-( | 98.6%-( | 3136 |
| 1 | 99.9%-(99.9%)-99.9% | 99.7%-(99.8%)-99.9% | 99.8%-(99.9%)-99.9% | |||
The minimum, maximum and average (in parenthesis) values are given for 10 runs with the BGL, Openstack and Thunderbird data sets. Positive labels are denoted by 1 and negative labels by 0
The precision, recall and F-measure for negative logs for (a) BGL, (b) Openstack and (c) Thunderbird data sets with the Auto-LSTM, IKNN, nLSALog and Deeplog algorithms
| Algorithm | Precision (%) | Recall (%) | F-measure (%) |
|---|---|---|---|
| (a) | |||
| Auto-LSTM | 98.0 | 91.3 | 94.5 |
| IKNN | 92.0 | 91.0 | 92.0 |
| nLSALog | 82.5 | 94.7 | 88.2 |
| (b) | |||
| Auto-LSTM | 99.4 | 92.8 | 96.0 |
| Deeplog | 94.0 | 99.0 | 97.0 |
| (c) | |||
| Auto-LSTM | 98.4 | 99.8 | 99.1 |
| IKNN | 96.0 | 96.0 | 96.0 |
Average testing accuracy, precision, recall, F-measure, and time for (a) BGL, (b) Openstack and (c) Thunderbird data sets with the BGM, EEnvelope, GMM, K-means, LOF and OC-SVM methods using 10-fold cross-validation
| Algorithm | Testing accuracy | Label | Precision | Recall | F-measure | Time (s) |
|---|---|---|---|---|---|---|
| (a) | ||||||
| BGM | 50.3% | 0 | 37.8% | 50.0% | 43.0% | 2632 |
| 1 | 63.9% | 50.3% | 52.2% | |||
| EEnvelope | 86.2% | 0 | 17.6% | 23.9% | 20.3% | 5287 |
| 1 | 93.8% | 91.1% | 92.4% | |||
| GMM | 50.3% | 0 | 38.2% | 50.0% | 43.3% | 1894 |
| 1 | 63.9% | 50.4% | 52.2% | |||
| K-means | 50.0% | 0 | 6.6% | 50.0% | 11.7% | 7759 |
| 1 | 93.4% | 50.0% | 64.9% | |||
| LOF | 83.6% | 0 | 7.1% | 10.3% | 8.4% | 605 |
| 1 | 92.6% | 89.4% | 91.0% | |||
| OC-SVM | 84.3% | 0 | 8.5% | 11.4% | 9.7% | 28469 |
| 1 | 92.8% | 90.1% | 91.4% | |||
| (b) | ||||||
| BGM | 50.3% | 0 | 17.0% | 50.0% | 21.9% | 90 |
| 1 | 82.9% | 50.3% | 59.8% | |||
| EEnvelope | 88.8% | 0 | 53.4% | 44.9% | 48.8% | 244 |
| 1 | 92.7% | 94.7% | 93.7% | |||
| GMM | 48.7% | 0 | 35.2% | 60.0% | 38.9% | 77 |
| 1 | 68.2% | 47.1% | 53.0% | |||
| K-means | 30.0% | 0 | 30.0% | 30.0% | 30.0% | 208 |
| 1 | 30.0% | 30.0% | 30.0% | |||
| LOF | 80.3% | 0 | 14.1% | 13.1% | 13.6% | 985 |
| 1 | 88.4% | 89.3% | 88.9% | |||
| OC-SVM | 38.5% | 0 | 0.3% | 1.3% | 0.5% | 13502 |
| 1 | 76.6% | 43.5% | 55.5% | |||
| (c) | ||||||
| BGM | 57.9% | 0 | 25.9% | 60.0% | 35.2% | 1901 |
| 1 | 83.6% | 57.6% | 63.4% | |||
| EEnvelope | 82.1% | 0 | 43.8% | 26.2% | 32.8% | 3872 |
| 1 | 86.3% | 93.3% | 89.6% | |||
| GMM | 66.2% | 0 | 27.1% | 70.0% | 37.2% | 760 |
| 1 | 92.8% | 65.8% | 71.5% | |||
| K-means | 50.0% | 0 | 5.9% | 50.0% | 10.6% | 4787 |
| 1 | 94.1% | 50.0% | 58.3% | |||
| LOF | 75.4% | 0 | 17.5% | 10.3% | 12.9% | 1141 |
| 1 | 82.2% | 89.5% | 85.7% | |||
| OC-SVM | 40.6% | 0 | 8.5% | 23.7% | 12.5% | 20353 |
| 1 | 72.8% | 44.3% | 55.1% | |||
Positive labels are denoted by 1 and negative labels by 0
Positive pruning testing accuracy, precision, recall, and F-measure with (a) GMM and (b) BGM for the BGL, Openstack and Thunderbird data sets
| Data set | Testing accuracy | Label | Precision | Recall | F-measure |
|---|---|---|---|---|---|
| (a) | |||||
| BGL | 97.1%-(97.8%)-98.4% | 0 | 71.6%-(76.8%)-81.9% | 99.9%-(99.9%)-100% | 83.4%-(86.9%)-90.0% |
| 1 | 99.9%-( | 96.8%-(97.6%)-98.2% | 98.4%-(98.8%)-99.1% | ||
| Openstack | 99.8%-(99.9%)-100% | 0 | 98.8%-(99.8%)-100% | 99.3%-(99.8%)-100% | 99.1%-(99.8%)-100% |
| 1 | 99.9%-( | 99.8%-(99.9%)-100% | 99.9%-(99.9%)-100% | ||
| Thunderbird | 89.9%-(90.0%)-90.5% | 0 | 49.3%-(50.4%)-59.6% | 99.9%-(99.9%)-100% | 66.0%-(67.0%)-74.7% |
| 1 | 99.9%-( | 88.8%-(88.9%)-89.0% | 94.1%-(94.1%)-94.2% | ||
| (b) | |||||
| BGL | 97.3%-(97.6%)-98.2% | 0 | 73.1%-(75.3%)-80.6% | 99.9%-(99.9%)-100% | 84.5%-(85.9%)-89.2% |
| 1 | 99.9%-( | 97.1%-(97.4%)-98.1% | 98.5%-(98.7%)-99.0% | ||
| Openstack | 99.8%-(99.9%)-99.9% | 0 | 98.5%-(99.4%)-99.9% | 99.6%-(99.8%)-99.9% | 99.1%-(99.6%)-99.9% |
| 1 | 99.9%-( | 99.8%-(99.9%)-99.9% | 99.9%-(99.9%)-99.9% | ||
| Thunderbird | 89.9%-(90.0%)-90.1% | 0 | 49.3%-(49.5%)-49.8% | 99.9%-(99.9%)-100% | 66.0%-(66.2%)-66.5% |
| 1 | 99.9%-( | 88.9%-(88.9%)-89.1% | 94.1%-(94.1%)-94.2% | ||
The minimum, maximum, and average (in parenthesis) values are given for 10 runs. Positive labels are denoted by 1 and negative labels by 0
Negative pruning testing accuracy, precision, recall, and F-measure for the BGL data set with (a) GMM, (b) K-means and (c) BGM methods for five rounds (with a GMM for positive pruning)
| Round | Testing accuracy | Label | Precision | Recall | F-measure |
|---|---|---|---|---|---|
| (a) | |||||
| 1 | 76.2%-(82.6%)-88.9% | 0 | 75.0%-(82.4%)-100% | 81.5%-(99.0%)-100% | 85.7%-(89.7%)-93.1% |
| 1 | 55.8%-(97.4%)-100% | 16.1%-(29.1%)-100% | 27.8%-(40.9%)-71.9% | ||
| 2 | 85.4%-(96.0%)-99.2% | 0 | 99.9%-(99.9%)-100% | 81.5%-(94.9%)-99.0% | 89.8%-(97.3%)-99.5% |
| 1 | 58.6%-(86.2%)-97.1% | 99.9%-(99.9%)-100% | 73.9%-(91.8%)-98.5% | ||
| 3 | 81.9%-(90.6%)-99.5% | 0 | 99.3%-(99.9%)-100% | 81.0%-(90.2%)-99.5% | 89.5%-(94.6%)-99.7% |
| 1 | 19.7%-(47.3%)-89.8% | 86.6%-(99.3%)-100% | 32.9%-(59.1%)-94.6% | ||
| 4 | 81.9%-(91.1%)-99.5% | 0 | 99.9%-(100%)-100% | 81.0%-(90.7%)-99.5% | 89.5%-(94.9%)-99.7% |
| 1 | 19.4%-(50.9%)-90.1% | 99.9%-(99.9%)-100% | 32.5%-(62.1%)-94.8% | ||
| 5 | 81.9%-(88.7%)-99.0% | 0 | 100%-( | 81.0%-(88.2%)-98.9% | 89.5%-(93.5%)-99.5% |
| 1 | 19.6%-(43.7%)-82.3% | 100%-(100%)-100% | 32.7%-(55.1%)-90.3% | ||
| (b) | |||||
| 1 | 76.1%-(81.8%)-87.1% | 0 | 75.0%-(80.8%)-86.4% | 99.9%-(99.9%)-100% | 85.7%-(89.4%)-92.7% |
| 1 | 99.9%-(99.9%)-100% | 16.0%-(21.7%)-28.9% | 27.6%-(35.5%)-44.8% | ||
| 2 | 96.0%-(96.2%)-96.3% | 0 | 95.3%-(95.5%)-95.7% | 99.9%-(99.9%)-100% | 97.6%-(97.7%)-97.8% |
| 1 | 99.9%-(99.9%)-100% | 72.6%-(80.8%)-85.3% | 84.1%-(89.3%)-92.1% | ||
| 3 | 86.9%-(87.7%)-91.1% | 0 | 94.9%-(95.8%)-99.9% | 91.1%-(91.2%)-91.2% | 93.0%-(93.4%)-95.3% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
| 4 | 86.9%-(87.5%)-90.8% | 0 | 94.9%-(95.6%)-99.9% | 90.8%-(91.1%)-91.2% | 93.0%-(93.3%)-95.1% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
| 5 | 87.0%-(87.7%)-90.8% | 0 | 95.0%-( | 90.8%-(91.1%)-91.2% | 93.0%-(93.4%)-95.1% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
| (c) | |||||
| 1 | 76.2%-(88.2%)-96.1% | 0 | 75.0%-(88.1%)-100% | 81.5%-(98.4%)-100% | 85.7%-(92.7%)-97.6% |
| 1 | 55.8%-(94.3%)-100% | 16.1%-(57.5%)-100% | 27.8%-(67.5%)-89.7% | ||
| 2 | 85.4%-(96.0%)-99.2% | 0 | 99.9%-(99.9%)-100% | 81.5%-(94.9%)-99.0% | 89.8%-(97.3%)-99.5% |
| 1 | 58.6%-(86.2%)-97.1% | 99.9%-(99.9%)-100% | 73.9%-(91.8%)-98.5% | ||
| 3 | 81.9%-(90.6%)-99.5% | 0 | 99.3%-(99.9%)-100% | 81.0%-(90.2%)-99.5% | 89.5%-(94.6%)-99.7% |
| 1 | 19.7%-(47.3%)-89.8% | 86.6%-(99.3%)-100% | 32.9%-(59.1%)-94.6% | ||
| 4 | 67.3%-(89.8%)-99.5% | 0 | 99.3%-(99.9%)-100% | 67.3%-(89.4%)-99.5% | 80.4%-(94.1%)-99.7% |
| 1 | 0.1%-(47.2%)-90.1% | 86.6%-(99.3%)-100% | 0.1%-(57.9%)-94.8% | ||
| 5 | 81.9%-(88.7%)-99.0% | 0 | 100%-( | 81.0%-(88.2%)-98.9% | 89.5%-(93.5%)-99.5% |
| 1 | 19.6%-(43.7%)-82.3% | 100%-(100%)-100% | 32.7%-(55.1%)-90.3% | ||
The minimum, maximum and average (in parenthesis) values are given for 10 runs. Positive labels are denoted by 1 and negative labels by 0
Negative pruning testing accuracy, precision, recall, and F-measure for the Openstack data set with (a) GMM, (b) K-means and (c) BGM methods for five rounds (with a GMM for positive pruning)
| Round | Testing accuracy | Label | Precision | Recall | F-measure |
|---|---|---|---|---|---|
| (a) | |||||
| 1 | 61.5%-(64.3%)-94.0% | 0 | 99.1%-(99.9%)-100% | 61.5%-(64.3%)-94.0% | 76.2%-(78.0%)-96.9% |
| 1 | 0.0%-(0.1%)-1.6% | 0.0%-(4.2%)-52.8% | 0.0%-(0.2%)-3.1% | ||
| 2 | 61.5%-(83.9%)-94.0% | 0 | 100%-(100%)-100% | 61.5%-(83.8%)-94.0% | 76.2%-(90.7%)-96.9% |
| 1 | 0.0%-(0.7%)-9.3% | 0.0%-(7.5%)-100% | 0.0%-(1.3%)-17.1% | ||
| 3 | 64.9%-(88.8%)-98.3% | 0 | 100%-(100%)-100% | 64.9%-(88.8%)-98.3% | 78.7%-(93.8%)-99.1% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
| 4 | 71.6%-(92.0%)-98.2% | 0 | 100%-(100%)-100% | 71.6%-(92.0%)-98.2% | 83.4%-(95.7%)-99.1% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
| 5 | 58.7%-(88.6%)-98.2% | 0 | 100%-( | 58.7%-(88.6%)-98.2% | 74.0%-(93.6%)-99.1% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
| (b) | |||||
| 1 | 62.7%-(62.8%)-62.9% | 0 | 99.0%-(99.9%)-100% | 62.7%-(62.8%)-62.9% | 76.9%-(77.1%)-77.3% |
| 1 | 0.0%-(0.1%)-1.4% | 0.0%-(4.5%)-46.7% | 0.0%-(0.3%)-2.8% | ||
| 2 | 62.8%-(84.1%)-90.4% | 0 | 100%-(100%)-100% | 62.8%-(84.1%)-90.4% | 77.1%-(90.9%)-95.0% |
| 1 | 0.0%-(1.0%)-9.3% | 0.0%-(11.4%)-100% | 0.0%-(1.9%)-17.1% | ||
| 3 | 90.4%-(91.6%)-94.6% | 0 | 100%-(100%)-100% | 90.4%-(91.6%)-94.6% | 95.0%-(95.6%)-97.2% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
| 4 | 91.4%-(93.2%)-94.6% | 0 | 100%-(100%)-100% | 91.4%-(93.2%)-94.6% | 95.5%-(96.5%)-97.2% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
| 5 | 78.6%-(91.8%)-94.6% | 0 | 100%-( | 78.6%-(91.8%)-94.6% | 88.0%-(95.7%)-97.2% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
| (c) | |||||
| 1 | 61.5%-(64.3%)-94.0% | 0 | 99.1%-(99.9%)-100% | 61.5%-(64.3%)-94.0% | 76.1%-(78.0%)-96.9% |
| 1 | 0.0%-(0.1%)-1.6% | 0.0%-(4.3%)-53.8% | 0.0%-(0.2%)-3.2% | ||
| 2 | 61.5%-(82.0%)-90.4% | 0 | 100%-(100%)-100% | 61.5%-(82.0%)-90.4% | 76.1%-(89.6%)-95.0% |
| 1 | 0.0%-(0.8%)-9.3% | 0.0%-(8.8%)-100% | 0.0%-(1.5%)-17.1% | ||
| 3 | 53.6%-(83.3%)-98.3% | 0 | 100%-(100%)-100% | 53.6%-(83.3%)-98.3% | 69.8%-(90.1%)-99.1% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
| 4 | 71.6%-(89.8%)-98.2% | 0 | 100%-(100%)-100% | 71.6%-(89.8%)-98.2% | 83.4%-(94.5%)-99.1% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
| 5 | 54.4%-(84.8%)-98.2% | 0 | 100%-( | 54.4%-(84.8%)-98.2% | 70.5%-(91.1%)-99.1% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
The minimum, maximum and average (in parenthesis) values are given for 10 runs. Positive labels are denoted by 1 and negative labels by 0
Negative pruning testing accuracy, precision, recall, and F-measure for the Thunderbird data set with (a) GMM, (b) K-means and (c) BGM methods for five rounds (with a GMM for positive pruning)
| Round | Testing accuracy | Label | Precision | Recall | F-measure |
|---|---|---|---|---|---|
| (a) | |||||
| 1 | 99.7%-(99.8%)-99.8% | 0 | 99.9%-(99.9%)-100% | 99.6%-(99.6%)-99.6% | 99.8%-(99.8%)-99.8% |
| 1 | 99.4%-(99.6%)-99.6% | 99.9%-(99.9%)-100% | 99.7%-(99.8%)-99.8% | ||
| 2 | 84.6%-(88.1%)-88.4% | 0 | 99.9%-(99.9%)-100% | 84.6%-(88.1%)-88.4% | 91.7%-(93.7%)-93.9% |
| 1 | 0.0%-(0.1%)-0.1% | 0.0%-(10.0%)-100% | 0.0%-(0.1%)-0.1% | ||
| 3 | 90.0%-(90.6%)-90.7% | 0 | 99.9%-(99.9%)-100% | 90.0%-(90.6%)-90.7% | 94.7%-(95.1%)-95.1% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
| 4 | 45.2%-(82.2%)-90.7% | 0 | 99.9%-(99.9%)-100% | 45.2%-(82.2%)-90.7% | 62.3%-(89.0%)-95.1% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
| 5 | 45.2%-(83.0%)-90.7% | 0 | 99.9%-( | 45.2%-(83.0%)-90.7% | 62.3%-(89.9%)-95.1% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
| (b) | |||||
| 1 | 70.9%-(78.0%)-99.8% | 0 | 100%-(100%)-100% | 51.2%-(57.0%)-99.6% | 67.7%-(72.1%)-99.8% |
| 1 | 58.1%-(69.4%)-99.6% | 100%-(100%)-100% | 73.5%-(81.7%)-99.8% | ||
| 2 | 35.0%-(35.3%)-35.8% | 0 | 100%-(100%)-100% | 35.0%-(35.3%)-35.8% | 51.8%-(52.2%)-52.7% |
| 1 | 0.0%-(0.1%)-0.1% | 0.0%-(17.4%)-100% | 0.0%-(0.1%)-0.1% | ||
| 3 | 36.1%-(36.5%)-37.9% | 0 | 100%-(100%)-100% | 36.1%-(36.5%)-37.9% | 53.0%-(53.4%)-55.0% |
| 1 | 0.0%-(0.1%)-0.1% | 0.0%-(5.9%)-100% | 0.0%-(0.1%)-0.1% | ||
| 4 | 36.1%-(36.6%)-38.0% | 0 | 100%-(100%)-100% | 36.1%-(36.6%)-38.0% | 53.0%-(53.6%)-55.0% |
| 1 | 0.0%-(0.1%)-0.1% | 0.0%-(5.5%)-100% | 0.0%-(0.1%)-0.1% | ||
| 5 | 29.7%-(38.2%)-61.2% | 0 | 100%-( | 29.7%-(38.2%)-61.2% | 45.8%-(54.9%)-76.0% |
| 1 | 0.0%-(0.1%)-0.1% | 0.0%-(4.5%)-100% | 0.0%-(0.1%)-0.1% | ||
| (c) | |||||
| 1 | 99.7%-(99.8%)-99.8% | 0 | 99.9%-(99.9%)-100% | 99.6%-(99.6%)-99.6% | 99.8%-(99.8%)-99.8% |
| 1 | 99.4%-(99.6%)-99.6% | 99.9%-(99.9%)-100% | 99.7%-(99.8%)-99.8% | ||
| 2 | 84.6%-(88.1%)-88.4% | 0 | 99.9%-(99.9%)-100% | 84.6%-(88.1%)-88.4% | 91.7%-(93.7%)-93.9% |
| 1 | 0.0%-(0.1%)-0.1% | 0.0%-(4.7%)-40.0% | 0.0%-(0.1%)-0.1% | ||
| 3 | 90.0%-(90.6%)-90.7% | 0 | 99.9%-(99.9%)-100% | 90.0%-(90.6%)-90.7% | 94.7%-(95.1%)-95.1% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
| 4 | 45.2%-(83.0%)-90.7% | 0 | 99.9%-(99.9%)-100% | 45.2%-(83.0%)-90.7% | 62.3%-(89.6%)-95.1% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
| 5 | 45.2%-(83.7%)-90.7% | 0 | 99.9%-( | 45.2%-(83.7%)-90.7% | 62.3%-(90.4%)-95.1% |
| 1 | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | 0.0%-(0.0%)-0.0% | ||
The minimum, maximum and average (in parenthesis) values are given for 10 runs. Positive labels are denoted by 1 and negative labels by 0