| Literature DB >> 28304378 |
Nino Arsov1, Martin Pavlovski1, Lasko Basnarkov1,2, Ljupco Kocarev1,2,3.
Abstract
Ensemble generation is a natural and convenient way of achieving better generalization performance of learning algorithms by gathering their predictive capabilities. Here, we nurture the idea of ensemble-based learning by combining bagging and boosting for the purpose of binary classification. Since the former improves stability through variance reduction, while the latter ameliorates overfitting, the outcome of a multi-model that combines both strives toward a comprehensive net-balancing of the bias-variance trade-off. To further improve this, we alter the bagged-boosting scheme by introducing collaboration between the multi-model's constituent learners at various levels. This novel stability-guided classification scheme is delivered in two flavours: during or after the boosting process. Applied among a crowd of Gentle Boost ensembles, the ability of the two suggested algorithms to generalize is inspected by comparing them against Subbagging and Gentle Boost on various real-world datasets. In both cases, our models obtained a 40% generalization error decrease. But their true ability to capture details in data was revealed through their application for protein detection in texture analysis of gel electrophoresis images. They achieve improved performance of approximately 0.9773 AUROC when compared to the AUROC of 0.9574 obtained by an SVM based on recursive feature elimination.Entities:
Mesh:
Year: 2017 PMID: 28304378 PMCID: PMC5356335 DOI: 10.1038/srep44649
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Notations used throughout the text (in order of appearance).
| Notation | Meaning |
|---|---|
| Data instance, i.e., input | |
| Class label, i.e., output | |
| Instance-label pair | |
| Set of instance-label pairs | |
| Number of ensemble members | |
| Number of boosting rounds | |
| Hypotheses, i.e., model outputs | |
| Loss function | |
| Probability distribution | |
| True (generalization) error | |
| Empirical (observed) error | |
| Algorithmic stability measure | |
| Data subset size as a fraction | |
| Probability | |
| Expected value | |
| Iterator used in the collaboration context | |
| Collaboration probability at each boosting round | |
| Number of instances to be exchanged during one collaboration |
Figure 1Graphical representation of the bi-level collaborative ensemble of learning machines.
Being able to solve a variety of real-world problems, it automatically employs artificial collaboration (shown in orange) within two distinct levels - between the weak components of the strong machines (W-CLB, i.e., left-hand side) or between the strong machines themselves (S-CLB, i.e., right-hand side). Strong machines are constructed via the Gentle Boost algorithm, using an automatically selected subset of available domain data (shown in green and deployed by a data sampling strategy). Machine learning algorithms (specified in SI) are used by inducers to generate predictive models. W-CLB operates during boosting, while S-CLB - employed afterwards - retrains the boosting ensembles after each successful data exchange (shown by the blue dashed arrows). W-CLB and S-CLB strive to improve the constituent models’ algorithmic stability, which in turn accounts for improved performance upon integrating them into an ensemble (shown in red). The theoretical definitions of the collaborative channels (shown in grey) have been carefully designed to promote parallelization by centralizing the input data source, resulting in time-decoupled ensemble members, making them highly applicable to prodigious learning tasks.
Minimal generalization error rates in per cent.
| Dataset | Subbagging | Gentle Boost | W-CLB | S-CLB |
|---|---|---|---|---|
| Australian | 16.5468 | 12.9496 | 10.0719 | 10.7914 |
| Breast Cancer | 3.9286 | 3.9286 | 1.7857 | 1.7857 |
| Diabetes | 20.7792 | 24.0260 | 18.8310 | 18.1818 |
| Heart | 16.6667 | 18.5185 | 11.1111 | 11.1111 |
| Ionosphere | 8.4906 | 7.5472 | 5.6604 | 5.6604 |
| Liver Disorders | 20.2899 | 18.8406 | 14.4928 | 14.4928 |
| Lung Cancer | 2.6846 | 18.1208 | 0.6711 | 0.6711 |
| Mammographic | 15.0259 | 16.5803 | 14.5078 | 13.9896 |
| Vote | 3.4091 | 2.2727 | 1.1364 | 1.1364 |
Decrease of the minimal generalization error by W-CLB and S-CLB, compared to Subbagging and Gentle Boost, respectively.
| Dataset | Subbagging vs. W-CLB | Gentle Boost vs. W-CLB | Subbagging vs. S-CLB | Gentle Boost vs. S-CLB |
|---|---|---|---|---|
| Australian | 39.13% | 22.22% | 34.78% | 16.67% |
| Breast Cancer | 54.55% | 54.55% | 54.55% | 54.55% |
| Diabetes | 9.38% | 21.62% | 12.50% | 24.32% |
| Heart | 33.33% | 40.00% | 33.33% | 40.00% |
| Ionosphere | 33.33% | 25.00% | 33.33% | 25.00% |
| Liver Disorders | 28.57% | 23.08% | 28.57% | 23.08% |
| Lung Cancer | 75.00% | 96.30% | 75.00% | 96.30% |
| Mammographic | 3.45% | 12.50% | 6.90% | 15.63% |
| Vote | 66.67% | 50.00% | 66.67% | 50.00% |
Total number of successful tentative collaborations in W-CLB and S-CLB until the minimal generalization error has been reached.
| Dataset | W-CLB | S-CLB |
|---|---|---|
| Australian | 50 | 284 |
| Breast Cancer | 25 | 100 |
| Diabetes | 100 | 549 |
| Heart | 50 | 155 |
| Ionosphere | 15 | 107 |
| Liver Disorders | 5 | 390 |
| Lung Cancer | 2 | 149 |
| Mammographic | 120 | 144 |
| Vote | 83 | 337 |
The total prospective number of collaborations for W-CLB is Tpn (a successful collaboration will assure that all S weak learners have successfully exchanged an instance within their own training set), while this number for S-CLB is equal to S(S − 1)n/2.
Figure 2AUROC - a comparison to texture analysis in gel electrophoresis images.
The line graphs show the AUROC, ranging from 0 to 1, for 10-Fold CV. The grey lines refer to the testing sets for each CV round, while the blue line is their mean. Furthermore, the red dashed line represents the highest AUROC value reported in ref. 12. The balloons on the mean curve (blue) show the number of successful collaborations throughout the model’s training in both cases, when W-CLB is used with probability p of 0.1 (left), as well as when S-CLB is the collaboration choice (right). The average AUROC over the 10 folds is 0.9758 after 30 rounds of boosting for W-CLB and 0.9789 upon adding 10 boosting ensembles for S-CLB. The best average AUROC reported in ref. 12 is 0.9574 for SVM-RFE (shown by the red dashed line), which substantiates the generalization effectiveness of our collaborative approaches.