| Literature DB >> 27835638 |
Qing-Hua Ling1,2, Yu-Qing Song1, Fei Han1, Dan Yang1, De-Shuang Huang3.
Abstract
For ensemble learning, how to select and combine the candidate classifiers are two key issues which influence the performance of the ensemble system dramatically. Random vector functional link networks (RVFL) without direct input-to-output links is one of suitable base-classifiers for ensemble systems because of its fast learning speed, simple structure and good generalization performance. In this paper, to obtain a more compact ensemble system with improved convergence performance, an improved ensemble of RVFL based on attractive and repulsive particle swarm optimization (ARPSO) with double optimization strategy is proposed. In the proposed method, ARPSO is applied to select and combine the candidate RVFL. As for using ARPSO to select the optimal base RVFL, ARPSO considers both the convergence accuracy on the validation data and the diversity of the candidate ensemble system to build the RVFL ensembles. In the process of combining RVFL, the ensemble weights corresponding to the base RVFL are initialized by the minimum norm least-square method and then further optimized by ARPSO. Finally, a few redundant RVFL is pruned, and thus the more compact ensemble of RVFL is obtained. Moreover, in this paper, theoretical analysis and justification on how to prune the base classifiers on classification problem is presented, and a simple and practically feasible strategy for pruning redundant base classifiers on both classification and regression problems is proposed. Since the double optimization is performed on the basis of the single optimization, the ensemble of RVFL built by the proposed method outperforms that built by some single optimization methods. Experiment results on function approximation and classification problems verify that the proposed method could improve its convergence accuracy as well as reduce the complexity of the ensemble system.Entities:
Mesh:
Year: 2016 PMID: 27835638 PMCID: PMC5106042 DOI: 10.1371/journal.pone.0165803
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Abbreviation comparison table.
| Abbreviation | Paraphrase | Abbreviation | Paraphrase |
|---|---|---|---|
| NNE | neural network ensemble | BP | backpropagation |
| RBF | radial basis function | RVFL | random vector functional link networks |
| SLFN | single hidden layered feedforward neural networks | RWSLFN | SLFN with random weights |
| ELM | extreme learning machine | E-ELM | an ensemble of ELM proposed in [ |
| EOS-ELM | an ensemble of online sequential ELM | LSTD-eELM | an ensemble of ELM proposed in [ |
| RMSE-ELM | an ensemble of ELM proposed in [ | GA | genetic algorithm |
| PSO | particle swarm optimization | GASEN | GA based selective ensembles |
| EE-ELM | an ensemble of ELM proposed in [ | PSOSEN | PSO based selective NNE |
| ARPSO | attractive and repulsive PSO | E-ARPSOELM | an ensemble of ELM proposed in [ |
| DGEELMBARPSO | an ensemble of ELM proposed in [ | MP | Moore-Penrose |
| LS | least square | APSO | adaptive PSO |
| DO-EELM | an ensemble of ELM based on double optimization | RMSE | root mean squared error |
| E-PSOELM | an ensemble of ELM based on PSO | SO-EELM | an ensemble of ELM proposed in [ |
Fig 1The frame of the DO-EELM method.
The results of approximating the Sinc function by the seven algorithms.
| Algorithms | Train RMSE | Test RMSE±Std. | Mean size of ELM ensemble |
|---|---|---|---|
| SVM | 0.1149 | 0.0130±0.0012 | / |
| E-ELM | 0.1157 | 0.0166±6.7086e-04 | 12 |
| E-OSELM | 0.1163 | 0.0167±6.3027e-04 | 12 |
| E-PSOELM | 0.1164 | 0.0163±8.6240e-04 | 10.3 |
| E-ARPSOELM | 0.1153 | 0.0161±8.6240e-04 | 9.7 |
| SO-EELM | 0.1161 | 0.0133±6.8036e-04 | 9.6 |
| DO-EELM | 0.1156 | 0.0113±5.7065e-04 | 6.83 |
Fig 2The diversity curves of different ensembles of ELM on approximating the SinC function with 20 independent runs.
Fig 3The number of the ELM in the four PSO based ensembles of ELM on approximating the SinC function with 20 independent runs.
The specifications of the six datasets.
| Dataset | Train set | Test set | Categories | Attributes |
|---|---|---|---|---|
| Diabetes | 576 | 192 | 2 | 8 |
| Satellite Image | 4435 | 2000 | 6 | 36 |
| Wine | 120 | 58 | 3 | 13 |
| Image Segmentation | 1500 | 810 | 7 | 19 |
| Brain cancer | 41 | 19 | 2 | 7129 |
| Lung | 140 | 63 | 5 | 3312 |
Classification results of the seven algorithms on the six data.
| Data | Algorithms | Train accuracy | Test accuracy±Std. | Mean size of ELM ensemble |
|---|---|---|---|---|
| Diabetes | SVM | 0.7807 | 0.7747±0.0252 | / |
| E-ELM | 0.7886 | 0.8271±0.0112 | 12 | |
| EOS-ELM | 0.7877 | 0.8279±0.0109 | 12 | |
| E-PSOELM | 0.8176 | 0.8316±0.0085 | 12.35 | |
| E-ARPSOELM | 0.8223 | 0.8359±0.0077 | 12.7 | |
| SO-EELM | 0.8147 | 0.8367±0.0073 | 11.05 | |
| DO-EELM | 0.8256 | 0.8536±0.0055 | 9.6 | |
| Satellite Image | SVM | 0.8825 | 0.8689±0.0035 | / |
| E-ELM | 0.9227 | 0.8927±0.0031 | 12 | |
| EOS-ELM | 0.9232 | 0.8926±0.0028 | 12 | |
| E-PSOELM | 0.9311 | 0.8936±0.0031 | 14.6 | |
| E-ARPSOELM | 0.9316 | 0.8976±0.0023 | 7.9 | |
| SO-EELM | 0.9279 | 0.9006±0.0023 | 6.5 | |
| DO-EELM | 0.9335 | 0.9018±0.0022 | 6 | |
| Wine | SVM | 0.9973 | 0.9529±0.0258 | / |
| E-ELM | 0.9875 | 0.9819±0.0163 | 12 | |
| EOS-ELM | 0.9975 | 0.9886±0.0087 | 12 | |
| E-PSOELM | 0.9997 | 0.9888±0.0083 | 15 | |
| E-ARPSOELM | 1 | 0.9914±0.0088 | 15.55 | |
| SO-EELM | 1 | 0.9936±0.0087 | 8.8 | |
| DO-EELM | 1 | 1±0 | 5.2 | |
| Image Segmentation | SVM | 0.9393 | 0.9336±0.0081 | / |
| E-ELM | 0.9722 | 0.9496±0.0040 | 12 | |
| EOS-ELM | 0.9736 | 0.9523±0.0035 | 12 | |
| E-PSOELM | 0.9772 | 0.9530±0.0030 | 14.85 | |
| E-ARPSOELM | 0.9828 | 0.9562±0.0030 | 12.7 | |
| SO-EELM | 0.9819 | 0.9575±0.0032 | 11.6 | |
| DO-EELM | 0.9822 | 0.9665±0.0028 | 9.3 | |
| Brain cancer | SVM | 0.8817 | 0.8368±0.0449 | / |
| E-ELM | 0.9853 | 0.7336±0.0290 | 12 | |
| EOS-ELM | 0.9676 | 0.7368±0.0301 | 12 | |
| E-PSOELM | 0.9769 | 0.7368±0.0273 | 13.35 | |
| E-ARPSOELM | 0.9808 | 0.7395±0.0118 | 11.05 | |
| SO-EELM | 0.9786 | 0.7893±0.0207 | 11.65 | |
| DO-EELM | 0.9808 | 0.8737±0.0106 | 8.75 | |
| Lung | SVM | 0.9861 | 0.9548±0.0165 | / |
| E-ELM | 0.9911 | 0.9865±0.0093 | 12 | |
| EOS-ELM | 0.9911 | 0.9873±0.0083 | 12 | |
| E-PSOELM | 0.9896 | 0.9906±0.0087 | 15 | |
| E-ARPSOELM | 0.9876 | 0.9960±0.0109 | 11.35 | |
| SO-EELM | 0.9763 | 0.9921±0.0096 | 10.8 | |
| DO-EELM | 1 | 1±0 | 8.3 |
Fig 4The diversity curves of different ensembles of ELM on the six classification problems with 20 independent runs.
Fig 5The number of the ELM in the four PSO based ensembles of ELM on the six classification problems with 20 independent runs.
Fig 6The convergence accuracy vs the different values of the parameter α in the DO-EELM method.
Fig 7The convergence accuracy vs the different values of the parameter β in the DO-EELM method.
Fig 8The convergence accuracy vs the size of the initial ELM pool in the DO-EELM method.
Fig 9The convergence accuracy of two approaches on the seven data with 20 independent runs.