| Literature DB >> 28546808 |
Adnan O M Abuassba1,2, Dezheng Zhang1,2, Xiong Luo1,2, Ahmad Shaheryar1, Hazrat Ali3.
Abstract
Extreme Learning Machine (ELM) is a fast-learning algorithm for a single-hidden layer feedforward neural network (SLFN). It often has good generalization performance. However, there are chances that it might overfit the training data due to having more hidden nodes than needed. To address the generalization performance, we use a heterogeneous ensemble approach. We propose an Advanced ELM Ensemble (AELME) for classification, which includes Regularized-ELM, L2-norm-optimized ELM (ELML2), and Kernel-ELM. The ensemble is constructed by training a randomly chosen ELM classifier on a subset of training data selected through random resampling. The proposed AELM-Ensemble is evolved by employing an objective function of increasing diversity and accuracy among the final ensemble. Finally, the class label of unseen data is predicted using majority vote approach. Splitting the training data into subsets and incorporation of heterogeneous ELM classifiers result in higher prediction accuracy, better generalization, and a lower number of base classifiers, as compared to other models (Adaboost, Bagging, Dynamic ELM ensemble, data splitting ELM ensemble, and ELM ensemble). The validity of AELME is confirmed through classification on several real-world benchmark datasets.Entities:
Mesh:
Year: 2017 PMID: 28546808 PMCID: PMC5435980 DOI: 10.1155/2017/3405463
Source DB: PubMed Journal: Comput Intell Neurosci
Algorithm 1AELME.
Figure 1The general scheme of the proposed AELME.
Datasets used in the experiments.
| Dataset | Number of Instances | No. of Attributes | No. of Classes |
|---|---|---|---|
| Iris | 150 | 4 | 3 |
| Climate | 540 | 18 | 2 |
| Credit | 690 | 15 | 3 |
| Wave | 5000 | 21 | 3 |
| Satellite | 6435 | 36 | 7 |
| Firm | 10800 | 20 | 4 |
| Letter | 20000 | 17 | 26 |
| Colon | 62 | 2000 | 2 |
| Liver | 345 | 6 | 2 |
| Vowel | 990 | 10 | 11 |
Optimal values of ELMs' parameters (nh: number of hidden nodes).
| Dataset |
|
| nh |
|---|---|---|---|
| Iris | 230 | 1.2 | 20 |
| Climate | 220 | 1.2 | 20 |
| Credit | 210 | 1.2 | 20 |
| Wave | 232 | 1.2 | 310 |
| Satellite | 232 | 1.4 | 310 |
| Firm | 232 | 1.2 | 310 |
| Letter | 232 | 1.2 | 700 |
| Colon | 2−16 | 0.3 | 80 |
| Liver | 230 | 1.2 | 20 |
| Vowel | 230 | 1.2 | 200 |
Figure 2Surface plot in terms of performance of AELME sensitivity to the user-specified parameters (C, λ): an example on Wave dataset.
Wilcoxon signed rank statistical test result of AELME versus all the algorithms upon all data sets. p values are small which implies the significance of the AELME approach as compared to the other algorithms (here Alg means Algorithm).
| AELME versus Alg |
|
|---|---|
| ELML2 | 0.0009770 |
| RELM | 0.0009766 |
| ELMK | 0.0097000 |
| Bagging | 0.0019500 |
| DELM | 0.0029200 |
| EnELM | 0.0019500 |
| DSELME | 0.0009766 |
| Adaboost | 0.0019500 |
| ELM | 0.0009765 |
| SVM | 0.0009760 |
Description of Q-Statistic experiments. (N is number of samples, #Att. is number of attributes, #Class is number of classes, and #Ens. is number of ensembles).
| Dataset |
| #Att. | #Class | #Ens. |
|---|---|---|---|---|
| Liver | 345 | 6 | 2 | 700 |
| Satellite | 1200 | 10 | 7 | 1300 |
|
| ||||
|
| ||||
| Dataset | Wave | |||
| | 700 | |||
| #Att. | 10 | |||
| #Class | 3 | |||
| #Ens. | 500 | |||
| The set of 10 features was divided into permutations subsets (first 500 permutations) of 4, 4, and 2. | ||||
Maximum improvement (Max-Impr) of ensemble accuracy over the single best classifier (Ensemble Accuracy − Maximum individual accuracy) with their corresponding Q-values.
| Dataset | Max-Impr |
|
|---|---|---|
| Liver | 0.320 | −0.8924 |
| Wave | 0.175 | −0.0600 |
| Satellite | 0.498 | −0.4000 |
Comparisons of the average accuracy rates with their corresponding standard deviations of all the algorithms on the datasets.
| Dataset | AELME | SVM | ELML2 | RELM | ELMK | Bagging | DELM | EnELM | DSELME | Adaboost | ELM |
|---|---|---|---|---|---|---|---|---|---|---|---|
| Iris | 0.9906 ± 0.0132 | 0.9670 ± 0.7100 | 0.9873 ± 0.0287 | 0.9901 ± 0.0338 | 1 ± 0.00 | 0.9656 ± 0.0099 | 0.9625 ± 0.00 | 0.9688 ± 0.0211 | 0.9500 ± 0.1234 | 0.9750 ± 0.0221 | 0.9856 ± 0.0151 |
| Climate | 0.9100 ± 0.0091 | 0.9000 ± 0.4470 | 0.8967 ± 0.0111 | 0.8933 ± 0.0064 | 0.9000 ± 0.00 | 0.9000 ± 0.0163 | 0.8707 ± 0.0158 | 0.8640 ± 0.0176 | 0.8920 ± 0.0250 | 0.9000 ± 0.0761 | 0.8953 ± 0.0095 |
| Credit | 0.7569 ± 0.0047 | 0.7500 ± 1.203 | 0.7495 ± 0.0094 | 0.7422 ± 0.0106 | 0.7426 ± 0.00 | 0.7245 ± 0.0237 | 0.7485 ± 0.0089 | 0.74657 ± 0.0143 | 0.7451 ± 0.0122 | 0.7034 ± 0.0280 | 0.7441 ± 0.0109 |
| Wave | 0.8674 ± 0.0022 | 0.8510 ± 1.68 | 0.8563 ± 0.0345 | 0.8495 ± 0.0323 | 0.8589 ± 0.00 | 0.8109 ± 0.0070 | 0.8712 ± 0.0026 | 0.8682 ± 0.0049 | 0.8628 ± 0.0051 | 0.7341 ± 0.0036 | 0.8587 ± 0.0245 |
| Satellite | 0.9017 ± 0.0051 | 0.8803 ± 2.176 | 0.8816 ± 0.0383 | 0.8903 ± 0.0345 | 0.8204 ± 0.00 | 0.8360 ± 0.0031 | 0.8593 ± 0.0014 | 0.8789 ± 0.0179 | 0.8454 ± 0.0017 | 0.8235 ± 0.0051 | 0.8595 ± 0.0399 |
| Firm | 0.9000 ± 0.00072 | 0.8888 ± 1.12 | 0.88244 ± 0.0219 | 0.85441 ± 0.0583 | 0.9001 ± 0.00 | 0.6873 ± 0.0060 | 0.8993 ± 0.0021 | 0.893 ± 0.0025 | 0.89445 ± 0.0019 | 0.8211 ± 0.0026 | 0.8831 ± 0.0414 |
| Letter | 0.9353 ± 0.0133 | 0.9237 ± 0.26 | 0.8103 ± 0.3265 | 0.9300 ± 0.2538 | 0.9315 ± 0.00 | 0.8400 ± 0.0085 | 0.8246 ± 0.0011 | 0.8221 ± 0.0016 | 0.9006 ± 0.0022 | 0.8338 ± 0.0074 | 0.8200 ± 0.9287 |
| Colon | 0.8900 ± 0.0211 | 0.8438 ± 0.00 | 0.8500 ± 0.0707 | 0.8400 ± 0.1137 | 0.8000 ± 0.00 | 0.85000 ± 0.0082 | 0.8450 ± 0.0158 | 0.8350 ± 0.0474 | 0.8600 ± 0.0211 | 0.8700 ± 0.0211 | 0.8200 ± 0.0882 |
| Liver | 0.7420 ± 0.0155 | 0.6900 ± 2.6 | 0.6580 ± 0.0454 | 0.6730 ± 0.0497 | 0.7200 ± 0.00 | 0.7000 ± 0.0194 | 0.6950 ± 0.0165 | 0.6540 ± 0.0433 | 0.6630 ± 0.0350 | 0.7000 ± 0.0236 | 0.6666 ± 0.0506 |
| Vowel | 0.6620 ± 0.0183 | 0.5660 ± 3.24 | 0.5120 ± 0.0488 | 0.5810 ± 0.0842 | 0.6380 ± 0.00 | 0.6700 ± 0.1180 | 0.6333 ± 0.00213 | 0.5680 ± 0.0306 | 0.5580 ± 0.0473 | 0.6673 ± 0.0374 | 0.5860 ± 0.0400 |
Accuracy rates of AELME upon the datasets using weighted sum method.
| Dataset | Accuracy |
|---|---|
| Iris | 0.9930 |
| Climate | 0.9067 |
| Credit | 0.7696 |
| Wave | 0.8748 |
| Satellite | 0.898 |
| Firm | 0.9008 |
| Letter | 0.9058 |
| Colon | 0.9 |
| Liver | 0.74 |
| Vowel | 0.71 |
Standard deviation of accuracy rates of AELME and base classifiers {ELM, ELMR, ELML2} upon Climate dataset. 200 + 8, 200 + 16, 200 + 32, 200 + 64, and 200 + 128 instances are selected in sequence, corresponding to 1st to the 5th group, respectively.
| Group | AELME | ELM | ELMR | ELML2 |
|---|---|---|---|---|
| 1st | 0.034577 | 0.116800 | 0.081347 | 0.110030 |
| 2nd | 0.025810 | 0.098455 | 0.153540 | 0.0769327 |
| 3rd | 0.044799 | 0.126870 | 0.085248 | 0.096722 |
| 4th | 0.021785 | 0.048463 | 0.056345 | 0.039907 |
| 5th | 0.026814 | 0.055462 | 0.050303 | 0.062163 |
Disagreement measurement between base classifiers in AELME.
| Dataset | Disagreement |
|---|---|
| Iris | 0.0169 |
| Climate | 0.0530 |
| Credit | 0.0946 |
| Wave | 0.1083 |
| Satellite | 0.0568 |
| Firm | 0.0751 |
| Letter | 0.1607 |
| Colon | 0.0500 |
| Liver | 0.0289 |
| Vowel | 0.0130 |
Mean Absolute Error (MAE), Relative Reduction Error (RErRed) and their Averages (AvgAbsEr, AvgRErRed, resp.), and Standard Deviation (STD) of all Algorithms.
| Dataset | Measure | AELME | DELM | DSELME | EnELM | Adaboost | Bagging |
|---|---|---|---|---|---|---|---|
| Iris | MAE | 0.0094 | 0.0375 | 0.05 | 0.0312 | 0.025 | 0.0344 |
| RErRed (%) | −299 | −432 | −232 | −166 | −266 | ||
| STD | 0.0132 | 0 | 0.1234 | 0.0211 | 0.0221 | 0.0099 | |
|
| |||||||
| Climate | MAE | 0.09000 | 0.1293 | 0.1080 | 0.1360 | 0.1000 | 0.1000 |
| RErRed (%) | −44 | −20 | −51 | −11 | −11 | ||
| STD | 0.0091 | 0.0158 | 0.0250 | 0.0176 | 0.0761 | 0.0163 | |
|
| |||||||
| Credit | MAE | 0.2431 | 0.2515 | 0.2549 | 0.2534 | 0.2966 | 0.2755 |
| RErRed (%) | −3 | −5 | −4 | −22 | −13 | ||
| STD | 0.0047 | 0.0089 | 0.0122 | 0.0143 | 0.0280 | 0.0237 | |
|
| |||||||
| Wave | MAE | 0.1326 | 0.1288 | 0.1372 | 0.1318 | 0.2659 | 0.1891 |
| RErRed (%) | 3 | −3 | 1 | −101 | −43 | ||
| STD | 0.0022 | 0.0026 | 0.0051 | 0.0049 | 0.0036 | 0.0070 | |
|
| |||||||
| Satellite | MAE | 0.0983 | 0.1407 | 0.1546 | 0.1211 | 0.1765 | 0.1640 |
| RErRed (%) | −43 | −57 | −23 | −80 | −67 | ||
| STD | 0.0051 | 0.0014 | 0.0017 | 0.0179 | 0.0051 | 0.0031 | |
|
| |||||||
| Firm | MAE | 0.1000 | 0.1007 | 0.1056 | 0.1070 | 0.1789 | 0.3127 |
| RErRed (%) | −1 | −6 | −7 | −79 | −213 | ||
| STD | 0.00072 | 0.0021 | 0.0019 | 0.0025 | 0.0026 | 0.0060 | |
|
| |||||||
| Letter | MAE | 0.0647 | 0.1754 | 0.0994 | 0.1779 | 0.1662 | 0.1600 |
| RErRed (%) | −171 | −54 | −175 | −157 | −147 | ||
| STD | 0.0133 | 0.0011 | 0.0022 | 0.0016 | 0.0074 | 0.0085 | |
|
| |||||||
| Colon | MAE | 0.8500 | 0.1550 | 0.1400 | 0.1650 | 0.1300 | 0.1500 |
| RErRed (%) | 82 | 84 | 81 | 85 | 82 | ||
| STD | 0.0211 | 0.0158 | 0.0211 | 0.0474 | 0.0211 | 0.0082 | |
|
| |||||||
| Liver | MAE | 0.6580 | 0.3050 | 0.3370 | 0.3460 | 0.3000 | 0.3000 |
| RErRed (%) | 54 | 49 | 47 | 54 | 54 | ||
| STD | 0.0155 | 0.0165 | 0.0350 | 0.0433 | 0.0236 | 0.0194 | |
|
| |||||||
| Vowel | MAE | 0.5120 | 0.3667 | 0.4420 | 0.4320 | 0.3327 | 0.3300 |
| RErRed (%) | 28 | 14 | 16 | 35 | 36 | ||
| STD | 0.0183 | 0.0213 | 0.0473 | 0.0306 | 0.0374 | 0.1180 | |
|
| |||||||
| AvgAbsEr | 0.2758 | 0.1791 | 0.1829 | 0.1901 | 0.1972 | 0.2016 | |
| AvgRErRed | − 39 | − 43 | − 35 | − 44 | −59 | ||
Figure 3Training time of Adaboost, Bagging, and AELME.
Average training time (×103 s) of the datasets in seconds.
| Dataset | AELME | EnELM | DSELME | DELM | Ada | Bag | ELM | RELM | ELML2 | ELMK |
|---|---|---|---|---|---|---|---|---|---|---|
| Iris | 0.00004 | 0.0008 | 0.00046 | 0.0003 | 0.00012 | 0.00106 | 0.000025 | 0.000036 | 0.00005 | 0.000013 |
| Climate | 0.0002 | 0.0022 | 0.00126 | 0.0018 | 0.00404 | 0.03736 | 0.000052 | 0.000204 | 0.00005 | 0.000095 |
| Credit | 0.0008 | 0.0044 | 0.00225 | 0.002 | 0.01219 | 0.10610 | 0.000047 | 0.001170 | 0.00007 | 0.000134 |
| Wave | 0.138 | 0.028 | 0.008 | 0.0107 | 0.1329307 | 1.09000 | 0.000222 | 0.001234 | 0.00028 | 0.004964 |
| Satellite | 0.1458 | 0.032 | 0.017 | 0.026 | 0.36782 | 2.30770 | 0.000304 | 0.210789 | 0.00038 | 0.012488 |
| Firm | 0.182 | 0.045 | 0.0393 | 0.11487 | 3.98310 | 3.9797 | 0.000373 | 0.401937 | 0.00046 | 0.019656 |
| Letter | 0.878 | 0.229 | 0.287 | 0.2911 | 6.68 | 6.88 | 0.000638 | 1.238600 | 0.00075 | 0.066160 |
| Colon | 0.000148 | 0.000808 | 0.000148 | 0.003449 | 0.000096 | 0.000131 | 0.000033 | 0.000080 | 0.000069 | 0.000005 |
| Liver | 0.000142 | 0.0001045 | 0.000509 | 0.002902 | 0.012979 | 0.0130564 | 0.000027 | 0.000076 | 0.0000421 | 0.000039 |
| Vowel | 0.0007082 | 0.0004571 | 0.0016411 | 0.0029796 | 0.1480 | 0.151 | 0.000075 | 0.001246 | 0.0000764 | 0.000184 |