| Literature DB >> 29601552 |
Yang Zhang1, Yun Liu2, Han-Chieh Chao3,4,5,6, Zhenjiang Zhang7, Zhiyuan Zhang8.
Abstract
In wireless sensor networks, the classification of incomplete data reported by sensor nodes is an open issue because it is difficult to accurately estimate the missing values. In many cases, the misclassification is unacceptable considering that it probably brings catastrophic damages to the data users. In this paper, a novel classification approach of incomplete data is proposed to reduce the misclassification errors. This method uses the regularized extreme learning machine to estimate the potential values of missing data at first, and then it converts the estimations into multiple classification results on the basis of the distance between interval numbers. Finally, an evidential reasoning rule is adopted to fuse these classification results. The final decision is made according to the combined basic belief assignment. The experimental results show that this method has better performance than other traditional classification methods of incomplete data.Entities:
Keywords: classification; evidence theory; extreme learning machine; incomplete data; wireless sensor network
Year: 2018 PMID: 29601552 PMCID: PMC5948797 DOI: 10.3390/s18041046
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1The presentation of three states of the transformer. No: normal state; Te: temperature fault; Di: discharge fault.
Figure 2The schematic diagram of extreme learning machine (ELM).
Figure 3The distributions of the training samples and test samples.
Figure 4Classification results by different approaches: (a) Classification result by regularized ELM with mean imputation (MI) (); (b) Classification result by regularized ELM with K-nearest neighbor imputation (KNNI) ().
Figure 5Classification result by our proposed method ().
The specification the selected data sets.
| Name | Classes | Attributes | Instances |
|---|---|---|---|
| Breast | 2 | 9 | 699 |
| Iris | 3 | 4 | 150 |
| Seeds | 3 | 7 | 210 |
| Wine | 3 | 13 | 178 |
| Segment | 7 | 19 | 2310 |
| Satimage | 6 | 36 | 6435 |
The classification results for different data sets.
| Data set | MI Re | KNNI Re | SVRI Re | ELMI {Re, Ri} | |
|---|---|---|---|---|---|
| Breast | 3 | 4.79% | 4.86% | 4.71% | {4.52%, 3.18%} |
| 5 | 8.23% | 8.38% | 5.96% | {4.90%, 3.31%} | |
| 7 | 14.75% | 14.29% | 12.45% | {10.42%, 5.59%} | |
| Iris | 1 | 23.96% | 5.65% | 5.13% | {4.77%, 2.09%} |
| 2 | 43.67% | 13.72% | 8.07% | {8.12%, 6.52%} | |
| 3 | 65.49% | 21.06% | 12.89% | {12.92%, 9.23%} | |
| Seeds | 2 | 21.05% | 12.38% | 9.26% | {9.47%, 4.13%} |
| 4 | 28.39% | 13.62% | 10.83% | {9.65%, 4.86%} | |
| 6 | 40.93% | 27.34% | 18.37% | {17.02%, 12.53%} | |
| Wine | 3 | 31.26% | 27.18% | 26.51% | {26.17%, 1.64%} |
| 6 | 33.98% | 27.93% | 27.14% | {26.83%, 1.58%} | |
| 9 | 38.02% | 30.36% | 28.33% | {27.25%, 3.67%} | |
| Segment | 3 | 12.71% | 10.46% | 9.52% | {6.89%, 1.37%} |
| 7 | 15.83% | 12.38% | 10.07% | {7.16%, 3.25%} | |
| 11 | 20.23% | 17.35% | 12.06% | {10.35%, 3.76%} | |
| Satimag | 7 | 41.28% | 40.70% | 33.62% | {29.63%, 12.81%} |
| 9 | 44.75% | 42.36% | 36.59% | {30.87%, 18.24%} | |
| 19 | 52.96% | 51.22% | 47.67% | {36.92%, 22.75%} |
SVRI: support vector regression imputation.
The comparison of training and testing time of SVRI and ELMI.
| Data Set | SVRI | ELMI | ||
|---|---|---|---|---|
| Training | Testing | Training | Testing | |
| Breast | 3.8371 s | 0.9353 s | 0.9761 s | 0.2572 s |
| Iris | 1.0639 s | 0.2108 s | 0.1924 s | 0.0497 s |
| Seeds | 2.4271 s | 0.5141 s | 0.4212 s | 0.1644 s |
| Wine | 2.1533 s | 0.5279 s | 0.3125 s | 0.1398 s |
| Segment | 753.2 s | 1.0422 s | 2.32 s | 0.5461 s |
| Satimag | 2157.3 s | 2.1823 s | 15.79 s | 0.5633 s |