| Literature DB >> 30395624 |
Changming Zhu1, Xiang Ji1, Chao Chen1, Rigui Zhou1, Lai Wei1, Xiafen Zhang1.
Abstract
Most data sets consist of interlaced-distributed samples from multiple classes and since these samples always cannot be classified correctly by a linear hyperplane, so we name them nonlinearly separable data sets and corresponding classifiers are named nonlinear classifiers. Traditional nonlinear classifiers adopt kernel functions to generate kernel matrices and then get optimal classifier parameters with the solution of these matrices. But computing and storing kernel matrices brings high computational and space complexities. Since INMKMHKS adopts Nyström approximation technique and NysCK changes nonlinearly separable data to linearly ones so as to reduce the complexities, we combines ideas of them to develop an improved NysCK (INysCK). Moreover, we extend INysCK into multi-view applications and propose multi-view INysCK (MINysCK). Related experiments validate the effectiveness of them in terms of accuracy, convergence, Rademacher complexity, etc.Entities:
Mesh:
Year: 2018 PMID: 30395624 PMCID: PMC6218068 DOI: 10.1371/journal.pone.0206798
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Full name and abbreviation for some used terms.
| Full name | Abbreviation |
|---|---|
| Ho-Kashyap algorithm | HK |
| Ho-Kashyap algorithm with squared approximation of the misclassification errors | MHKS |
| support vector machine | SVM |
| nonlinearly combined classifiers | NCC |
| fuzzy clustering with nonlinearly transformed data | FC-NTD |
| kernelized modification of MHKS | KMHKS |
| kernel SVM | KSVM |
| multi-views KMHKS | MultiV-KMHKS |
| multi-view learning developed from single-view patterns with Ho-Kashyap linear classification strategy | MVMHKS |
| regularized MVMHKS | RMVMHKS |
| double-fold localized multiple matrix learning machine | DLMMLM |
| Universum based DLMMLM | UDLMMLM |
| Nyström approximation matrix with multiple KMHKSs | NMKMHKS |
| improved NMKMHKS | INMKMHKS |
| cluster kernel | CK |
| Nyström CK | NysCK |
| improved NysCK | INysCK |
| multi-view INysCK | MINysCK |
| multi-view L2-SVM | MSVM |
| multiple graph regularized generative model | MGGM |
| multi-view least squares support vector machines | MV-LSSVM |
| multi-view and multi-feature learning | MVMFL |
| semi-supervised multi-view maximum entropy discrimination approach | SMVMED |
| multi-view low-rank sparse subspace clustering | MLRSSC |
| kernel MLRSSC | KMLRSSC |
| multi-view kernel spectral clustering | MVKSC |
| matrix-pattern-oriented MHKS with boundary projection discrimination | BPDMatMHKS |
| regularized weighted least square support vector classifier | rWLSSVC |
| novel dissimilarity learning | NDL |
| locality constrained dictionary learning | LCDL |
| scale-invariant feature transform | SIFT |
| singular value decomposition | SVD |
| radial basis function | RBF |
Algorithm: INysCK.
| 1. Generate |
| 2. For p = 1,2,…,M do |
| 3. Construct kernel matrix |
| 4. Centralize, normalize, and decompose |
| 5. Carry out SVD on |
| 6. End for |
| 7. Compute coefficient |
| 8. On the base of |
| 9. Carry out SVD on |
| 10. Compute |
| 11. Get |
Algorithm: MINysCK.
| 1. For g = 1,2,…,V do |
| 2. Change |
| 3. End for |
| 4. Obtain |
Description of the used UCI data sets.
| Data set | No. dimensions | No. classes | No. samples |
|---|---|---|---|
| YouTube Comedy Slam (YCS) | 2 | 2 | 1138562 |
| Authorship Attribution (AA) | 1000 | 50 | 93600 |
| Breast Cancer (BC) | 10 | 2 | 699 |
| Arrhythmia | 279 | 16 | 452 |
Used classifiers.
| nonlinear | linear | |
|---|---|---|
| single-view | KMHKS [ | SVM [ |
| NDL [ | BPDMatMHKS [ | |
| multi-view | MultiV-KMHKS [ | MSVM [ |
| MGGM [ | ||
| MVMFL [ | ||
| KMLRSSC [ |
Fig 1Accuracy with related classifiers and CK-based methods on used single-view data sets.
CK-related method in italic represents baseline one and one in bold denotes the proposed one. For classifiers, SVM is used as the baseline one and we just clarify this point in words rather than in font. In other figures and tables, we have similar representations.
Fig 2Accuracy with related classifiers and CK-based methods on used multi-view data sets.
CK-related method in italic represents baseline one and ones in bold denote the proposed ones. For classifiers, MSVM is used as the baseline one and we just clarify this point in words rather than in font. In other figures and tables, we have similar representations.
Comparison about time (in seconds) cost for the three NysCK-related methods.
| Data set | INysCK | MINysCK | Data set | INysCK | MINysCK | ||
|---|---|---|---|---|---|---|---|
| YCS | 43.405 | 44.642 | / | AA | 146.672 | 154.527 | / |
| BC | 0.087 | 0.095 | / | Arrhythmia | 1.577 | 1.634 | / |
| NUS-WIDE | 1380.392 | 1434.962 | 1406.573 | YMVG | 241078.011 | 249127.624 | 247240.890 |
| DBLP | 4.166 | 4.282 | 4.266 | Cora | 29.663 | 32.586 | 32.069 |
Fig 3Distributions of samples with different CK-related methods on a binary-class data set.
The numbers of iterations comparisons.
| single-view | Null | NysCK | INysCK | multi-view | Null | NysCK | INysCK | MINysCK | ||
|---|---|---|---|---|---|---|---|---|---|---|
| SVM | 17.20 | 17.13 | 16.06 | 14.96 | MSVM | 17.57 | 16.35 | 14.72 | 14.44 | 13.43 |
| MHKS | 21.30 | 19.69 | 18.79 | 18.06 | MLRSSC | 29.18 | 26.60 | 24.44 | 23.26 | 22.05 |
| BPDMatMHKS | 22.94 | 22.94 | 22.29 | 21.89 | MultiV-KMHKS | 25.63 | 24.77 | 22.71 | 21.24 | 20.16 |
| rWLSSVC | 20.13 | 19.30 | 18.96 | 18.64 | DLMMLM | 28.91 | 27.10 | 25.68 | 24.84 | 23.18 |
| KMHKS | 20.82 | 18.98 | 17.55 | 17.13 | MGGM | 27.03 | 24.46 | 23.62 | 22.07 | 21.26 |
| KSVM | 17.03 | 16.60 | 15.53 | 15.29 | MV-LSSVM | 43.56 | 40.94 | 38.92 | 37.87 | 36.39 |
| NDL | 19.82 | 18.92 | 18.34 | 17.46 | MVMFL | 66.96 | 66.33 | 64.75 | 62.91 | 62.31 |
| LCDL | 22.42 | 21.81 | 19.92 | 19.26 | SMVMED | 37.19 | 33.95 | 31.37 | 29.06 | 27.69 |
| KMLRSSC | 32.69 | 32.65 | 32.00 | 30.33 | 28.16 | |||||
| MVKSC | 26.78 | 25.55 | 23.48 | 21.46 | 20.70 |
Fig 4The average Rademacher complexity comparison.
Average rank comparisons for different CK-related methods and classifiers.
| single-view | Null | NysCK | INysCK | multi-view | Null | NysCK | INysCK | MINysCK | ||
|---|---|---|---|---|---|---|---|---|---|---|
| SVM | 3.25 | 3.75 | 1.75 | 1.25 | MSVM | 4.80 | 4.00 | 3.20 | 1.60 | 1.40 |
| MHKS | 3.00 | 4.00 | 2.00 | 1.00 | MLRSSC | 5.00 | 3.60 | 3.40 | 1.40 | 1.60 |
| BPDMatMHKS | 3.25 | 3.75 | 1.75 | 1.25 | MultiV-KMHKS | 5.00 | 3.60 | 3.40 | 1.60 | 1.40 |
| rWLSSVC | 3.25 | 3.75 | 1.75 | 1.25 | DLMMLM | 5.00 | 3.60 | 3.40 | 1.60 | 1.40 |
| KMHKS | 3.00 | 4.00 | 2.00 | 1.00 | MGGM | 4.80 | 4.00 | 3.20 | 1.60 | 1.40 |
| KSVM | 3.25 | 3.75 | 1.75 | 1.25 | MV-LSSVM | 5.00 | 3.60 | 3.40 | 1.60 | 1.40 |
| NDL | 3.25 | 3.75 | 1.75 | 1.25 | MVMFL | 4.60 | 4.20 | 3.20 | 1.60 | 1.40 |
| LCDL | 3.00 | 4.00 | 2.000 | 1.000 | SMVMED | 5.00 | 3.60 | 3.40 | 1.60 | 1.40 |
| Average | 3.16 | 3.84 | 1.84 | 1.16 | KMLRSSC | 4.80 | 3.80 | 3.40 | 1.60 | 1.40 |
| MVKSC | 4.80 | 3.80 | 3.40 | 1.40 | 1.60 | |||||
| Average | 4.88 | 3.78 | 3.34 | 1.56 | 1.44 |
Critical values for the two-tailed Nemenyi test.
| No. algorithms | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
|---|---|---|---|---|---|---|---|---|---|
| 1.960 | 2.343 | 2.569 | 2.728 | 2.850 | 2.949 | 3.031 | 3.102 | 3.164 | |
| 1.645 | 2.052 | 2.291 | 2.459 | 2.589 | 2.693 | 2.780 | 2.855 | 2.920 |
Fig 5Average influence of ratio of training samples on accuracy with INysCK and MINysCK and corresponding classifiers used.