| Literature DB >> 35797275 |
María J Nueda1, Carmen Gandía1, Mariola D Molina1.
Abstract
The search of separation hyperplanes is an efficient way to find rules with classification purposes. This paper presents an alternative mathematical programming formulation to existing methods to find a discriminant hyperplane. The hyperplane H is found by minimizing the sum of all the distances to the area assigned to the group each individual belongs to. It results in a convex optimization problem for which we find an equivalent linear programming problem. We demonstrate that H exists when the centroids of the two groups are not equal. The method is effective dealing with low and high dimensional data where reduction of the dimension is proposed to avoid overfitting problems. We show the performance of this approach with different data sets and comparisons with other classifications methods. The method is called LPDA and it is implemented in a R package available in https://github.com/mjnueda/lpda.Entities:
Mesh:
Year: 2022 PMID: 35797275 PMCID: PMC9262202 DOI: 10.1371/journal.pone.0270403
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.752
Fig 1The objective is finding H that separates X from Y.
Fig 2Situation of X and Y related to H, H+1 and H−1.
Fig 3Examples where both methods are successfull but the hyperplanes are slight different.
Fig 4Examples where LPDA gets less separation errors than SVM.
Sensitivity, specificity and classification error for default data.
| Sensitivity | Specificity | Classification error | |
|---|---|---|---|
|
| 0.9009 | 0.8646 | 0.1342 |
| weighted- | 0.9039 | 0.8555 | 0.1429 |
| Logistic | 0.3153 | 0.2372 | 0.0267 |
|
| 0.2372 | 0.9977 | 0.0276 |
Classification error test average and confidence interval in cervical cancer dataset.
| Method | 8 samples | 10 samples | 12 samples |
|---|---|---|---|
|
| 0.102 (0.096, 0.108) | 0.106 (0.101, 0.112) | 0.100 (0.095, 0.104) |
|
| 0.078 (0.072, 0.084) | 0.078 (0.072, 0.083) | 0.081 (0.076, 0.086) |
|
| 0.076 (0.070, 0.082) | 0.078 (0.073, 0.083) | 0.081 (0.076, 0.085) |
|
| 0.102 (0.096, 0.109) | 0.105 (0.099, 0.110) | 0.106 (0.101, 0.111) |
|
| 0.076 (0.071, 0.082) | 0.082 (0.077, 0.087) | 0.079 (0.075, 0.084) |