| Literature DB >> 28336938 |
Yong Zhang1, Dun-Wei Gong2, Xiao-Yan Sun2, Yi-Nan Guo2.
Abstract
Feature selection is an important data preprocessing technique in multi-label classification. Although a large number of studies have been proposed to tackle feature selection problem, there are a few cases for multi-label data. This paper studies a multi-label feature selection algorithm using an improved multi-objective particle swarm optimization (PSO), with the purpose of searching for a Pareto set of non-dominated solutions (feature subsets). Two new operators are employed to improve the performance of the proposed PSO-based algorithm. One operator is adaptive uniform mutation with action range varying over time, which is used to extend the exploration capability of the swarm; another is a local learning strategy, which is designed to exploit the areas with sparse solutions in the search space. Moreover, the idea of the archive, and the crowding distance are applied to PSO for finding the Pareto set. Finally, experiments verify that the proposed algorithm is a useful approach of feature selection for multi-label classification problem.Entities:
Year: 2017 PMID: 28336938 PMCID: PMC5428503 DOI: 10.1038/s41598-017-00416-0
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1The pseudocode of the function MUTATION.
Figure 2The flowchart of the proposed algorithm.
Format of six datasets.
| Data sets | Number of training samples | Number of testing samples | Number of labels | Number of features |
|---|---|---|---|---|
| Flags | 129 | 65 | 7 | 19 |
| CAL500 | 250 | 252 | 174 | 68 |
| Emotions | 391 | 202 | 6 | 72 |
| Yeast | 1500 | 917 | 14 | 103 |
| Birds | 322 | 323 | 19 | 260 |
| Scene | 1211 | 1196 | 6 | 294 |
Solutions with the smallest Hloss value found by the four comparison algorithms.
| Datasets | Proposed algorithm | MI-PPT | RF-BR | NSGA-II | ||||
|---|---|---|---|---|---|---|---|---|
| Hamming loss | Number of features | Hamming loss | Number of features | Hamming loss | Number of features | Hamming loss | Number offeatures | |
| Emotions |
| 27 | 0.229 | 43 | 0.220 |
| 0.183 | 47 |
| Yeast |
| 45 | 0.194 | 91 | 0.240 |
| 0.196 | 56 |
| Scene |
| 164 | 0.092 | 278 | 0.120 |
| 0.092 | 123 |
Figure 3Pareto optimal sets found by MI-PPT and our algorithm on Emotions.
Figure 4Pareto optimal sets found by MI-PPT and our algorithm on Yeast.
Figure 5Pareto optimal sets found by MI-PPT and our algorithm on Scene.
The average set coverage values of the two algorithms.
| Datasets | (Proposed algorithm, MI-PPT) | (MI-PPT, Proposed algorithm) |
|---|---|---|
| Emotions | 0.8571 | 0.1429 |
| Yeast | 1.0 | 0 |
| Scene | 0.6429 | 0.4286 |
The average HV values obtained by the two algorithms on the six datasets.
| Data sets | Proposed algorithm | NSGA-II | t-test |
|---|---|---|---|
| Flags | 0.738/0.076 | 0.592/0.063 | Y+ |
| CAL500 | 0.798/0.077 | 0.595/0.022 | Y+ |
| Emotions | 0.766/0.029 | 0.553/0.036 | Y+ |
| Yeast | 0.709/0.036 | 0.508/0.011 | Y+ |
| Birds | 0.859/0.043 | 0.580/0.015 | Y+ |
| Scene | 0.756/0.021 | 0.560/0.014 | Y+ |
Figure 6Solutions obtained by our algorithm and NSGA-II on six datasets.
Figure 7The curve of HV values with respect to the iterations obtained by MPSOFS and MPSOFS/LLS.
Figure 8The curve of HV values with respect to the iterations obtained by MPSOFS and MPSOFS/M.
The average HV values obtained by MPSOFS, MPSOFS-NM and MPSOFS-PRM on the three datasets.
| Data sets | MPSOFS | MPSOFS-NM | MPSOFS-PRM |
|---|---|---|---|
| Emotions | 0.766/0.029 | 0.762/0.025 | 0.764/0.027 |
| Yeast | 0.709/0.036 | 0.686/0.031 | 0.710/0.039 |
| Scene | 0.756/0.021 | 0.738/0.020 | 0.754/0.024 |