| Literature DB >> 31851696 |
Youcheng Qian1,2, Xueyan Yin3, Jun Kong4, Jianzhong Wang4, Wei Gao1.
Abstract
Graph-based dimensionality reduction methods have attracted substantial attention due to their successful applications in many tasks, including classification and clustering. However, most classical graph-based dimensionality reduction approaches are only applied to data from one view. Hence, combining information from different data views has attracted considerable attention in the literature. Although various multi-view graph-based dimensionality reduction algorithms have been proposed, the graph construction strategies utilized in them do not adequately take noise and different importance of multiple views into account, which may degrade their performance. In this paper, we propose a novel algorithm, namely, Low-Rank Graph Optimization for Multi-View Dimensionality Reduction (LRGO-MVDR), that overcomes these limitations. First, we construct a low-rank shared matrix and a sparse error matrix from the graph that corresponds to each view for capturing potential noise. Second, an adaptive nonnegative weight vector is learned to explore complementarity among views. Moreover, an effective optimization procedure based on the Alternating Direction Method of Multipliers scheme is utilized. Extensive experiments are carried out to evaluate the effectiveness of the proposed algorithm. The experimental results demonstrate that the proposed LRGO-MVDR algorithm outperforms related methods.Entities:
Mesh:
Year: 2019 PMID: 31851696 PMCID: PMC6919611 DOI: 10.1371/journal.pone.0225987
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Fig 1Overview of the shared graph matrix construction.
Fig 2The examples of similarity matrices.
The elements of zero values are illustrated with white and those of non-zero values are illustrated with blue.
Detailed information on the simulated dataset.
| Cluster 1 | Cluster 2 | Cluster 3 | ||
|---|---|---|---|---|
| View 1 | Centers | |||
| Covariances | ||||
| View 2 | Centers | |||
| Covariances | ||||
| View 3 | Centers | |||
| Covariances |
Fig 3Simulated dataset example.
Detailed information on the datasets that are used for clustering.
| Dataset | Size | Views | Clusters |
|---|---|---|---|
| Simulated | 600 | 3 | 3 |
| Wisconsin | 265 | 2 | 5 |
| WebKB | 1051 | 2 | 2 |
| Cornell | 195 | 2 | 5 |
Best clustering results (ACC±std) of various algorithms.
| Methods | Simulated | Wisconsin | WebKB | Cornell |
|---|---|---|---|---|
| BSV | 0.8295±0.0048 (2, 1.2883) | 0.6298±0.0254 (4, 2.0602) | 0.9519±0.0784 (2, 5.7666) | 0.5205±0.0373 (4, 2.8671) |
| FeaConc | 0.9583±0.0000 (2, 0.7068) | 0.6049±0.0428 (4, 1.3095) | 0.9429±0.0030 (2, 3.3374) | 0.5477±0.0349 (4, 1.7544) |
| KA | 0.9540±0.0008 (2, 0.7436) | 0.4668±0.0362 (10, 1.1770) | 0.9353±0.0000 (2, 3.3640) | 0.4364±0.0045 (4,1.7907) |
| MVSim | 0.6510±0.0052 (2, 1.6887) | 0.5868±0.0621 (2, 3.7267) | 0.9479±0.0064 (7, 15.6147) | 0.4626±0.0153 (2, 4.6397) |
| MVSpec | 0.9560±0.0016 (2, 1.1547) | 0.4438±0.0222 (4, 2.2537) | 0.9667±0.0000 (3, 22.8464) | 0.4113±0.0164 (3, 3.1782) |
| MvSpecCE | 0.8700±0.0000 (2, 60.3064) | 0.4053±0.0184 (6, 201.025) | 0.7275±0.0512 (6, 5118.3) | 0.4267±0.0093 (2, 439.42) |
| Cospectral | 0.9283±0.0000 (2, 6.1690) | 0.5457±0.0164 (4, 14.3779) | 0.9600±0.0000 (3, 79.8809) | 0.4374±0.0123 (16, 21.1995) |
| Corespectral | 0.9600±0.0000 (2, 2.5740) | 0.4706±0.0199 (6, 7.7287) | 0.9365±0.0005 (3, 41.8010) | 0.4318±0.0203 (4, 10.3666) |
| MSE | 0.9517±0.0000 (2, 1.3142) | 0.5189±0.0368 (6, 1.7598) | 0.9667±0.0000 (3, 19.0447) | 0.4621±0.0294 (4, 2.4776) |
| RMSC | 0.9497±0.0004 (2, 7.3515) | 0.5381±0.0211 (6, 2.0915) | 0.9667±0.0000 (3, 26.0721) | 0.4718±0.0331 (15, 2.2652) |
| LRGO-MVDR |
The numbers in parentheses are the feature dimension corresponds to the best result and the training time of each algorithm, respectively.
Fig 4NMI results of various algorithms.
Fig 5Performance of LRGO-MVDR under various parameter values for clustering tasks on four datasets.
Fig 6Convergence curves of LRGO-MVDR on four datasets for clustering.
Best clustering results (ACC±std) of three methods for solving Eq (19).
| Methods | Simulated | Wisconsin | WebKB | Cornell |
|---|---|---|---|---|
| LRGO-MVDR-L | 0.96129239±0.0007 (2, 4.1361) | 0.64145186±0.0205 (6, 2.4216) | 0.96857136±0.0000 (3, 18.4467) | 0.60262186±0.0081 (4, 2.8211) |
| LRGO-MVER-S | 0.95758912±0.0008 (2, 12.1196) | 0.63234186±0.0116 (6, 6.6156) | 0.96574672±0.0000 (2, 39.8058) | 0.59854761±0.0321 (4, 5.1145) |
| LRGO-MVDR-C |
The numbers in parentheses are the feature dimension corresponds to the best result and the training time of each algorithm, respectively.
Detailed information on the datasets for classification.
| Dataset | Size | Views | Classes | ||
|---|---|---|---|---|---|
| Caltech101 | 512 | 3 | 11 | 258 | 254 |
| Wiki | 693 | 2 | 10 | 349 | 344 |
| Yale | 165 | 3 | 15 | 90 | 75 |
| Cornell | 195 | 2 | 5 | 99 | 96 |
Fig 7Classification performance of various algorithms.
Best classification results (HVA±std) of various algorithms.
| Methods | Caltech101 | Wiki | Yale | Cornell |
|---|---|---|---|---|
| BSV | 0.5110±0.0337 (50, 3.5896) | 0.6500±0.0181 (10,3.0380) | 0.9187±0.0284 (20, 3.0975) | 0.6646±0.0367 (12, 5.0114) |
| FeaConc | 0.3028±0.0225 (35, 1.8866) | 0.6195±0.0185 (20, 2.0563) | 0.8773±0.0250 (24, 1.1578) | 0.6917±0.0602 (12, 3.2670) |
| KA | 0.4339±0.0256 (50, 1.8555) | 0.6224±0.0189 (10, 1.9641) | 0.8027±0.0300 (28, 1.6784) | 0.6312±0.0165 (21, 2.8715) |
| MVSim | 0.2807±0.0209 (40, 3.9675) | 0.5625±0.0179 (40, 3.2889) | 0.6053±0.0475 (40, 3.0893) | 0.5844±0.0487 (21, 6.3791) |
| MVSpec | 0.2165±0.0201 (50, 3.5650) | 0.6497±0.0253 (30, 4.6592) | 0.4040±0.0361 (8, 2.2567) | 0.5615±0.0349 (9, 4.3756) |
| MvSpecCE | 0.2795±0.0000 (15, 254.109) | 0.1988±0.1727(30, 318.7920) | 0.5333±0.0000 (8, 88.5287) | 0.5521±0.0000 (6, 504.64) |
| Cospectral | 0.2764±0.0096 (30, 23.9225) | 0.6360±0.0160 (45, 20.9476) | 0.7453±0.0341 (20, 15.3160) | 0.6646±0.0264 (15, 26.1359) |
| Corespectral | 0.3028±0.0225 (35, 11.1016) | 0.2032±0.0079 (40, 10.3300) | 0.6253±0.0528 (16, 6.2484) | 0.5094±0.0581 (30, 14.7143) |
| MSE | 0.4169±0.0303 (50, 3.3677) | 0.6032±0.0173 (10, 4.4454) | 0.7953±0.0385 (28, 1.7592) | 0.6271±0.0319 (24, 3.4674) |
| RMSC | 0.4232±0.0246 (45, 6.8901) | 0.6125±0.0219 (10, 9.3732) | 0.80000±0.0377 (32, 2.2527) | 0.64381±0.0274 (21, 3.3620) |
| LRGO-MVDR |
The numbers in parentheses are the feature dimension corresponds to the best result and the training time of each algorithm, respectively.
Fig 8Performance of LRGO-MVDR under various parameter values for classification tasks on four datasets.
Fig 9Convergence curves of LRGO-MVDR on four datasets for classification.
Best classification results (HVA±std) of three methods for solving Eq (19).
| Methods | Caltech101 | Wiki | Yale | Cornell |
|---|---|---|---|---|
| LRGO-MVDR-L | 0.54995924±0.0257 (50, 6.1097) | 0.66367536±0.0168 (15, 8.8546) | 0.94528513±0.0203 (20, 2.2008) | 0.71767461±0.0271 (9, 3.8976) |
| LRGO-MVDR-S | 0.52323286±0.0263 (50, 16.1447) | 0.66153208±0.0384 (25,10.4051) | 0.92404011±0.0474 (28, 7.88366) | 0.70308023±0.0253 (6, 5.1217) |
| LRGO-MVDR-C |
The numbers in parentheses are the feature dimension corresponds to the best result and the training time of each algorithm, respectively.
p-values of the Wilcoxon rank sum tests on clustering tasks (ACC).
| LRGO -MVDR vs. BSV | 0.0038 |
| LRGO -MVDR vs. FeaConc | 0.0011 |
| LRGO -MVDR vs. KA | 5.7275e-06 |
| LRGO -MVDR vs. MVSim | 4.5902e-04 |
| LRGO -MVDR vs. MVSpec | 0.00019 |
| LRGO -MVDR vs. MvSpecCE | 5.7392e-05 |
| LRGO -MVDR vs. Cospectral | 1.2268e-04 |
| LRGO -MVDR vs. Corespectral | 8.2736e-05 |
| LRGO -MVDR vs. MSE | 0.0019 |
| LRGO -MVDR vs. RMSC | 0.0019 |
p-values of the Wilcoxon rank sum tests on classification tasks (HVA).
| LRGO -MVDR vs. BSV | 0.0479 |
| LRGO -MVDR vs. FeaConc | 0.0470 |
| LRGO -MVDR vs. KA | 0.0030 |
| LRGO -MVDR vs. MVSim | 2.8704e-07 |
| LRGO -MVDR vs. MVSpec | 2.2929e-09 |
| LRGO -MVDR vs. MvSpecCE | 3.1995e-11 |
| LRGO -MVDR vs. Cospectral | 0.0020 |
| LRGO -MVDR vs. Corespectral | 5.4697e-11 |
| LRGO -MVDR vs. MSE | 0.0039 |
| LRGO -MVDR vs. RMSC | 0.0064 |
p-values of the Wilcoxon rank sum tests on clustering tasks (NMI).
| LRGO -MVDR vs. BSV | 1.0348e-04 |
| LRGO -MVDR vs. FeaConc | 0.0031 |
| LRGO -MVDR vs. KA | 0.0019 |
| LRGO -MVDR vs. MVSim | 3.63e-09 |
| LRGO -MVDR vs. MVSpec | 0.0019 |
| LRGO -MVDR vs. MvSpecCE | 3.5697e-09 |
| LRGO -MVDR vs. Cospectral | 0.0019 |
| LRGO -MVDR vs. Corespectral | 0.0035 |
| LRGO -MVDR vs. MSE | 0.0019 |
| LRGO -MVDR vs. RMSC | 0.0019 |