Literature DB >> 21385668

Structural regularized support vector machine: a framework for structural large margin classifier.

Hui Xue1, Songcan Chen, Qiang Yang.   

Abstract

Support vector machine (SVM), as one of the most popular classifiers, aims to find a hyperplane that can separate two classes of data with maximal margin. SVM classifiers are focused on achieving more separation between classes than exploiting the structures in the training data within classes. However, the structural information, as an implicit prior knowledge, has recently been found to be vital for designing a good classifier in different real-world problems. Accordingly, using as much prior structural information in data as possible to help improve the generalization ability of a classifier has yielded a class of effective structural large margin classifiers, such as the structured large margin machine (SLMM) and the Laplacian support vector machine (LapSVM). In this paper, we unify these classifiers into a common framework from the concept of "structural granularity" and the formulation for optimization problems. We exploit the quadratic programming (QP) and second-order cone programming (SOCP) methods, and derive a novel large margin classifier, we call the new classifier the structural regularized support vector machine (SRSVM). Unlike both SLMM at the cross of the cluster granularity and SOCP and LapSVM at the cross of the point granularity and QP, SRSVM is located at the cross of the cluster granularity and QP and thus follows the same optimization formulation as LapSVM to overcome large computational complexity and non-sparse solution in SLMM. In addition, it integrates the compactness within classes with the separability between classes simultaneously. Furthermore, it is possible to derive generalization bounds for these algorithms by using eigenvalue analysis of the kernel matrices. Experimental results demonstrate that SRSVM is often superior in classification and generalization performances to the state-of-the-art algorithms in the framework, both with the same and different structural granularities.

Mesh:

Year:  2011        PMID: 21385668     DOI: 10.1109/TNN.2011.2108315

Source DB:  PubMed          Journal:  IEEE Trans Neural Netw        ISSN: 1045-9227


  7 in total

1.  The sparseness of mixed selectivity neurons controls the generalization-discrimination trade-off.

Authors:  Omri Barak; Mattia Rigotti; Stefano Fusi
Journal:  J Neurosci       Date:  2013-02-27       Impact factor: 6.167

2.  Multi-Model Adaptation Learning With Possibilistic Clustering Assumption for EEG-Based Emotion Recognition.

Authors:  Yufang Dan; Jianwen Tao; Di Zhou
Journal:  Front Neurosci       Date:  2022-05-04       Impact factor: 5.152

3.  Support vector machine with hypergraph-based pairwise constraints.

Authors:  Qiuling Hou; Meng Lv; Ling Zhen; Ling Jing
Journal:  Springerplus       Date:  2016-09-23

4.  A Novel and Effective Method for Congestive Heart Failure Detection and Quantification Using Dynamic Heart Rate Variability Measurement.

Authors:  Wenhui Chen; Lianrong Zheng; Kunyang Li; Qian Wang; Guanzheng Liu; Qing Jiang
Journal:  PLoS One       Date:  2016-11-11       Impact factor: 3.240

5.  Feature space learning model.

Authors:  Renchu Guan; Xu Wang; Maurizio Marchese; Mary Qu Yang; Yanchun Liang; Chen Yang
Journal:  J Ambient Intell Humaniz Comput       Date:  2018-05-09

6.  Possibilistic Clustering-Promoting Semi-Supervised Learning for EEG-Based Emotion Recognition.

Authors:  Yufang Dan; Jianwen Tao; Jianjing Fu; Di Zhou
Journal:  Front Neurosci       Date:  2021-06-23       Impact factor: 4.677

7.  A new approach to detect congestive heart failure using short-term heart rate variability measures.

Authors:  Guanzheng Liu; Lei Wang; Qian Wang; Guangmin Zhou; Ying Wang; Qing Jiang
Journal:  PLoS One       Date:  2014-04-18       Impact factor: 3.240

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.