Literature DB >> 24779683

Probability-enhanced sufficient dimension reduction for binary classification.

Seung Jun Shin1, Yichao Wu1, Hao Helen Zhang1, Yufeng Liu1.   

Abstract

In high-dimensional data analysis, it is of primary interest to reduce the data dimensionality without loss of information. Sufficient dimension reduction (SDR) arises in this context, and many successful SDR methods have been developed since the introduction of sliced inverse regression (SIR) [Li (1991) Journal of the American Statistical Association 86, 316-327]. Despite their fast progress, though, most existing methods target on regression problems with a continuous response. For binary classification problems, SIR suffers the limitation of estimating at most one direction since only two slices are available. In this article, we develop a new and flexible probability-enhanced SDR method for binary classification problems by using the weighted support vector machine (WSVM). The key idea is to slice the data based on conditional class probabilities of observations rather than their binary responses. We first show that the central subspace based on the conditional class probability is the same as that based on the binary response. This important result justifies the proposed slicing scheme from a theoretical perspective and assures no information loss. In practice, the true conditional class probability is generally not available, and the problem of probability estimation can be challenging for data with large-dimensional inputs. We observe that, in order to implement the new slicing scheme, one does not need exact probability values and the only required information is the relative order of probability values. Motivated by this fact, our new SDR procedure bypasses the probability estimation step and employs the WSVM to directly estimate the order of probability values, based on which the slicing is performed. The performance of the proposed probability-enhanced SDR scheme is evaluated by both simulated and real data examples.
© 2014, The International Biometric Society.

Entities:  

Keywords:  Binary classification; Conditional class probability; Fisher consistency; Sufficient dimension reduction; Weighted support vector machines (WSVMs)

Mesh:

Year:  2014        PMID: 24779683      PMCID: PMC4670268          DOI: 10.1111/biom.12174

Source DB:  PubMed          Journal:  Biometrics        ISSN: 0006-341X            Impact factor:   2.571


  3 in total

1.  ASYMPTOTIC PROPERTIES OF SUFFICIENT DIMENSION REDUCTION WITH A DIVERGING NUMBER OF PREDICTORS.

Authors:  Yichao Wu; Lexin Li
Journal:  Stat Sin       Date:  2011       Impact factor: 1.261

2.  Two-Dimensional Solution Surface for Weighted Support Vector Machines.

Authors:  Seung Jun Shin; Yichao Wu; Hao Helen Zhang
Journal:  J Comput Graph Stat       Date:  2014-04-03       Impact factor: 2.302

3.  PLS dimension reduction for classification with microarray data.

Authors:  Anne-Laure Boulesteix
Journal:  Stat Appl Genet Mol Biol       Date:  2004-11-23
  3 in total
  3 in total

1.  Comments on: Probability Enhanced Effective Dimension Reduction for Classifying Sparse Functional Data.

Authors:  Chong Zhang; Yufeng Liu
Journal:  Test (Madr)       Date:  2016-01-25       Impact factor: 2.345

2.  Principal weighted support vector machines for sufficient dimension reduction in binary classification.

Authors:  Seung Jun Shin; Yichao Wu; Hao Helen Zhang; Yufeng Liu
Journal:  Biometrika       Date:  2017-01-19       Impact factor: 2.445

3.  Receiver operating characteristic curves and confidence bands for support vector machines.

Authors:  Daniel J Luckett; Eric B Laber; Samer S El-Kamary; Cheng Fan; Ravi Jhaveri; Charles M Perou; Fatma M Shebl; Michael R Kosorok
Journal:  Biometrics       Date:  2020-09-12       Impact factor: 1.701

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.