| Literature DB >> 33824723 |
Hang Yu1, Yuanjia Wang2, Donglin Zeng3.
Abstract
With growing interest to use black-box machine learning for complex data with many feature variables, it is critical to obtain a prediction model that only depends on a small set of features to maximize generalizability. Therefore, feature selection remains to be an important and challenging problem in modern applications. Most of existing methods for feature selection are based on either parametric or semiparametric models, so the resulting performance can severely suffer from model misspecification when high-order nonlinear interactions among the features are present. A very limited number of approaches for nonparametric feature selection were proposed, but they are computationally intensive and may not even converge. In this paper, we propose a novel and computationally efficient approach for nonparametric feature selection in regression field based on a tensor-product kernel function over the feature space. The importance of each feature is governed by a parameter in the kernel function which can be efficiently computed iteratively from a modified alternating direction method of multipliers (ADMM) algorithm. We prove the oracle selection property of the proposed method. Finally, we demonstrate the superior performance of our approach compared to existing methods via simulation studies and application to the prediction of Alzheimer's disease.Entities:
Keywords: Alternating direction method of multipliers; Fisher consistency; Oracle property; Reproducing kernel Hilbert space; Tensor product
Year: 2020 PMID: 33824723 PMCID: PMC8021131 DOI: 10.1002/sta4.300
Source DB: PubMed Journal: Stat (Int Stat Inst) ISSN: 2049-1573