| Literature DB >> 30283157 |
Hanyang Peng1,2, Yong Fan3.
Abstract
A unified framework is proposed to select features by optimizing computationally feasible approximations of high-dimensional conditional mutual information (CMI) between features and their associated class label under different assumptions. Under this unified framework, state-of-the-art information theory based feature selection algorithms are rederived, and a new algorithm is proposed to select features by optimizing a lower bound of the CMI with a weaker assumption than those adopted by existing methods. The new feature selection method integrates a plug-in component to distinguish redundant features from irrelevant ones for improving the feature selection robustness. Furthermore, a novel metric is proposed to evaluate feature selection methods based on simulated data. The proposed method has been compared with state-of-the-art feature selection methods based on the new evaluation metric and classification performance of classifiers built upon the selected features. The experiment results have demonstrated that the proposed method could achieve promising performance in a variety of feature selection problems.Entities:
Keywords: Conditional mutual information; Feature selection; Lower Bound; Weak assumptions
Year: 2017 PMID: 30283157 PMCID: PMC6167022 DOI: 10.1016/j.ins.2017.08.036
Source DB: PubMed Journal: Inf Sci (N Y) ISSN: 0020-0255 Impact factor: 6.795