Literature DB >> 28422666

Efficient kNN Classification With Different Numbers of Nearest Neighbors.

Shichao Zhang, Xuelong Li, Ming Zong, Xiaofeng Zhu, Ruili Wang.   

Abstract

nearest neighbor (kNN) method is a popular classification method in data mining and statistics because of its simple implementation and significant classification performance. However, it is impractical for traditional kNN methods to assign a fixed value (even though set by experts) to all test samples. Previous solutions assign different values to different test samples by the cross validation method but are usually time-consuming. This paper proposes a kTree method to learn different optimal values for different test/new samples, by involving a training stage in the kNN classification. Specifically, in the training stage, kTree method first learns optimal values for all training samples by a new sparse reconstruction model, and then constructs a decision tree (namely, kTree) using training samples and the learned optimal values. In the test stage, the kTree fast outputs the optimal value for each test sample, and then, the kNN classification can be conducted using the learned optimal value and all training samples. As a result, the proposed kTree method has a similar running cost but higher classification accuracy, compared with traditional kNN methods, which assign a fixed value to all test samples. Moreover, the proposed kTree method needs less running cost but achieves similar classification accuracy, compared with the newly kNN methods, which assign different values to different test samples. This paper further proposes an improvement version of kTree method (namely, k*Tree method) to speed its test stage by extra storing the information of the training samples in the leaf nodes of kTree, such as the training samples located in the leaf nodes, their kNNs, and the nearest neighbor of these kNNs. We call the resulting decision tree as k*Tree, which enables to conduct kNN classification using a subset of the training samples in the leaf nodes rather than all training samples used in the newly kNN methods. This actually reduces running cost of test stage. Finally, the experimental results on 20 real data sets showed that our proposed methods (i.e., kTree and k*Tree) are much more efficient than the compared methods in terms of classification tasks.

Year:  2017        PMID: 28422666     DOI: 10.1109/TNNLS.2017.2673241

Source DB:  PubMed          Journal:  IEEE Trans Neural Netw Learn Syst        ISSN: 2162-237X            Impact factor:   10.451


  45 in total

1.  Group sparse reduced rank regression for neuroimaging genetic study.

Authors:  Xiaofeng Zhu; Heung-Il Suk; Dinggang Shen
Journal:  World Wide Web       Date:  2018-09-17       Impact factor: 2.716

2.  Predicting Patient Length of Stay in Australian Emergency Departments Using Data Mining.

Authors:  Sai Gayatri Gurazada; Shijia Caddie Gao; Frada Burstein; Paul Buntine
Journal:  Sensors (Basel)       Date:  2022-06-30       Impact factor: 3.847

3.  A New Fault Diagnosis of Rolling Bearing Based on Markov Transition Field and CNN.

Authors:  Mengjiao Wang; Wenjie Wang; Xinan Zhang; Herbert Ho-Ching Iu
Journal:  Entropy (Basel)       Date:  2022-05-25       Impact factor: 2.738

4.  A Lightweight Pose Sensing Scheme for Contactless Abnormal Gait Behavior Measurement.

Authors:  Yuliang Zhao; Jian Li; Xiaoai Wang; Fan Liu; Peng Shan; Lianjiang Li; Qiang Fu
Journal:  Sensors (Basel)       Date:  2022-05-27       Impact factor: 3.847

5.  Parameter-Free Centralized Multi-Task Learning for Characterizing Developmental Sex Differences in Resting State Functional Connectivity.

Authors:  Xiaofeng Zhu; Hongming Li; Yong Fan
Journal:  Proc Conf AAAI Artif Intell       Date:  2018-04-26

6.  Brain functional connectivity analysis based on multi-graph fusion.

Authors:  Jiangzhang Gan; Ziwen Peng; Xiaofeng Zhu; Rongyao Hu; Junbo Ma; Guorong Wu
Journal:  Med Image Anal       Date:  2021-04-09       Impact factor: 8.545

7.  Integrative analysis of genomic, epigenomic and transcriptomic data identified molecular subtypes of esophageal carcinoma.

Authors:  Mingyang Ma; Yang Chen; Xiaoyi Chong; Fangli Jiang; Jing Gao; Lin Shen; Cheng Zhang
Journal:  Aging (Albany NY)       Date:  2021-02-26       Impact factor: 5.682

8.  A novel artificial intelligence protocol to investigate potential leads for Parkinson's disease.

Authors:  Zhi-Dong Chen; Lu Zhao; Hsin-Yi Chen; Jia-Ning Gong; Xu Chen; Calvin Yu-Chian Chen
Journal:  RSC Adv       Date:  2020-06-16       Impact factor: 4.036

9.  Improving Risk Identification of Adverse Outcomes in Chronic Heart Failure Using SMOTE+ENN and Machine Learning.

Authors:  Ke Wang; Jing Tian; Chu Zheng; Hong Yang; Jia Ren; Chenhao Li; Qinghua Han; Yanbo Zhang
Journal:  Risk Manag Healthc Policy       Date:  2021-06-08

10.  Machine learning classification of texture features of MRI breast tumor and peri-tumor of combined pre- and early treatment predicts pathologic complete response.

Authors:  Lal Hussain; Pauline Huang; Tony Nguyen; Kashif J Lone; Amjad Ali; Muhammad Salman Khan; Haifang Li; Doug Young Suh; Tim Q Duong
Journal:  Biomed Eng Online       Date:  2021-06-28       Impact factor: 2.819

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.