Literature DB >> 30964868

Classification of high dimensional biomedical data based on feature selection using redundant removal.

Bingtao Zhang1,2, Peng Cao3.   

Abstract

High dimensional biomedical data contain tens of thousands of features, accurate and effective identification of the core features in these data can be used to assist diagnose related diseases. However, there are often a large number of irrelevant or redundant features in biomedical data, which seriously affect subsequent classification accuracy and machine learning efficiency. To solve this problem, a novel filter feature selection algorithm based on redundant removal (FSBRR) is proposed to classify high dimensional biomedical data in this paper. First of all, two redundant criteria are determined by vertical relevance (the relationship between feature and class attribute) and horizontal relevance (the relationship between feature and feature). Secondly, to quantify redundant criteria, an approximate redundancy feature framework based on mutual information (MI) is defined to remove redundant and irrelevant features. To evaluate the effectiveness of our proposed algorithm, controlled trials based on typical feature selection algorithm are conducted using three different classifiers, and the experimental results indicate that the FSBRR algorithm can effectively reduce the feature dimension and improve the classification accuracy. In addition, an experiment of small sample dataset is designed and conducted in the section of discussion and analysis to clarify the specific implementation process of FSBRR algorithm more clearly.

Entities:  

Mesh:

Year:  2019        PMID: 30964868      PMCID: PMC6456288          DOI: 10.1371/journal.pone.0214406

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

The analysis of high dimensional disease data [1-2] is a very important research field, especially cancer [3], or mental disease (e.g. Depressive [4-5]). It is unrealistic to cure these diseases completely, so early diagnosis or prevention plays an important role in the treatment related disease. However, high dimension biomedical data usually contain a large number of weak relevant or irrelevant features. If all the features are treated equally, the time complexity, spatial complexity and accuracy of the prediction can be seriously affected. Therefore, feature selection is considered to be an essential step in the diagnosis of related disease using high dimension biomedical data. As is stated in [6], feature selection is also referred to as feature subset selection. The main purpose of feature selection is to remove irrelevant and redundant features in the classification process, while retaining the most valuable information of the original data. In other words, the objective of feature selection is to select an optimal feature subset [7] from the original feature set, which lays a good foundation for subsequent classification or learning work. As one of the important part of knowledge discovery technology, feature selection [8] can effectively improve the computing speed of subsequent prediction algorithm, enhance the compactness of the prediction model, increase the generalization ability of the corresponding model, and avoid over fitting. Based on the above factors, feature selection has always been a hot research topic, and new achievements are constantly emerging. For example, in [9], a feature selection method based on multi-objective binary based biogeography optimization (MOBBBO) is proposed for gene selection, which combines the non-dominated sorting method and the crowding distances method into the binary based biogeography optimization (BBBO) framework. In [10], a novel feature selection method via chaotic optimization is developed to solve the problem of balance between exploration of the search space and exploitation of the best solutions. In [11], Liu et al. used a discrete biogeography based optimization (DBBO) method by integrating discrete migration model and discrete mutation model for feature selection in molecular signatures. In [12], Li et al. proposed that a feature selection algorithm based on the multi-objective ranking binary artificial bee colony to select the optimal subset from the original high dimensional data while retaining a subset that satisfies the defined objective. In addition, recent advances on feature selection can be found in [13-14], in which [14] focuses on the review of the latest research work on evolutionary computation (EC) of feature selection, and identifies the contributions of these different algorithms. To better introduce the research status of high-throughput data or high-dimensional data based on feature selection algorithm [1, 15], the following is an overview of some representative studies in recent years. Tan et al. [1] proposed a new minimax sparse LR model for very high-dimensional feature selections, which can be efficiently solved by a cutting plane algorithm. In order to solve the problem of effectively identifying chromosome-wide spatial clusters from high-throughput chromatin conformation capture data, a population based optimization algorithm coordinates and guides the non-negative matrix factorization toward global optima was proposed in [16]. In [17], a novel feature selection method based on high dimensional model representation (HDMR) was proposed to solve the hyper-spectral image classification problem. The core idea of this method is to rank the global sensitivity index calculated via the HDMR to find the most relevant features. In order to explore and identify small clusters of spatially proximal genomic regions, Li et al. [18] proposed evolutionary computation methods to evolve and confirm functionally related genomic regions. Chen et al. [19] proposed a feature selection algorithm, which was named genetic programming (GP) with permutation importance (GPPI), to select features of high-dimensional symbolic regression (SR) using GP. Based on two typical applications of microarray analysis and target detection, Augusto et al. [20] discussed the feature selection of high-dimensional spatial data. In order to solve the problem of high-dimensional data classification, Zhang [21] proposed an improved artificial bee colony (ABC) algorithm to select the optimal feature subset. Meanwhile, to improve the convergence of ABC, the modified ABC algorithm (named OGR-ABC algorithm) introduces three modified strategies including opposite initialization, global optimum based search equations and ranking based selection mechanism. Through the analysis of the above research, it is not difficult to find that feature selection process mainly includes two steps [22]: search strategy and evaluation criterion. Based on whether or not the classifier itself is used as feature evaluation index, evaluation criterion can be categorized into the wrapper method and the filter method. The wrapper method [23, 24] to evaluate superiority and inferiority of the optimal feature subset under the premise of the classification algorithm unchanged. Meanwhile, the corresponding classification accuracy is adopted as an index to select optimal feature subset. So the feature subset selected by the wrapper method is not universal. It is necessary to execute the feature selection process again when the classification algorithm is changed. Therefore, its time complexity is too high, especially for high dimension data, and the execution time of the algorithm may be longer. Another evaluation criterion based on the filter method [25-27], the search of feature space depends on the intrinsic correlation of the data itself rather than the classification algorithm. The filter method is increasingly attractive because of its simplicity and fast speed. Therefore this method is more popular than the wrapper method. Four intrinsic correlation metrics are often adopted by the filter method to evaluate feature subset, including MI [28], fractal dimension [29], dependency degree [30] and distance [31]. Among them, MI is considered as the most acceptable criteria due to two major advantages [32]: (1) Measuring different relationship between nonlinear (random) variables. (2) Preserving stability of transformations in the high dimensional feature space that is invertible. According to the above analysis, a filter feature selection method is proposed for high dimensional biomedical data based on redundant removal in this paper. Firstly, we analyze the four boundary extremes of correlation between feature and target class and the correlation between feature and feature. Based on this, two redundant criterions are proposed. And then in order to quantify the redundancy criterion, the core module based on MI is proposed: the definition of approximate redundancy feature. Finally, the algorithm implementation is given.

Mathematical symbols and basic concepts

Mathematical symbols

There are many mathematical symbols used in this study. To improve the readability, we list these mathematical symbols and their abbreviations in below. P:a probability measure. F: feature set, F = {F1,F2,…,F,…,F}. F: F = {fi,1, f,…,f}. A: A = F −{F}. C: class attribute, C = {C1,C2,…,C,…,C}. R: the relevance between any two variables. R: the relevance between any pair of feature F and F,i≠j. R: the relevance between any feature F and class attribute C. Rmax: the maximum value of R. : the mean value of Ri,c, that is .

Basic concepts

For laying a base for further investigation, three basic concepts [33] (strongly relevant feature, weakly relevance feature, and irrelevant feature) used in this study are listed as follows. Strong relevance: F is strongly relevant feature iff there exists P(F,A)>0 such that Weak relevance: F is weakly relevant feature iff it is not strongly relevant (i.e. P(C|F,A) ≠ P(C|A)), there exists and such that Irrelevance: F is irrelevant feature iff it are not strongly relevant and weakly relevant, there all and such that Strong relevance indicates that the feature is very important for classification accuracy improvement, so it can’t be arbitrarily removed. Weak relevance indicates that the feature can sometimes contribute to improve prediction accuracy. Irrelevance indicates that the feature is useless on the improvement of classification accuracy, so it can be directly deleted.

Method

Determination of redundancy criterion

It is difficult to determine that the complete correlation between any pair of feature in the actual calculation process, and then determine whether there is redundancy among features. To combat this, a redundancy criterion based on the correlation is proposed in this study to lay the foundation for further feature selection. Based on three basic concepts, the redundancy of feature F is analyzed under different extreme values of R and R. Different cases of extreme value are shown in Table 1.
Table 1

Different cases of extreme value.

Ri,cRi,j
Case 1largelarge
Case 2largesmall
Case 3smalllarge
Case 4smallsmall
It is easy to draw the following four conclusions after analyzing four cases: Conclusion 1: R is large, which means that F contains more information about C. R is large, which means that the correlation between F and F is strong. If R = 1, then F and F is complete correlation, hence F is redundant. If R≠1, it is difficult to determine the feature F whether or not is redundant. Conclusion 2: R is small, which means that the correlation between F and F is weak. Hence F can’t replace F. In other words, no matter the size of the R, the feature F is not redundant. Conclusion 3: R is small, which means that F contains less information about C. R is large, which means that the correlation between F and F is strong. In this case, the feature F is redundant with higher probability. With the increase of R, this probability is also increasing. Conclusion 4: R is small, which means that the correlation between F and F is weak. This conclusion is consistent with the conclusions 2, no matter the size of the R, the feature F is not redundant. Based on the above four conclusions, two redundant criteria can be obtained: Criteria 1: when R is large, whether F is redundant is uncertain. Criteria 2: when R is small, no matter the size of the R, the feature F is not redundant.

Definition approximate redundancy feature

Based on the two redundant criterions inthe previous section, the approximate redundancy feature is proposed and defined in this section. Assuming that the R of the feature F is very close to Rmax, it indicates that F contains a lot of information about class attribute C. In this condition, only if the value of R is large enough, F can be considered as an approximate redundancy feature. Otherwise, it can’t be considered as redundancy feature. The reason is that F plays an important role in improving the accuracy of classification, and can’t be easily removed as redundancy. By contrast, Assuming that the R of feature F is not very close to Rmax, it indicates that F contains relatively less information about class attribute C. In this condition, as long as the value of R is relatively large, F is considered as an approximate redundancy feature. The reason is that F not plays a main role in improving the accuracy of classification. Based on the above analysis and discussion, the approximate redundant feature is formally described in definition 1. Definition 1 (approximate redundancy feature): There is any pair of correlation feature F and F, and RR. (1) F is an approximate redundancy feature iff there exists, Rmax-R≤δ, 0.05≤δ≤0.13, such that (2) F is an approximate redundancy feature iff there exists R − R > δ && R − R ≤ α, 0.05≤δ≤0.13, 0.60≤α≤0.66, such that Definition 1indicates that F can be approximated as an alternative for F.

Correlation calculation

In general, correlation measure methods include linear and nonlinear. A nonlinear method based on MI is adopted in this study, and the reason is that the high dimensional biomedical data usually exist in the form of nonlinear in the real world. The correlation between any pair of variables (X, Y) can be calculated in the following formulas (6) or (7). where H(X), H(X|Y) and H(X,Y) can be calculated on the basis of formulas (8), (9) and (10). According on the above three formulas, the value of H(X|Y) or H(X,Y) is smaller when Y contains more information about X. In other words, the greater value of IG(X;Y), which means there are more relevant between X and Y. To prevent the scale of data is not unified and to reduce the effect of extreme value, each IG(X;Y) is normalized to the range [0, 1] using formula (11).

Algorithm implementation

Based on the definition of the approximate redundancy feature and the correlation calculation method, a feature selection algorithm for high dimension biomedical data classification based on redundant removal (FSBRR) is given in Algorithm 1. Algorithm 1: FSBRR Input: Feature set: F = {F1,F2,…,F,…,F}; Class label: C = {C1,C2, …, Ci…,C}; Parameter: τ, δ, α. 1.for i = 1:n 2. = 2 * (;)/(() + ()); 3.if ≥                 % a preset threshold value, remove irrelevance features 4.Addto(F’, F); 5.Addto(R’, R); 6.end 7. end 8. ;                  % the mean value of R, 1≤i≤n 9. [X, I] = sort(R, ‘descend’);                      % where RϵR’ 10.F’ = F’(I);                                                % order F’ in descending R value 11 for i = 1:size(I,2)-1; 12.F = F’(:,i); R;                        % to select current first feature in each cycle, i.e. first variable 13. for j = i+1: size(I,2) 14.F = F’(:,j);                        % to select next feature (or variable) 15. = 2 * (;)/(() + ()) 16.         = 2 * (;)/(() + ()) 17.            if F ≠ Null 18.                if R- R≤δ&&R≥R% 0.05≤δ≤0.13 19.                    remove(F’, F); % removing approximate redundancy features 20.else if R- R>δ&&R- R<α&&R> (+R)/2% 0.05≤δ≤0.13, 0.60≤α≤0.66 21.                    remove(F’, F); 22.                end 23.            end 24.        end 25.end 26.Foptimal = F’; Output: Foptimal The time consumption of FSBRR algorithm is mainly used to calculate R and R, so its atomic operation is the calculation of R and R. Assuming that a dataset contains n features, the time complexity of this algorithm used for removing irrelevant feature is linear time order O(n) (line 1 to line 7). For the removing approximate redundant feature and approximate irrelevant feature (line 11 to line 25): In the worst case, the time complexity is square time order O(n), and all features are not redundant features at this time. In the best case, the time complexity is linear time order O(n), and except for the first feature, the remaining n-1 features are redundant features at this time.

Performance evaluation function

In this study, classification accuracy and the number of selected features are used to design the performance evaluation function [34-35], which is shown in formula (12). n is the number of selected features and N is the total number of features. w1 and w are predefined weight coefficients, which are used to adjust the importance of two indicators in the performance evaluation function. In this study, the values of w1 and w are set to 0.999 and 0.001 respectively. The main reason for this setting can be attributed to the following three aspects: (1) Under the prerequisite of data dimensionality reduction, this study main focuses on the use of classification accuracy as a metric of feature selection algorithm. (2) The number of selected features is significantly reduced, but the classification accuracy is not improved, such dimensionality reduction of high-dimensional data will lose its original application value. (3) The performance evaluation function with high weight coefficient of classification accuracy and low weight coefficient of the number of selected features has been recognized and widely applied in many feature selection studies, such as [34]. In addition, Acc is classification accuracy as defined in formula(13). C and I are the number of correct and incorrect classification labels respectively.

Experiments

Data description

Eight well-known biomedical datasets (Table 2) were used to evaluate the performance of FSBRR algorithm. These dataset includes eight aspects of disease diagnosis data. The data dimension range was from 319 to 21548. The first three datasets were taken from the Kent Ridge Biomedical [36]. p53 Mutants and Arcene were taken from the UCI dataset [37]. Breast invasive carcinoma (BRCA), Glioblastoma multiforme (GBM), and tumour sequencing project (TSP) were taken from the TCGA [38].
Table 2

High dimension biomedical datasets.

DatasetAttributesInstancesClasses
ColonTumor2000622
Nervous-System7129602
DLBCL-Stanford4026472
p53 Mutants5409167722
Arcene100002002
BRCA2154810972
GBM183485282
TSP3191632

Experimental design

To evaluate the performance of FSBRR algorithm, under the same conditions, we designed and conducted the following experiments: eight high dimensional biomedical data were compared and analyzed by FSBRR, Relief (a filter methodbased on the nearest neighbor distance) [39], maximum relevance and minimum redundancy (mRmR) [40] and genetic algorithm (GA) [41], respectively. In this experiment, the same conditions contain two meanings: (1) Random forest (RF, numTrees = 10), K-nearest neighbor (KNN, k = 1), and Support Vector Machine (SVM, Linear Kernel) were adopted as classifier to evaluate classification performance. (2) In FSBRR algorithm, the parameter τ is set to 0, and the purpose is that to avoid losing the weak correlation feature without prior knowledge. In addition, after adaptive testing, the values of δ and α were set to 0.08 and 0.64 respectively. To obtain an unbiased experimental result, 10 fold cross validation was adopted to evaluate the classification performance. Each dataset was stratified into 10 folds, of which 9 folds were used as a training sample and the remaining 1 fold was used as a testing sample. Moreover, in order to get a statistically meaningful result, each experiment was executed 100 times independently. This means that the classification task is executed 1000 times in total, and the average value is taken as the result in finally. The above experiments were implemented in Matlab 2017a. The experimental hardware and software configuration is shown in Table 3.
Table 3

Hardware and software configuration of experimental.

No.ComponentsParameters
1CPUIntel i7-8550U 4.0GHz
2RAM16G
3Operating system (OS)Windows 7
4Software platformMatlab 2017a

Results

For the eight datasets, we have conducted the experiments described in above section. The six major statistical indicators were compared and analyzed, and the results were shown in Table 4.
Table 4

Experimental results based on eight data sets.

ClassifierDatasetAlgorithmMean (%)Max(%)Min(%)StdMeanFNRT(s)
RFColonTumorFull Set78.2182.3271.463.512000
FSBRR92.0195.4591.211.97341.04
Relief85.4989.9481.233.16382.19
mRmR86.6190.7282.773.85422.75
GA89.1395.7184.123.913619.18
Nervous-SystemFull Set61.3374.6154.865.687129
FSBRR80.1784.0274.643.02375.23
Relief71.5681.2468.883.974414.72
mRmR72.4276.4369.613.244216.37
GA75.7679.2169.563.874890.01
DLBCL-StanfordFull Set76.6080.2173.792.874026
FSBRR82.9984.1679.302.33292.41
Relief75.4881.6471.633.34453.84
mRmR76.5884.6169.156.42544.10
GA79.4682.4675.163.129131.80
p53 MutantsFull Set89.6692.1881.613.425409
FSBRR94.3198.0090.563.19392.49
Relief90.4294.2584.254.06544.14
mRmR87.9191.1479.795.48374.52
GA94.1698.3191.213.016437.44
ArceneFull Set73.1480.6066.604.8710000
FSBRR85.6787.3082.202.115112.34
Relief81.0184.0177.782.067324.94
mRmR78.8084.0272.024.827120.34
GA77.1480.0274.313.1691128.67
BRCAFull Set80.5785.5176.993.2921548
FSBRR86.2689.2481.142.5114818.84
Relief85.2290.1079.444.0124126.25
mRmR83.5485.9280.362.2118924.01
GA85.1688.9081.122.6224629.60
GBMFull Set69.9280.1662.778.4518348
FSBRR80.9586.7875.493.95614.56
Relief76.1482.1570.654.29534.87
mRmR74.9079.1269.443.96685.42
GA82.0390.2575.025.489312.84
TSPFull Set68.7775.8763.214.26319
FSBRR78.6381.2376.991.831080.47
Relief67.0169.8961.323.011350.75
mRmR56.9662.4250.424.121240.50
GA67.2670.5160.233.961370.80
KNNColonTumorFull Set75.8081.3271.023.692000
FSBRR91.9195.0188.092.38340.75
Relief83.4188.9178.133.76381.87
mRmR84.6989.1279.583.88422.04
GA86.6788.3382.332.53389.16
Nervous-SystemFull Set56.8669.6653.944.177129
FSBRR74.2481.2570.113.22372.88
Relief65.6970.8963.283.36449.14
mRmR65.7769.1262.143.28428.31
GA73. 6276.5869.142.935673.82
DLBCL-StanfordFull Set78.2482.4662.365.834026
FSBRR83.6888.7879.473.42291.88
Relief76.3284.4771.954.73452.43
mRmR76.9887.7171.545.86543.01
GA80.0682.7872.833.799724.31
p53 MutantsFull Set84.9390.1580.614.015409
FSBRR88.3092.0185.502.09392.04
Relief85.2089.2381.992.45543.78
mRmR84.2787.5080.092.29374.14
GA89.9095.1487.042.995639.77
ArceneFull Set67.6077.3862.574.3810000
FSBRR85.6789.3079.203.02514.98
Relief82.0187.0179.042.89739.04
mRmR78.9386.3375.184.34719.79
GA79.1384.7875.903.418784.46
BRCAFull Set78.5183.4272.814.8721548
FSBRR84.4287.2480.022.7114810.19
Relief83.7587.0674.595.2324119.10
mRmR82.4387.8875.314.8518920.21
GA83.0187.1178.013.0225162.62
GBMFull Set68.8280.6465.554.5818348
FSBRR80.1287.0778.302.87611.85
Relief74.8580.0270.883.81532.42
mRmR74.8280.2566.405.56682.87
GA81.3988.5175.243.95885.50
TSPFull Set62.7274.8761.214.86319
FSBRR77.6982.3176.462.161080.46
Relief61.7569.2360.863.231350.69
mRmR57.8763.7953.184.271240.51
GA69.5671.8866.812.261390.81
SVMColonTumorFull Set73.8680.6370.663.982000
FSBRR87.6891.4683.213.48341.12
Relief80.4684.1875.444.23383.22
mRmR81.7885.7175.814.58423.96
GA84.1689.6681.624.404828.86
Nervous-SystemFull Set53.3465.1550.866.587129
FSBRR73.1878.1270.693.28378.78
Relief68.5672.6162.813.924417.89
mRmR65.7872.4861.155.024219.96
GA68.1672.1163.563.524542.11
DLBCL-StanfordFull Set70.1679.2160.795.864026
FSBRR78.9183.4574.414.31292.51
Relief72.9674.6162.695.95454.41
mRmR72.1577.6569.153.82543.25
GA72.8577.4568.063.859112.54
p53 MutantsFull Set85.5687.7680.134.115409
FSBRR86.2590.0282.153.11392.35
Relief84.1587.2680.363.09542.85
mRmR82.1587.5876.054.85373.91
GA84.5588.1278.513.986120.51
ArceneFull Set70.5674.1566.583.5510000
FSBRR80.6184.5278.153.54515.55
Relief76.5980.1270.543.85738.52
mRmR74.7579.6570.254.65718.15
GA71.3575.8565.354.0511555.85
BRCAFull Set77.2580.2474.213.1721548
FSBRR82.0286.5580.851.801487.56
Relief79.0585.8976.193.292418.58
mRmR78.4986.0171.715.851899.77
GA79.9686.3675.453.252628.40
GBMFull Set65.1271.5561.103.4018348
FSBRR80.0985.5577.352.01616.45
Relief74.0178.2371.082.75537.88
mRmR73.9979.0070.593.44688.58
GA80.0387.1076.403.361019.51
TSPFull Set64.4572.1460.455.84319
FSBRR72.3576.1869.982.251080.42
Relief62.6567.1858.103.941350.70
mRmR56.9061.5451.954.751240.52
GA64.6772.0960.465.181400.75

Note: (1) full set: a set of all features that have not been processed by feature selection algorithm. (2) Mean (%): the mean of performance. (3) Max(%): the highest of performance. (4) Min(%): the lowest of performance. (5) Std: the standard deviation. (6) MeanFN: the mean number of selected feature. (7) RT(s): running time, unit is second.

Boldface indicates the best experimental result.

Note: (1) full set: a set of all features that have not been processed by feature selection algorithm. (2) Mean (%): the mean of performance. (3) Max(%): the highest of performance. (4) Min(%): the lowest of performance. (5) Std: the standard deviation. (6) MeanFN: the mean number of selected feature. (7) RT(s): running time, unit is second. Boldface indicates the best experimental result. From Table 4, we can observe the following aspects: (1)The FSBRR algorithm obtained the highest Mean among all the four feature selection algorithms for twenty-one out of twenty-four experimental results. The highest Mean(s) were 92.01%, 80.17%, 82.99%, 94.31%, 85.67%, 86.26%, 78.63%, 91.91%, 74.24%, 83.68%, 85.67%, 84.42%, 77.69%, 87.68%, 73.18%, 78.91%, 86.25%, 80.61%, 82.02%, 80.09%, and 72.35% respectively. Meanwhile, we notice that the maximum Mean improvement of FSBRR was 19.84% compared with the full set. (2) RF-based GA algorithm uses GBM dataset, and KNN-based GA algorithm uses p53 Mutants and GBM datasets to obtain highest Mean of 82.03%, 89.90%, and 81.39%, respectively. However, Std, MeanFN and RT of GA were significantly higher than FSRRR in these three experiments. (3) The Std among for eighteen out of twenty-four experimental results obtained by FSBRR was smaller than other three algorithms. (4) Four feature selection algorithms can effectively reduce the feature dimension, but the dimension reduction of the FSBRR algorithm was the most obvious. In FSBRR, Relief, mRmR and GA, GA belongs to the wrapper feature selection algorithm, so there were differences in the number of feature subsets for RF, KNN and SWM classifiers. (5) In most experiments, except for the running time index, the other performances of RF were significantly better than the KNN and SVM for the same dataset. Such results indicate that for the specified dataset, to get the best experimental results, a matching classification (or learning) algorithm must be found. Fig 1 was obtained by statistical analysis of Table 4. It shows four average attribute values (avg(Mean), avg(Std), avg(MeanFN), and avg(RT)) of four feature selection algorithms. From Fig 1 we can observe that performance, stability, the feature number of optimal subset, and time complexity of FSBRR was superior to the other three algorithms. We believe that the reasons for obtaining this experiment were: (1) FSBRR algorithm uses R to explore the horizontal correlation between features and features, and the longitudinal correlation between features and classes was explored by R. Meanwhile, based on the basic relevance theory, the horizontal relevance and the longitudinal relevancewere effectively combined. (2) FSBRR algorithm not only removes the irrelevant features, but also removes the approximately redundant features.
Fig 1

The average attributes value of four feature selection algorithms based on eight datasets.

Discussion and analysis

To verify the effect of parameters δ and α on the performance of FSBRR algorithm, the classification accuracy of 8 datasets was tested on RF, KNN, and SVM, respectively. In the first experiment, when the parameter α = 0.6 keep constant, the parameter δ increased from 0 to 0.2 with a step length of 0.01. In the second experiment, when the parameter δ = 0.1 keep constant, the parameter α increased from 0.5 to 0.7 with a step length of 0.01. Other experimental procedures are described in experimental design section. Moreover, to facilitate discussion and analysis, the classification accuracy was only considered in here. The results of the two experiments were shown in Figs 2 and 3. Statistical analysis of the experimental results in Figs 2 and 3 reveals that: when the highest classification accuracy was obtained for eight datasets, although the values of parameters were different, their ranges of value were overlapped. By selecting the results of the top 20% of the classification accuracy, we can clearly observe that the saliency overlap of the parameter ranges, and the results were shown in Fig 4. The optimum range of parameter δ is [0.05, 0.13], and the optimum range of parameter α is [0.60, 0.66]. See the dotted line marker range in the Fig 4.
Fig 2

The relationship between parameter δ and classification accuracy.

Fig 3

The relationship between parameter α and classification accuracy.

Fig 4

The range of parameter values when the classification accuracy is located in the top 20% based on FSBRR algorithm.

The approximate optimal feature subset of each original dataset may be 1 or more. The purpose of feature selection is to find one of the optimal subset. However, the dimensions of the optimal subset for different original datasets are different, and the correlation values (R and R) distribution of different datasets is vary, which leads to the difference of parameter values (δ or α). According to the statistical analysis of the above two experimental results, the parameters δ and α should be selected within the range in [0.05, 0.13], and the range in [0.60, 0.66], respectively. Because the classification accuracy may be reach the highest value in this range. In order to further analysis and interpretation of our proposed algorithm, a Breast cancer Wisconsin (Diagnostic) [42] dataset from UCI was used in this subsection. This dataset contains two kinds of data: malignant and benign. It was composed of 569 instances, and each instance contains 32 attributes. For FSBRR algorithm, the relationship between relevance and redundant features was analyzed using RF algorithm. Other experimental procedures are described in experimental design section. Besides, to facilitate discussion and analysis, the classification accuracy was only considered in here. The main results of this experiment were listed in Table 5. For Table 5 there were two attributes need to be explained which are: (1) "Feature set", for example, {5} was the subscript of the feature, which is the fifth feature. (2) "Change" was the accuracy change based on the current feature set. From Table 5 we can observe the following points: (1) {5} as the current feature set, the accuracy rate increases after adding feature 7 (see the third row), in contrast, the accuracy rate reduces after adding feature 21 (see the fourth row). These results verify the redundancy Criterion 1 (when R is large, whether F is redundant is uncertain). In this case, the approximate redundant feature definitionis needed to determine whether the current feature is redundant. (2) {5, 7} as the current feature set, the accuracy rate increases after adding feature 18 (see the fifth row); {8, 3} as the current feature set, the accuracy rate increases after adding feature 30 (see the seventh row). These results verify the redundancy Criterion 2 (when R is smaller, no matter the size of the R, the feature F is not redundant).
Table 5

The relationship between relevance and redundant feature based on FSBRR algorithm.

Feature setAccuracy (%)ChangeRi,cRi,c-valueRi,jRi,j-value
{5}83.30R5,c0.9157
{5,7}87.12R7,c0.9011R5,70.9454
{5,21}83.16R21,c0.8045R5,210.9101
{5,7,18}87.64R18,c0.4562R5,18 /R7,180.2131/0.1983
{8,3}84.50R8,c / R3,c0.8934/0.9013
{8,3,30}87.86R30,c0.7120R8,30 /R3,300.3010/0.2139

Conclusions

In this paper, the relationship between two kinds of correlation (R and R) is established, which effectively combines the correlation between features and classes and the correlation between features and features to eliminate redundant features. Because the determination of completely redundant features in actual operation is difficult to realize, so we first analyze four kinds of boundary conditions between R and R, and then a redundancy feature criteria is proposed. On this basis, the approximate redundancy features are defined in this study. Finally, we have proposed a new feature selection algorithm based on approximate redundancy removal (FSBRR) for high dimensional biomedicine data classification. To verify the effectiveness of the FSBRR algorithm, three classification algorithms (RF, KNN and SVM) are used to compare the FSBRR and three typical feature selection algorithms on eight high dimension biomedical datasets. The experiment results show that the FSBRR algorithm can effectively remove redundant features to improve the classification performance. Additionally, we also designed a set of comparative experiments to discuss and analyze the effects of parameters δ and α on the performance of FSBRR algorithm.
  14 in total

1.  Feature selection based on mutual information: criteria of max-dependency, max-relevance, and min-redundancy.

Authors:  Hanchuan Peng; Fuhui Long; Chris Ding
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2005-08       Impact factor: 6.226

2.  Normalized mutual information feature selection.

Authors:  Pablo A Estévez; Michel Tesmer; Claudio A Perez; Jacek M Zurada
Journal:  IEEE Trans Neural Netw       Date:  2009-01-13

3.  A Comparative Study for Identifying the Chromosome-Wide Spatial Clusters from High-Throughput Chromatin Conformation Capture Data.

Authors:  Xiangtao Li; Ka-Chun Wong
Journal:  IEEE/ACM Trans Comput Biol Bioinform       Date:  2017-03-20       Impact factor: 3.710

4.  Feature Selection Based on High Dimensional Model Representation for Hyperspectral Images.

Authors:  Gulsen Taskin; Huseyin Kaya; Lorenzo Bruzzone
Journal:  IEEE Trans Image Process       Date:  2017-03-24       Impact factor: 10.856

5.  Feature Selection for Optimized High-Dimensional Biomedical Data Using an Improved Shuffled Frog Leaping Algorithm.

Authors:  Bin Hu; Yongqiang Dai; Yun Su; Philip Moore; Xiaowei Zhang; Chengsheng Mao; Jing Chen; Lixin Xu
Journal:  IEEE/ACM Trans Comput Biol Bioinform       Date:  2016-08-24       Impact factor: 3.710

6.  Multiobjective binary biogeography based optimization for feature selection using gene expression data.

Authors:  Xiangtao Li; Minghao Yin
Journal:  IEEE Trans Nanobioscience       Date:  2013-12       Impact factor: 2.935

7.  Minimax sparse logistic regression for very high-dimensional feature selection.

Authors:  Mingkui Tan; Ivor W Tsang; Li Wang
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2013-10       Impact factor: 10.451

8.  Evolving Spatial Clusters of Genomic Regions From High-Throughput Chromatin Conformation Capture Data.

Authors:  Xiangtao Li; Shijing Ma; Ka-Chun Wong
Journal:  IEEE Trans Nanobioscience       Date:  2017-07-11       Impact factor: 2.935

9.  Molecular classification of endometriosis and disease stage using high-dimensional genomic data.

Authors:  John S Tamaresis; Juan C Irwin; Gabriel A Goldfien; Joseph T Rabban; Richard O Burney; Camran Nezhat; Louis V DePaolo; Linda C Giudice
Journal:  Endocrinology       Date:  2014-09-22       Impact factor: 4.736

10.  Feature Selection via Chaotic Antlion Optimization.

Authors:  Hossam M Zawbaa; E Emary; Crina Grosan
Journal:  PLoS One       Date:  2016-03-10       Impact factor: 3.240

View more
  2 in total

1.  Big data and artificial intelligence (AI) methodologies for computer-aided drug design (CADD).

Authors:  Jai Woo Lee; Miguel A Maria-Solano; Thi Ngoc Lan Vu; Sanghee Yoon; Sun Choi
Journal:  Biochem Soc Trans       Date:  2022-02-28       Impact factor: 4.919

2.  A framework model using multifilter feature selection to enhance colon cancer classification.

Authors:  Murad Al-Rajab; Joan Lu; Qiang Xu
Journal:  PLoS One       Date:  2021-04-16       Impact factor: 3.240

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.