Literature DB >> 34198638

Feature Ranking and Screening for Class-Imbalanced Metabolomics Data Based on Rank Aggregation Coupled with Re-Balance.

Guang-Hui Fu1, Jia-Bao Wang1, Min-Jie Zong1, Lun-Zhao Yi2.   

Abstract

Feature screening is an important and challenging topic in current class-imbalance learning. Most of the existing feature screening algorithms in class-imbalance learning are based on filtering techniques. However, the variable rankings obtained by various filtering techniques are generally different, and this inconsistency among different variable ranking methods is usually ignored in practice. To address this problem, we propose a simple strategy called rank aggregation with re-balance (RAR) for finding key variables from class-imbalanced data. RAR fuses each rank to generate a synthetic rank that takes every ranking into account. The class-imbalanced data are modified via different re-sampling procedures, and RAR is performed in this balanced situation. Five class-imbalanced real datasets and their re-balanced ones are employed to test the RAR's performance, and RAR is compared with several popular feature screening methods. The result shows that RAR is highly competitive and almost better than single filtering screening in terms of several assessing metrics. Performing re-balanced pretreatment is hugely effective in rank aggregation when the data are class-imbalanced.

Entities:  

Keywords:  class-imbalance; feature screening; filtering algorithm; rank aggregation; re-balance

Year:  2021        PMID: 34198638      PMCID: PMC8232202          DOI: 10.3390/metabo11060389

Source DB:  PubMed          Journal:  Metabolites        ISSN: 2218-1989


1. Introduction

Datasets with imbalanced distribution are quite common in classification. In the settings of binary category, a dataset is called “imbalanced” if the number of one class is far larger than the others in the training data. Generally, the majority class is called negative while the minority class is called positive. Thus, the number of positive instances is often much lower than that of negative ones. A hindrance in class-imbalance learning is that standard classifiers are often biased towards the majority classes. Therefore, there is a higher misclassification rate in the minority instances [1,2]. Re-sampling is the standard strategy to deal with class-imbalance learning tasks. Many studies [2,3,4] have shown that re-sampling the dataset is an effective way to enhance the overall performance of the classification for several types of classifiers. Re-sampling methods concentrate on modifying the training set to make it suitable for a standard classifier. There are generally three types of re-sampling strategies to balance the class distribution: over-sampling, under-sampling, and hybrid sampling. Over-sampling adds a set sampled from the minority class. Randomly duplicating the minority instances, SMOTE [5] and smoothed bootstrap [6] are three widely used over-sampling methods. Under-sampling removes some of the data points from the majority class to alleviate the harms of imbalanced distribution. Random under-sampling (RUS) is a simple but effective way to randomly remove part of the majority class. Hybrid-sampling is a combination of over-sampling and under-sampling. Let D be a dataset with p features , the target of feature screening is to extract a part of features such that and these selected features satisfy the specified conditions of the task at hand [7]. For instance, the target is to select the subset of candidate features to maximize classifier accuracy in a classification setting. In the past two decades, many papers in studies have adopted the feature screening methods [8,9,10]. Feature screening has many advantages such as reducing susceptibility to over-fitting, training models faster and offsetting the pernicious effects of the curse of dimensionality [8]. The disadvantage of feature screening is that some crucial features may be omitted, thus harming classification performance. Filtering [11], wrapping [12], and embedding [13] are three kinds of approaches for feature screening. Filter algorithms screen top-ranked variables via a certain metric. Wrapper methods perform a search in all the combinations to find the best subsets of all features. Generally, a complete search is often time-consuming and greedy, so the heuristic technique is frequently utilized to explore the solutions. Embedded algorithms screen important variables while building the classifier. Of all the three types of feature screening, filter methods are the simplest and the most frequently used to solve real-world imbalanced problems [14] in class-imbalance learning community. Many metrics have been utilized to perform filtering feature screening algorithms, such as t test, Fisher score [15], Hellinger distance [16], Relief [17], ReliefF [18], information gain [19], Gini index [20], [21], [22], geometric mean [23], F-measure [24], and R-value [25]. Ensemble feature selection has been widely applied to the field of classification [26], such as Nazrul et al. [27] provided an ensemble feature selection method using feature–class and feature mutual information to select an optimal subset of features by combining multiple subsets of features. Yang et al. [28] proposed an ensemble-based wrapper approach for feature selection from data with highly imbalanced class distribution. Nowadays, feature selection methods are popular in metabolomics data analysis. In order to resolve the problem of filtering the discriminative metabolites from high-dimension metabolomics data, Lin et al. [29] proposed a mutual information (MI)-SVM-RFE method that filters out noise and non-informative variables by means of artificial variables and MI, then conducts SVM-RFE to select the most discriminative features. Fu et al. [30] proposed two feature selection algorithms that, by minimizing the overlap degree between the majority and the minority, are effective in recognizing key features and control false discoveries for class-imbalanced metabolomics data. The above feature screening methods are usually established for balanced datasets, but they are also directly utilized in class-imbalance situations. Different filtered approaches give different feature rankings because of their different theories, even when just counting top-ranked features. Motivated by this problem, we propose a simple strategy called rank aggregation with re-balance (RAR) to combine all methods’ ranking results in this study. It is an essential tool to fuse each rank to generate a synthetic rank that takes every ranking into account for class-imbalanced data. Different from the general feature selection methods, the proposed method combines different feature selection methods rather than simply accepting the result of one method, which can enhance the stability of the algorithm. At the same time, the great performances of the experiments in balanced and imbalanced metabolomics datasets verify the strong generalization abilities of RAR.

2. Results

2.1. Kendall’s Rank Correlation of Eight Filtering Methods on Class-Imbalanced Data

Each filtered method above can be employed to perform feature screening. However, we noted that different filtering feature screening techniques may give different rankings, especially when the data are extremely class-imbalanced. In this section, we compare methods using Kendall’s rank correlation [31]. The Kendall’s rank correlation of eight filtering methods (t test, Fisher score, Hellinger distance, Relief, ReliefF, Information gain, Gini index, and R-value) are computed with simulated data that are generated by multivariate normal distributions, namely, and , where the label denotes the majority class and minority class, respectively. The predictors in two classes have the same covariance matrix , which is set to be a unit matrix for the purpose of simplicity. Two cases are considered in this study. In case one, the number of , and eight variables are all set to be key features. The difference of mean values . In case two, , and the first eight variables are set to be the same with case one, but another eight irrelevant predictors are added. The number of total instances is set to 960. The negative to the positive ratios here are set be 1:1, 3:1, 9:1, 31:1, and 95:1, respectively. There are 28 Kendall’s rank correlation coefficients among 8 filtering methods, and the mean of these coefficients (with 100 repeats) is shown in Figure 1. As stated above, if all pairs are concordant. Whereas the maximum of is 0.88 in case one (left, Figure 1) where there are no irrelevant predictors, and 0.76 in case two (right, Figure 1) where one-half of features are irrelevant variables. The two maximal values are reached when two classes are exactly balanced, and reduces as the imbalance ratio increases in two cases. It indicates that these filtering methods probably generate different feature rankings, and such differences tend to be intensified when the class imbalanced ratio increases. Consequently, it is hard to say that a filtering approach is better or worse than another one, and it is a big risk to just depend on a single filter algorithm to make decisions. We have known that such a difference will occur due to the different principles of the filtering methods, but we also presume that class imbalance intensifies this difference. A natural way to combat this challenge may combine each filtering approach’s information and relieve the effect of class imbalance. This is the motivation for why we propose the strategy of rank aggregation with re-balance.
Figure 1

Kendall’s rank correlation coefficient under different imbalance ratios (with 100 repeats). Left: eight key variables; right: eight key plus eight irrelevant variables.

2.2. Rank Aggregation(RA) on Original Balanced Data

In our computation, eight filtering methods—t test, Fisher score, Hellinger distance, Relief, ReliefF, Information gain, Gini index, and R-value—are aggregated to generate an incorporative rank. Rank aggregation is firstly tested with the original balanced dataset “NPC”. Artificial rebalancing is unnecessary, and just case 1 (no resampling) is performed. Rank aggregation is compared with eight filtering methods: t test, Fisher score, Hellinger distance, Relief, ReliefF, Information gain, Gini index, and R-value. , , and are utilized as evaluation measurements. The rank lists ordered by their importance are shown on the x-axis in Figure 2. The top seven features are selected according to all four assessment metrics.
Figure 2

Rank lists of rank aggregation with the dataset “NPC” in case 1.

2.3. Rank Aggregation with Re-Balance (RAR) on Imbalanced Data

Figure 3, Figure 4, Figure 5 and Figure 6 show the aggregated rank lists on seven cases with the datasets “TBI”, “CHD2-1”, “CHD2-2”, and “ATR”, respectively. Rank aggregation combines each ranking into a list reflective of the overall preference, and each subgraph of four figures shows the aggregation results based on the CE algorithm. The x-axis is the optimal list obtained by the rank aggregation algorithm. The y-axis also ranks, and the gray line is the rank of the original data; the black line is their average rank; and the red line is the aggregate result of the CE algorithm. The order of the x-axis rank is based on the aggregate ranks obtained by the red line. The performances measured by , , , and are given in Table 1, Table 2, Table 3 and Table 4, respectively.
Figure 3

Rank lists of rank aggregation with the dataset “TBI” on seven cases.

Figure 4

Rank lists of rank aggregation with the dataset “CHD2-1” on seven cases.

Figure 5

Rank lists of rank aggregation with the dataset “CHD2-2” on seven cases.

Figure 6

Rank lists of rank aggregation with the dataset “ATR” on seven cases.

Table 1

from rank aggregation and eight filtering techniques (The best result is in bold).

DatasetResamplingNo.RARARt TestFisherHellingerReliefReliefFIGGiniR-Value
NPCCase 17 1.00 0.870.970.950.950.950.95 1.00 0.92
TBICase 16 0.96 0.410.680.680.700.880.720.580.63
Case 211 1.00 0.840.880.950.900.940.860.900.90
Case 31 1.00 0.950.840.950.900.800.950.90 1.00
Case 410 1.00 0.96 1.00 0.890.930.970.930.85 1.00
Case 57 1.00 0.930.900.970.970.930.970.860.93
Case 612 1.00 0.750.830.820.830.910.850.71 1.00
Case 729 1.00 0.710.700.650.820.850.780.710.71
CHD2-1Case 110 0.87 0.670.71 0.87 0.000.770.770.47 0.87
Case 2110.940.850.85 1.00 0.870.930.940.850.91
Case 36 1.00 0.86 1.00 1.00 1.00 0.93 1.00 0.93 1.00
Case 431 1.00 1.00 1.00 1.00 1.00 1.00 1.00 0.95 1.00
Case 537 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00
Case 6110.870.610.750.870.87 0.89 0.870.500.87
Case 710 0.87 0.610.71 0.87 0.87 0.87 0.87 0.710.71
CHD2-2Case 130.580.480.580.480.580.550.680.00 0.73
Case 234 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00
Case 314 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00
Case 424 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00
Case 528 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00
Case 67 1.00 1.00 0.820.82 1.00 0.820.820.470.82
Case 740.860.82 1.00 0.82 1.00 0.61 1.00 0.000.75
ATRCase 125 1.00 1.00 1.00 1.00 1.00 1.00 1.00 0.50 1.00
Case 29 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00
Case 38 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00
Case 42 1.00 1.00 1.00 1.00 1.00 1.00 1.00 0.89 1.00
Case 52 1.00 1.00 1.00 1.00 1.00 1.00 1.00 0.80 1.00
Case 69 1.00 1.00 1.00 1.00 1.00 1.00 1.00 0.71 1.00
Case 710 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00
Table 2

from rank aggregation and eight filtering techniques (The best result is in bold).

DatasetResamplingNO.RARARt TestFisherHellingerReliefReliefFIGGiniR-Value
NPCCase 18 1.00 0.880.950.950.97 1.00 0.950.950.92
TBICase 111 0.93 0.820.860.870.900.900.880.830.88
Case 213 1.00 0.840.900.940.950.940.840.870.86
Case 31 1.00 0.870.87 1.00 1.00 0.960.950.83 1.00
Case 47 1.00 0.910.970.930.970.900.930.89 1.00
Case 510 1.00 0.870.930.930.930.930.970.880.93
Case 619 1.00 0.730.800.860.860.770.800.710.91
Case 711 1.00 0.830.920.800.860.860.750.500.86
CHD2-1Case 110 0.88 0.87 0.88 0.870.870.870.870.830.87
Case 211 0.94 0.94 0.710.930.890.930.880.670.89
Case 36 1.00 0.89 1.00 1.00 1.00 0.930.940.940.92
Case 431 1.00 0.950.950.960.900.960.900.860.95
Case 537 1.00 1.00 0.910.950.95 1.00 0.950.860.95
Case 6110.890.500.750.730.89 1.00 0.830.550.67
Case 710 0.86 0.730.570.800.75 0.86 0.670.550.57
CHD2-2Case 130.910.870.86 0.95 0.870.910.900.870.91
Case 234 1.00 1.00 1.00 0.880.920.93 1.00 0.880.93
Case 314 1.00 1.00 0.920.860.92 1.00 0.930.860.93
Case 424 1.00 0.900.910.950.950.95 1.00 0.950.95
Case 528 1.00 0.950.950.950.95 1.00 0.960.95 1.00
Case 67 0.86 0.570.800.800.75 0.86 0.86 0.50 0.86
Case 74 0.86 0.800.750.57 0.86 0.86 0.86 0.500.67
ATRCase 125 1.00 1.00 0.89 1.00 1.00 1.00 1.00 0.89 1.00
Case 29 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00
Case 38 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00
Case 42 1.00 1.00 1.00 1.00 1.00 1.00 1.00 0.89 1.00
Case 52 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00
Case 69 1.00 1.00 1.00 1.00 1.00 1.00 1.00 0.67 1.00
Case 710 1.00 1.00 1.00 1.00 1.00 1.00 1.00 0.67 1.00
Table 3

from rank aggregation and eight filtering techniques (The best result is in bold).

DatasetResamplingNO.RARARt TestFisherHellingerReliefReliefFIGGiniR-Value
NPCCase 116 0.96 0.900.910.940.930.960.930.930.95
TBICase 13 0.70 0.480.610.520.610.650.580.490.67
Case 211 0.89 0.800.810.780.820.870.790.720.83
Case 31 0.95 0.830.740.940.850.830.850.76 0.95
Case 428 0.91 0.850.860.870.880.860.860.840.89
Case 526 0.93 0.910.920.900.900.920.900.880.90
Case 622 0.77 0.650.680.690.730.740.730.500.71
Case 725 0.71 0.630.560.680.530.610.610.350.66
CHD2-1Case 1100.600.480.51 0.61 0.500.590.600.450.59
Case 2110.760.660.600.75 0.77 0.730.720.610.73
Case 36 0.92 0.850.81 0.92 0.900.810.880.800.79
Case 4310.860.810.790.86 0.87 0.820.840.750.84
Case 5370.870.850.840.860.860.810.86 0.90 0.82
Case 611 0.64 0.570.450.500.570.57 0.64 0.360.62
Case 710 0.60 0.520.550.520.480.52 0.60 0.360.52
CHD2-2Case 13 0.60 0.520.490.540.500.520.550.390.57
Case 2340.860.790.84 0.88 0.730.800.800.700.79
Case 314 0.90 0.850.840.82 0.90 0.680.880.810.84
Case 424 0.93 0.880.880.880.900.860.910.87 0.93
Case 528 0.91 0.880.900.880.85 0.91 0.890.890.90
Case 670.630.530.56 0.66 0.590.440.560.280.59
Case 74 0.66 0.560.630.50 0.66 0.66 0.630.220.56
ATRCase 1250.88 0.98 0.510.850.850.760.850.500.76
Case 29 0.96 0.790.86 0.96 0.96 0.93 0.96 0.750.93
Case 380.97 1.00 0.900.970.900.970.970.800.93
Case 42 0.98 0.930.830.88 0.98 0.95 0.98 0.810.95
Case 52 0.98 0.810.860.760.930.930.950.710.88
Case 69 0.94 0.810.750.88 0.94 0.880.810.190.81
Case 710 0.94 0.810.630.690.63 0.94 0.880.190.75
Table 4

from rank aggregation and eight filtering techniques (The best result is in bold).

DatasetResamplingNO.RARARt TestFisherHellingerReliefReliefFIGGiniR-Value
NPCCase 115 0.96 0.870.900.910.920.960.920.910.91
TBICase 18 0.62 0.270.580.400.470.560.470.260.46
Case 2180.850.790.810.78 0.86 0.840.810.650.75
Case 31 0.93 0.770.600.910.850.740.850.740.89
Case 427 0.89 0.840.810.830.850.860.830.810.85
Case 513 0.91 0.810.820.810.820.900.860.790.81
Case 630 0.78 0.610.690.640.620.660.670.500.70
Case 724 0.71 0.650.540.640.590.570.620.410.58
CHD2-1Case 110 0.61 0.330.300.520.280.450.430.230.45
Case 211 0.77 0.560.580.690.650.640.680.550.65
Case 36 0.88 0.79 0.88 0.810.770.810.860.790.86
Case 431 0.86 0.810.800.800.820.830.850.730.80
Case 537 0.87 0.810.750.840.850.820.840.780.82
Case 611 0.66 0.460.450.480.650.530.550.400.50
Case 710 0.63 0.440.530.520.550.560.560.400.48
CHD2-2Case 13 0.45 0.220.330.270.320.310.270.210.26
Case 234 0.87 0.680.780.760.820.850.860.730.80
Case 314 0.87 0.810.740.810.860.760.830.790.78
Case 424 0.91 0.840.900.870.850.850.870.850.81
Case 528 0.90 0.860.880.800.880.860.890.790.79
Case 67 0.71 0.420.620.640.660.450.570.350.63
Case 740.640.640.570.500.630.55 0.66 0.400.60
ATRCase 125 0.93 0.750.520.520.820.820.820.250.60
Case 29 1.00 0.810.880.920.940.940.940.680.88
Case 38 1.00 1.00 0.78 1.00 0.880.93 1.00 0.820.88
Case 42 0.95 0.840.790.910.910.880.910.700.91
Case 52 0.95 0.790.780.780.880.90 0.95 0.610.88
Case 69 1.00 0.890.690.850.890.760.850.360.80
Case 710 1.00 0.610.540.640.540.800.890.370.80

3. Discussion

Table 1 and Table 2 show that RA reached the maximal values of and . It can be seen from Table 3 and Table 4 that RA and ReliefF obtained the maximal values of and . Therefore, RA outperformed single filtering methods when assessed with , , , and . The NPC dataset had a completely balanced distribution, and RA worked well on it. Thus, rank aggregation is necessary to integrate different results, even if in a totally balanced situation, and a consensual feature ranking list is provided. Aggregation ranking lists in Figure 3, Figure 4, Figure 5 and Figure 6 tell us the order of importance of each feature. Though the rank lists derived from different subsampling methods were not the same, the top features were approximately consistent. After obtaining the rank list, another task is to figure out how many features should be considered as key variables. In this computation, we performed 5-fold cross-validation [32] to find the optimal number of key features. As recent studies have showed that is more informative in imbalanced learning [32,33], was employed as perfomance metric in this section, and random forest classifier was utilized to implement classification. Namely, the value of was calculated, as the top k ranked features were used each time, where k varies from 1 to p (see Figure 7, Figure 8, Figure 9 and Figure 10). We chose the optimal k- value such that the random forest classifier had the maximal in identifying classification. It can be seen from Table 1, Table 2, Table 3 and Table 4 that the optimal number of important features varied greatly under different re-balanced strategies. One possible reason is that the artificial data generated by different subsampling have difference to some extent. Another possible reason is the measurement changes sightly as the number of candidate features changes. This seems to be true from the Figure 7, Figure 8, Figure 9 and Figure 10 where each curve tends to be flat as the changes in the number of features used for classification. It also noted that the under no re-sampling (case 1) was generally lower than that under six re-sampling methods.
Figure 7

when top k features are used with the dataset “TBI” on seven cases.

Figure 8

when top k features are used with the dataset “CHD2-1” on seven cases.

Figure 9

when top k features are used with the dataset “CHD2-2” on seven cases.

Figure 10

when top k features are used with the dataset “ATR” on seven cases.

Table 1, Table 2, Table 3 and Table 4 report the results of these real datasets with the assessing metrics , , , and , respectively. We can perform comparisons in several aspects. Original imbalanced datasets are employed in case 1 from Table 1, Table 2, Table 3 and Table 4 (except NPC dataset). Of all the 16 “no re-balance” situations, the aggregation rank method reached the maximal measures in 12 situations compared with the other 8 filtering methods (t test, Fisher score, Hellinger distance, Relief, ReliefF, information gain, Gini, and R-value). It indicates that aggregation rank was better than a single filtering rank with the proportion of 75.00% when the data are class-imbalanced. If the original dataset NPC was counted in, this proportion was 80.00%. Therefore, rank aggregation is generally superior to single filtering methods, no matter how the data are balanced or imbalanced. Re-balanced datasets were artificially generated and utilized in cases 2–7 from Table 1, Table 2, Table 3 and Table 4. Of all the 96 scenarios with re-balance, the aggregation rank method reached the maximal measures in 83 scenarios compared to the other eight filtering methods. It means that aggregation rank outperformed single filtering rank with the proportion of 86.46% when the class-imbalanced data were treated with re-balance strategies. Thus, performing aggregation rank is extremely effective in dealing with class-imbalanced data. Rank aggregation was performed on both imbalanced datasets (RA) and re-balanced datasets (RAR). Of all the 96 scenarios with re-balance (cases 2–7 in Table 1, Table 2, Table 3 and Table 4), there were 93 situations whose measurements were equal or greater than those from case 1 (no re-balance). It shows that aggregation rank with re-balance strategies performed better with the proportion of 96.88% than that with original class-imbalanced data. Therefore, performing re-balance can play a crucial role in improving the performance of rank aggregation when the data are class-imbalanced. Figure 7, Figure 8, Figure 9 and Figure 10 show the curves of seven cases on four imbalanced datasets. from re-balanced data (cases 2–7) was generally higher than that from imbalanced data (case 1). In other words, the performance can be promoted after re-sampling to balance the imbalanced data artificially. Case 5 and case 6 are two under-sampling methods, and the was generally lower than that from over-sampling or hybrid sampling (cases 2–4). The possible reason is that some of the useful information is missed in doing under-sampling when the size of the minority instances is too small (see Table 5). Therefore, one should be cautious about using under-sampling in practice.
Table 5

The summary of five datasets.

DatasetsAttributesInstancesMajorityMinorityRatio
NPC242001001001.00
TBI4210473312.35
CHD2-1507251212.43
CHD2-2506751163.19
ATR104292182.63
In sum, different filter methods generate different rankings. Rank aggregation is necessary to integrate different results and provide a consensual feature ranking list. Class-imbalance usually leads to degraded performance from a filtering method on feature importance ranking. This harmfulness can be alleviated via different re-balance strategies in sample space.

4. Materials and Methods

4.1. Notations

The notations used in this study are listed below:

4.2. Eight Filtering Methods

4.2.1. t Test

Feature screening using the t test statistic [34] is similar to performing a hypothesis test (the null hypothesis is that there is no difference in the means) on the class’s distribution, and its significance indicates the difference between majority and minority classes. The lower the p value of this t test, the higher the majority and minority classes’ significant difference. Consequently, the considered feature is more relevant to the separation of two classes.

4.2.2. Fisher Score

Fisher score [35] is simple and generally quite effective, which can be a criterion of feature screening. Fisher score of a single feature is defined as follows: , , , and can be replaced by their corresponding sample statistics in computation, namely, A feature with a large Fisher score is more crucial for discriminating the two categories.

4.2.3. Hellinger Distance

Hellinger distance can be used to measure a distributional divergence [36]. Denoted the two normal distributions by P and Q, Hellinger distance is calculated as follows: where , , , and are the expectation and variance of P and Q, respectively, and their corresponding sample statistics are used in practice [37]. The larger the Hellinger distance is, the more divergent the two distributions are.

4.2.4. Relief and ReliefF

The Relief is an iteration method that tries to give each feature a score to indicate its level of relevance to the response [37,38]. Let be an instance; and be its two nearest neighbors from the same class and the other class by the Euclidean distance, respectively. The score vector is refreshed as follows: where , and are the jth element of , and , respectively. A feature with a higher score is more crucial to the response. Though ReliefF [39] is originally developed for dealing with multi-class and noise datasets, it can be applied to binary classification cases. Compared with Relief that searches one nearest instance from the same class and one from the other class in updating the weights, ReliefF finds k nearest neighbors. Similarly, a feature with a higher score is more important to the response.

4.2.5. Information Gain (IG)

Information gain [40] is the measurement of informational theory and can be utilized to assess the importance of a given feature. In the settings of binary classification, the information entropy of the set D is defined as follows: Assuming that a discrete feature (attribute) has V different values , and is the subset of instance set D satisfying . The information gain of the variable is The larger the information gain is, the more important the feature is for separating the classes. A continuous feature should be discretized before using the IG metric.

4.2.6. Gini Index

Gini index [41] fits binary digits, continuous numerical values, ordinal numbers, etc. It is a non-purity split method. Gini index of D is defined as follows: where is the probability that any instance belongs to , and it is replaced with , in practice. If we divide D into M subsets , the Gini index after splitting is: The smaller the Gini index is, the more important the feature is.

4.2.7. R-Value

R-value [30,42] indicates the degree of overlap for the class-imbalanced dataset. R-value for a dataset D is defined as follows: where where is the subset of k nearest neighbors of instance P that belong to the set of instances , and is the threshold generally set to be [43]. The smaller the R-value is, the more important the feature is for discriminating the categories.

4.3. Four Evaluation Metrics

4.3.1. Geometric Mean and F-Measure

True positive, true negative, false positive, and false negative are denoted by , , , and , respectively. Some common metrics are listed below: The range of both and is . The larger they are, the better the classifer works.

4.3.2. and

is the area under the receiver operating characteristic curve (ROC) [44]. is the area under the precision recall curve (PRC) [45]. Both and range from 0 to 1, and the larger they are, the better the classifier is built for the imbalanced learning. More details on or can be found in our previous studies [37,46]. , , , and are more widely used than the metric in class-imbalance learning. These metrics actually pay more attention to the minority samples.

4.4. Kendall’S Rank Correlation

Kendall’s rank correlation statistic [47] can be applied to calculate the degree of comparability between the feature rankings of two filtering techniques. Let the two feature rankings generated by two filters be and there are no ties in each of ranking list. Then Kendall’s is calculated as follows, where is the sign function, namely it equals 1 if x is positive and if x negative. A pair of is called concordant if and or and . Otherwise, they are considered discordant. The numerator is the difference between the number of concordant pairs and the number of discordant pairs, and the denominator is the number of all distinct pairs of p elements. The range of is . If , the correlation of two rankings is weak; if , then all pairs will be discordant, and the two rankings are exactly opposite; if , then all pairs are exactly concordant [48].

4.5. Rank Aggregation with Re-Balance for Class-Imbalanced Data

As mentioned above, there are differences among the ranks from different filtering methods, but we assume that they are equal in match, namely, no one is better or worse than another. Rank aggregation (RA) is a greatly intuitive metric that computes the absolute differences between the ranks of all individual features [49]. Rank aggregation with re-balance (RAR) consists of two stages for class-imbalanced data and is illustrated in Figure 11. In sample space, the data are artificially balanced by generating new instances of the minority class or (and) removing some of the majority class instances. In feature space, m rank lists are first computed using m different filtering methods. Each rank list is the full permutation of all the features. Then, they are merged to be an aggregated rank. Feature screening and classification can be performed according to this aggregated rank.
Figure 11

The frame of rank aggregation with re-balance.

4.5.1. Rank Aggregation

As mentioned above, different filter techniques will give different feature ranking results. The rank aggregation method [34,50] combines all the rankings together, by aggregating all feature ranking lists generated from different filtering methods. RA is to find an optimal ranking such that where is the ith feature ranking list, represents a ranking list with the same length of , d is a distance function, and is the important weight related with list . In this study, d is chosen to be the Spearman’s foot rule distance [50]: denotes an ordered list of top m algorithms produced by the validation measure M. Let be the scores for the top m algorithms in , where is the best score given by measure M and so on. Let be the rank of A under M (1 means “best”) if A is within top m, and be equal to ; otherwise, is defined likewise. The optimization of the objective (12) is achieved by using the Monte Carlo cross-entropy (CE) algorithm [51,52]. CE Monte Carlo algorithm is a stochastic search method, which produces a “better” sample in the future, which is concentrated around an x that corresponds to an optimal [50].

4.5.2. Strategies to Generate New Samples

Before performing rank aggregation, the training instances are to be modified to produce a more balanced class distribution. To achieve this task, new minority or (and) majority class samples need to be generated or drawn from the original dataset. We employ the following three strategies to gain new samples:

Randomly Sampling

In the over-sampling, some (all) the minority class instances are randomly duplicated; in the under-sampling, a portion of majority samples are randomly removed.

SMOTE

Synthetic minority over-sampling technique (SMOTE) is a popular over-sampling algorithm [5]. Figure 12 illustrates how to generate new samples according to the selected point in SMOTE. The five selected nearest neighbors of are to . to are the synthetic data points created by the randomized interpolation. Namely, where is a random number between 0 and 1. The above operation can be repeated to obtain requested synthetic minority instances.
Figure 12

An illustration of how to create the synthetic data points in the SMOTE algorithm.

Smoothed Bootstrap

Smoothed bootstrap technique repeatedly bootstraps the data from the two classes and employs smoothed kernel functions to generate new approximately balanced samples [53]. A new instance is generated by performing the following three steps: 1: choose with probability ; 2: choose in the original daa set such that with probability ; 3: sample from a probability distribution , which is centered at and depends on the smoothing matrix . In brief, smoothed bootstrap firstly draws randomly from the original dataset an instance from one of the two categories, then generates a new instance in its neighborhood.

4.6. Experiment and Assessing Metrics

As shown in Table 5, five metabolomics datasets were employed to test our algorithm. NPC is a nasopharyngeal carcinoma dataset [32,54] that is exactly balanced. In this study, NPC was utilized to investigate the performance of rank aggregation strategy on original balanced data, which included 100 patients with nasopharyngeal carcinoma and 100 healthy controls. Traumatic brain injury (TBI) is from our previous studies [32,55], which reports the serum metabolic profiling of TBI patients with (or without) cognitive impairment (CI). The TBI dataset included 73 TBI patients with CI and 31 TBI patients without CI. CHD2-1 and CHD2-2 datasets are actually from the same experiment about coronary heart disease (CHD) [30]. The CHD2-1 dataset contains 21 patients with CHD, and the CHD2-2 dataset contains 16 patients with coronary heart disease associated with type 2 diabetes mellitus (CHD-T2DM), which are compared with a control group of 51 healthy adults. ATR is an Acori Tatarinowii Rhizoma dataset, which included 21 samples collected from Sichuan Province, and 8 samples were from Anhui Province in China [56]. Table 5 lists the summary of five datasets; included are the numbers of attributes, total instances, the majority, the minority instances, and the imbalance ratio. The NPC dataset was utilized to test the performance of rank aggregation under original balanced distribution. The other four imbalanced data sets were used to evaluate the RAR algorithm with artificially re-balanced data. This section shows the efficacy of the proposed RAR algorithm on one original balanced dataset and four class-imbalanced datasets and compares it with other filtering feature screening methods via several assessing metrics. Rank aggregation was performed under the following seven situations: : no re-sampling: the original datasets are directly utilized to perform rank aggregation. Denoted case 1 by “RA” because there is no re-sampling in it. : hybrid-sampling A: some instances of the majority class are randomly eliminated, and new synthetic minority examples are generated by SMOTE. The size of the remaining majority is equal to the size of the (original plus new generated) minority class. : hybrid-sampling B: A new synthetic dataset is generated according to the smoothed bootstrap re-sampling technique. The sizes of the majority and minority classes are approximately equal. : over-sampling A: new minority class instances are randomly duplicated according to the original minority group. : over-sampling B: new synthetic minority examples are generated on the basis of the smoothed bootstrap re-sampling technique. : under-sampling A: some instances from the majority class are randomly removed so that the size of the remaining majority class is equal to the size of the minority. : under-sampling B: new synthetic majority examples are generated according to the smoothed bootstrap re-sampling technique. Note that NPC is balanced, and just case 1 is performed on it. Table 6 lists the summary of the six re-balanced strategies. In this study, , , , and are employed to assess the performance of RA or RAR algorithm on five datasets under seven cases.
Table 6

Re-balanced strategies.

MethodsRe-Sampling ProcessAlgorithm Process
Under-Sampling Over-Sampling Hybrid SMOTE Random Smoothed Bootstrap
Case 2 YesYes
Case 3 Yes Yes
Case 4 Yes Yes
Case 5 Yes Yes
Case 6Yes Yes
Case 7Yes Yes

5. Conclusions

In this paper, we propose a simple but effective strategy called RAR for feature screening of class-imbalanced data by aggregating rankings from individual filtering algorithms and modifying the class-imbalanced data with various re-sampling methods to provide balanced or more adequate data. RAR can address the problem of inconsistency between different feature ranking methods to a large extent. The results on real datasets show that RAR is highly competitive and almost better than single filtering screening in terms of geometric mean, F-measure, , and . After performing re-balanced pretreatment, the performance of rank aggregation can be highly improved, so re-sampling to balance the classes is extremely useful in rank aggregation when the data are class-imbalanced in metabolomics. Our proposed method serves as a reference for future research on feature selection for the diagnosis of diseases. Rank aggregation is a general idea to investigate the importance of features. In this study, rankings from eight filtering algorithms are employed to generate the aggregated rank. There are many other filter techniques, such as Chi-squared, power, Kolmogorov–Smirnov statistic, and signal-to-noise ratio [57], which are all widely utilized in class-imbalance learning. In addition, considering that a re-sampling method can also generate a rank list, rank aggregation can be performed according to the various re-sampling algorithms rather than different filtering methods. Further, if necessary, ensemble multiple rank aggregations could be performed to combine those aggregated rankings derived from different algorithms. Finally, although RAR is used in the metabolomics datasets in this study, it is potentially available for handing high-dimensional imbalanced data from other fields, such as economics and biology.
  12 in total

1.  A support vector machine-recursive feature elimination feature selection method based on artificial contrast variables and mutual information.

Authors:  Xiaohui Lin; Fufang Yang; Lina Zhou; Peiyuan Yin; Hongwei Kong; Wenbin Xing; Xin Lu; Lewen Jia; Quancai Wang; Guowang Xu
Journal:  J Chromatogr B Analyt Technol Biomed Life Sci       Date:  2012-05-24       Impact factor: 3.205

2.  Weighted rank aggregation of cluster validation measures: a Monte Carlo cross-entropy approach.

Authors:  Vasyl Pihur; Susmita Datta; Somnath Datta
Journal:  Bioinformatics       Date:  2007-05-05       Impact factor: 6.937

Review 3.  A review of feature selection techniques in bioinformatics.

Authors:  Yvan Saeys; Iñaki Inza; Pedro Larrañaga
Journal:  Bioinformatics       Date:  2007-08-24       Impact factor: 6.937

4.  Tuning model parameters in class-imbalanced learning with precision-recall curve.

Authors:  Guang-Hui Fu; Lun-Zhao Yi; Jianxin Pan
Journal:  Biom J       Date:  2018-12-12       Impact factor: 2.207

5.  RFS: efficient feature selection method based on R-value.

Authors:  Jimin Lee; Nomin Batnyam; Sejong Oh
Journal:  Comput Biol Med       Date:  2012-12-20       Impact factor: 4.589

6.  Variable importance analysis based on rank aggregation with applications in metabolomics for biomarker discovery.

Authors:  Yong-Huan Yun; Bai-Chuan Deng; Dong-Sheng Cao; Wei-Ting Wang; Yi-Zeng Liang
Journal:  Anal Chim Acta       Date:  2016-01-07       Impact factor: 6.558

7.  Finding common genes in multiple cancer types through meta-analysis of microarray experiments: a rank aggregation approach.

Authors:  V Pihur; Somnath Datta; Susmita Datta
Journal:  Genomics       Date:  2008-06-20       Impact factor: 5.736

8.  uEFS: An efficient and comprehensive ensemble-based feature selection methodology to select informative features.

Authors:  Maqbool Ali; Syed Imran Ali; Dohyeong Kim; Taeho Hur; Jaehun Bang; Sungyoung Lee; Byeong Ho Kang; Maqbool Hussain
Journal:  PLoS One       Date:  2018-08-28       Impact factor: 3.240

9.  RankAggreg, an R package for weighted rank aggregation.

Authors:  Vasyl Pihur; Susmita Datta; Somnath Datta
Journal:  BMC Bioinformatics       Date:  2009-02-19       Impact factor: 3.169

10.  Hellinger distance-based stable sparse feature selection for high-dimensional class-imbalanced data.

Authors:  Guang-Hui Fu; Yuan-Jiao Wu; Min-Jie Zong; Jianxin Pan
Journal:  BMC Bioinformatics       Date:  2020-03-23       Impact factor: 3.169

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.