Literature DB >> 28659973

Forest Pruning Based on Branch Importance.

Xiangkui Jiang1, Chang-An Wu2, Huaping Guo2.   

Abstract

A forest is an ensemble with decision trees as members. This paper proposes a novel strategy to pruning forest to enhance ensemble generalization ability and reduce ensemble size. Unlike conventional ensemble pruning approaches, the proposed method tries to evaluate the importance of branches of trees with respect to the whole ensemble using a novel proposed metric called importance gain. The importance of a branch is designed by considering ensemble accuracy and the diversity of ensemble members, and thus the metric reasonably evaluates how much improvement of the ensemble accuracy can be achieved when a branch is pruned. Our experiments show that the proposed method can significantly reduce ensemble size and improve ensemble accuracy, no matter whether ensembles are constructed by a certain algorithm such as bagging or obtained by an ensemble selection algorithm, no matter whether each decision tree is pruned or unpruned.

Entities:  

Mesh:

Year:  2017        PMID: 28659973      PMCID: PMC5474283          DOI: 10.1155/2017/3162571

Source DB:  PubMed          Journal:  Comput Intell Neurosci


1. Introduction

Ensemble learning is a very important research topic in machine learning and data mining. The basic heuristic is to create a set of learners and aggregate the prediction of each learner for classifying examples. Many approaches such as bagging [1], boosting [2], and COPEN [3] have been proposed to create ensembles, and the key to the success of these approaches is that base learners are accurate and diverse [4]. Ensemble methods have been applied to many applications such as image detection [5-7] and imbalanced learning problem [8]. However, an important drawback existing in ensemble learning approaches is that they try to train unnecessarily large ensembles. Large ensembles need a large memory for storing the bases learners and much response time for prediction. Besides, large ensemble may reduce its generalization ability instead of increasing the performance [9]. Therefore, a lot of researches to tackle this problem have been carried out, and the researches mainly focus on ensemble selection: selecting a subset of ensemble members for prediction, such as ordered-based ensemble selection methods [10-12] and greedy heuristic based ensemble selection methods [13-21]. The research results indicate that a well-designed ensemble selection method can reduce ensemble size and improve ensemble accuracy. Besides ensemble selection, we can prune an ensemble through the following two approaches if ensemble members are decision trees: (1) pruning individual members separately and combining the pruned members together for prediction and (2) repeatedly pruning individual members by considering the overall performance of the ensemble. For the first strategy, many decision tree pruning methods such as those used in CART [22] and C4.5 [23] have been studied. Although pruning can simplify model structure, whether pruning can improve model accuracy is still a controversial topic in machine learning [24]. The second strategy coincides with the expectation of improving model generalization ability globally. However, this method has not been extensively studied. This paper focuses on this strategy and names the strategy as forest pruning (FP). The major job of forest pruning is to define an effective metric evaluating the importance of a certain branch of trees. Traditional metrics can not be applied to forest pruning, since these metrics just consider the influence on a single decision tree when a branch is pruned. Therefore, we need a new metric for pruning forest. Our contributions in this paper are as follows: Introduce a new ensemble pruning strategy to prune decision tree based ensemble; propose a novel metric to measure the improvement of forest performance when a certain node grows into a subtree; present a new ensemble pruning algorithm with the proposed metric to prune a decision tree based ensemble. The ensemble can be learned by a certain algorithm or obtained by some ensemble selection method. Each decision tree can be pruned or unpruned. Experimental results show that the proposed method can significantly reduce the ensemble size and improve its accuracy. This result indicates that the metric proposed in this paper reasonably measures the influence on ensemble accuracy when a certain node grows into a subtree. The rest of this paper is structured as follows. Section 2 provides a survey of ensemble of decision trees; Section 3 presents the formal description of forest trimming and the motivation of this study by an example. Section 4 introduces a new forest pruning algorithm. Section 5 reports and analyzes experimental results and we conclude the paper with simple remark and future work in Section 6.

2. Forests

A forest is an ensemble whose members are learned by decision tree learning method. Two approaches are often used to train a forest: traditional approaches and the methods specially designed for forests. Bagging [1] and boosting [2] are the two most often used traditional methods to build forests. Bagging takes bootstrap samples of objects and trains a tree on each sample. The classifier votes are combined by majority voting. In some implementations, classifiers produce estimates of the posterior probabilities for the classes. These probabilities are averaged across the classifiers and the most probable class is assigned, called “average” or “mean” aggregation of the outputs. Bagging with average aggregation is implemented in Weka and used in the experiments in this paper. Since each individual classifier is trained on a bootstrap sample, the data distribution seen during training is similar to the original distribution. Thus, the individual classifiers in a bagging ensemble have relatively high classification accuracy. The factor encouraging diversity between these classifiers is the proportion of different examples in the training set. Boosting is a family of methods and Adaboost is the most prominent member. The idea is to boost the performance of a “weak” classifier (can be decision tree) by using it within an ensemble structure. The classifiers in the ensemble are added one at a time so that each subsequent classifier is trained on data which have been “hard” for the previous ensemble members. A set of weights is maintained across the objects in the data set so that objects that have been difficult to classify acquire more weight, forcing subsequent classifiers to focus on them. Random forest [25] and rotation forest [26] are two important approaches specially designed for building forests. Random forest is a variant version of bagging. The forest is built again on bootstrap samples. The difference lies in the construction of the decision tree. The feature to split a node is selected as the best feature among a set of M randomly chosen features, where M is a parameter of the algorithm. This small alteration appeared to be a winning heuristic in that diversity was introduced without much compromising the accuracy of the individual classifiers. Rotation forest randomly splits the feature set into K subsets (K is a parameter of the algorithm) and Principal Component Analysis (PCA) [27] is applied to each subset. All principal components are retained in order to preserve the variability information in the data. Thus, K axis rotations take place to form the new features and rotation forest building a tree using all training set in the new space defined by a given new feature space.

3. Problem Description and Motivation

3.1. Problem Description

Let D = {(x, y)∣i = 1,2,…, N} be a data set, and let F = {T1,…, T} be an ensemble with decision tree, T, learning from D. Denote by v ∈ T a node in tree T and by E(v) ∈ D, the set of the examples reaching v from the root of T, root(T). Suppose each node v ∈ T contains a vector (p1, p2,…, p), where p is the proportion of the examples in E(v) associated with label k. If v ∈ T is a leaf and x ∈ E(v), the prediction of T on x is Similarly, for each example x to be classified, ensemble F returns a vector (p, p,…, p) indicating that x belongs to label k with probability p, where The prediction of F on x is F(x) =  argmax⁡p. Now, our problem is, given a forest F with M decision trees, how to prune each tree to reduce F's size and improve its accuracy, where F is either constructed by some algorithm or obtained by some ensemble selection method.

3.2. Motivation

First, let us look at an example, which shows the possibility that forest trimming can improve ensemble accuracy.

Example 1 .

Let F = {T0, T1,…, T9} be a forest with ten decision trees, where T1 is shown in Figure 1. Suppose that p1 = 0.60, p2 = 0.40; p1 = 1.00, p2 = 0.00; p1 = 0.20 and p2 = 0.80. Let ten examples x0, x1,…, x9 reach node v, where x0,…, x5 associate with label 1 and x6,…, x9 associate with label 2. Assume examples x0, x1,…, x4 reach leaf node v1, and x5,…, x9 reach leaf node v1.
Figure 1

Decision tree T0. v is a test node and v1 and v2 are two leaves.

Obviously, for T0, we can not prune the children of node v, since treating v as a leaf would lead to more examples incorrectly classified by T0. Assume that F's predictions on x0, x1,…, x9 are as follows:where p is the probability of x associated with label k. From F's predictions shown above, we have that x6 is incorrectly classified by F. Update T0 to T0′ by pruning v's children and update F to F′ = {T0′, T1,…, T9}. A simple calculation tells us that, for the ten examples, F′ returns: It is easy to see that F′ correctly classifies all of the ten examples. This example shows that if a single decision tree is considered, maybe it should not be pruned any more. However, for the forest as a whole, it is still possible to prune some branches of the decision tree, and this pruning will probably improve the ensemble accuracy instead of reducing it. Although the example above is constructed by us, similar cases can be seen everywhere when we study ensembles further. It is this observation that motivates us to study forest trimming methods. However, more efforts are needed to turn possibility into feasibility. Further discussions about this problem will be presented in the next section.

4. Forest Pruning Based on Branch Importance

4.1. The Proposed Metric and Algorithm Idea

To avoid trapping in detail too early, we assume that I(v, F, x) has been defined, which is the importance of node v when forest F classifies example x. If x ∉ E(v), then I(v, F, x) = 0. Otherwise, the details of the definition of I(v, F, x) are presented in Section 3.2. Let T ∈ F be a tree and let v ∈ T be a node. The importance of v with respect to forest F is defined aswhere D′ is a pruning set and E(v) is the set of the example in D′ reaching node v from root(T). I(v, F) reflects the impact of node v on F's accuracy. Let L(v) be the set of leaf nodes of branch(v), the branch (subtree) with v as the root. The contribution of branch(v) to F is defined aswhich is the sum of the importance of leaves in branch(v). Let v ∈ T be a nonterminal node. The importance gain of v to F is defined by the importance difference between branch(v) and node v, that is,IG(v, F) can be considered as the importance gain of branch(v), and its value reflects how much improvement of the ensemble accuracy is achieved when v grows into a subtree. If IG(v, F) > 0, then this expansion is helpful to improve F's accuracy. Otherwise it is unhelpful to improve or even reduce F's accuracy. The idea of the proposed method of pruning ensemble of decision trees is as follows. For each nonterminal node v in each tree T, calculate its importance gain IG(v, F) on the pruning set. If IG(v, F) is smaller than a threshold, prune branch(v) and treat v as a leaf. This procedure continues until all decision trees can not be pruned. Before presenting the specific details of the proposed algorithm, we introduce how to calculate I(v, F, x) in the next subsection.

4.2. Con(v, F, x) Calculation

Let h be a classifier and let S be an ensemble. Partalas et al. [28, 29] identified that the prediction of h and S on an example x can be categorized into four cases: (1)  e: h(x) = y∧S(x) ≠ y, (2)  e: h(x) = y∧S(x) = y, (3)  e: h(x) ≠ y∧S(x) = y, (4)  e: h(x) ≠ y∧S(x) ≠ y. They concluded that considering all four cases is crucial to design ensemble diversity metrics. Based on the four cases above, Lu et al. [11] introduced a metric, IC(, to evaluate the contribution of the ith classifier to S when S classifies the jth instance. Partalas et al. [28, 29] introduced a measure called Uncertainty Weighted Accuracy, UWA(h, S, x), to evaluate h's contribution when S classifies example x. Similar to the discussion above, we defineIn the following discussions, we assume that v ∈ T and x ∈ E(v). Let f and f be the subscripts of the largest element and the second largest element in {p,…, p}, respectively. Obviously, f is the label of x predicted by ensemble F. Similarly, let t = arg⁡max(p1,…, p). If v is a leaf node, then t is the label of x predicted by decision tree T. Otherwise, t is the label of x predicted by T′, where T′ is the decision tree obtained from T by pruning branch(v). For simplicity, we call t the label of x predicted by node v and say node v correctly classifies x if t = y. We define I(v, F, x) based on the four cases in formula (8), respectively. If x ∈ e(v) or x ∈ e(v), then Con(v, F, x) ≥ 0, since v correctly classifies x. Otherwise, Con(v, F, x) < 0, since v incorrectly classifies x. For x ∈ e(v), Con(v, F, x) is defined aswhere M is the number of base classifiers in F. Here, t = y and f ≠ y, then p ≥ p, p ≥ p, and thus 0 ≤ Con(v, F, x) ≤ 1. Since p is the contribution of node v to the probability that F correctly predicates x belonging to class t while p is the contribution of node v to p, the probability that F incorrectly predicates x belongs to class f, (p − p)/M can be considered as the net importance of node v when F classifies x. p − p is the weight of v's net contribution, which reflects the importance of node v for classifying x correctly. The constant 1/M is to avoid p − p being zero or too small. For x ∈ e(v), Con(v, F, x) is defined asHere, 0 ≤ Con(v, F, x) ≤ 1. In this case, both v and F correctly classify x. We treat (p − p)/M as the net contribution of node v to F and x, and p − p as the weight of v's net contribution. For x ∈ e(v), Con(v, F, x) is defined asIt is easy to prove −1 ≤ Con(v, F, x) ≤ 0. This case is opposed to the first case. In this case, we treat −(p − p)/M as the net contribution of node v to F and x, and p − p as the weight of v's net contribution. For x ∈ e(v), Con(v, F, x) is defined aswhere y ∈ {1,…, K} is the label of x, −1 ≤ Con(v, F, x) ≤ 0. In this case, both v and F incorrectly classify x, namely, t ≠ y and f ≠ y. We treat −(p − p)/M as the net contribution of node v to F and x, and p − p as the weight of v's net contribution.

4.3. Algorithm

The specific details of forest pruning (FP) are shown in Algorithm 1, where
Algorithm 1

The procedure of forest pruning.

D′ is a pruning set containing n instances, p is the probability that ensemble F predicts x ∈ D′ associated with label k, p is the probability that current tree T predicts x ∈ D′ associated with label k, I is a variant associated with node v to save v's importance, Ibr( is a variant associated with node v to save the contribution of branch(v). FP first calculates the probability of F's prediction on each instance x (lines (1)~(2)). Then it iteratively deals with each decision tree T (lines (3)~(14)). Lines (4)~(10) calculate the importance of each node v ∈ T, where I(v, F, x) in line (10) is calculated using one of the equations (9)~(12) based on the four cases in equation (8). Line (11) calls PruningTree(v) to recursively prune T. Since forest F has been changed after pruning T, we adjust F's prediction in lines (12)–(14). Lines (3)–(14) can be repeated many times until all decision trees can not be pruned. Experimental results show that forest performance is stable after this iteration is executed 2 times. The recursive procedure PruningTree(v) adopts a bottom-up fashion to prune the decision tree with v as the root. After pruning  branch(v)  (subtree(v)), I saves the sum of the importance of leaf nodes in branch(v). Then I(branch(v), F) is equal to the sum of importance of the tree with v as root. The essence of using T's root to call PruningTree is to travel T. If current node v is a nonleaf, the procedure calculates v's importance gain IG, saves into I the importance sum of the leaves of branch(v) (lines (2)~(7)), and determines pruning branch(v) or not based on the difference between CG and the threshold value δ (lines (8)~(9)).

4.4. Discussion

Suppose pruning set D′ contains n instances, forest F contains M decision trees, and dmax is the depth of the deepest decision tree in F. Let |T| be the number of nodes in decision tree T, and tmax = max1≤⁡(|T|). The running time of FP is dominated by the loop from lines (4) to (19). The loop from lines (5) to (7) traverses T, which is can be done in O(tmax); the loop from lines (8) to (14) searches a path of T for each instance in D′, which is complexity of O(ndmax); the main operation of PruningTree(root(T)) is a complete traversal of T, whose running time is O(tmax); the loop from lines (16) to (18) scans a linear list of length n in O(n). Since tmax ≪ ndmax, we conclude the running time of FP is O(nMdmax). Therefore, FP is a very efficient forest pruning algorithm. Unlike traditional metrics such as those used by CART [22] and C4.5 [23], the proposed measure uses a global evaluation. Indeed, this measure involves the prediction values that result from a majority voting of the whole ensemble. Thus, the proposed measure is based on not only individual prediction properties of ensemble members but also the complementarity of classifiers. From equations (9), (10), (11), and (12), our proposed measure takes into account both the correctness of predictions of current classifier and the predictions of ensemble and the measure deliberately favors classifiers with a better performance in classifying the samples on which the ensemble does not work well. Besides, the measure considers not only the correctness of classifiers, but also the diversity of ensemble members. Therefore, using the proposed measure to prune an ensemble leads to significantly better accuracy results.

5. Experiments

5.1. Experimental Setup

19 data sets of which the details are shown in Table 1 are randomly selected from UCI repertory [30], where #Size, #Attrs, and #Cls are the size, attribute number, and class number of each data set, respectively. We design four experiments to study the performance of the proposed method (forest pruning, FP):
Table 1

The details of data sets used in this paper.

Data set#Attrs#Size#Cls
Australian2267024
Autos205266
Backache180332
Balance-scale62553
Breast-cancer268102
Cars172874
Credit-rating690162
German-credit1000212
Ecoli33688
Hayes-roth16054
Heart-c303145
Horse-colic368242
Ionosphere351352
Iris15053
Lymph148194
Page-blocks5473115
Pima76892
prnn-fglass214106
Vote439172
The first experiment studies FP's performance versus the times of running FP. Here, four data sets, that is, autos, balance-scale, German-credit, and pima, are selected as the representatives, and each data set is randomly divided into three subsets with equal size, where one is used as the training set, one as the pruning set, and the other one as the testing set. We repeat 50 independent trials on each data set. Therefore a total of 300 trials of experiments are conducted. The second experiment is to evaluate FP's performance versus FL's size (number of base classifiers). The experimental setup of data sets is the same as the first experiment. The third experiment aims to evaluate FP's performance on pruning ensemble constructed by bagging [1] and random forest [26]. Here, tenfold cross-validation is employed: each data set is divided into tenfold [31, 32]. For each one, the other ninefold is to train model, and the current one is to test the trained model. We repeat 10 times the tenfold cross-validation and thus, 100 models are constructed on each data set. Here, we set the training set as the pruning set. Besides, algorithm rank is used to further test the performance of algorithms [31-33]: on a data set, the best performing algorithm gets the rank of 1.0, the second best performing algorithm gets the rank of 2.0, and so on. In case of ties, average ranks are assigned. The last experiment is to evaluate FP's performance on pruning the subensemble obtained by ensemble selection method. EPIC [11] is selected as the candidate of ensemble selection methods. The original ensemble is a library with 200 base classifiers, and the size of subsembles is 30. The setup of data sets is the same as the third experiment. In the experiments, bagging is used to train original ensemble, and the base classifier is J48, which is a Java implementation of C4.5 [23] from Weka [34]. In the third experiment, random forest is also used to build forest. In the last three experiments, we run FP two times.

5.2. Experimental Results

The first experiment is to investigate the relationship of the performance of the proposed method (FP) and the times of running FP. In each trial, we first use bagging to learn 30 unpruned decision trees as a forest and then iteratively run lines (3)~(14) of FP many times to trim the forest. More experimental setup refers to Section 5.1. The corresponding results are shown in Figure 2, where the top four subfigures are the variation trend of forest nodes number with the iteration number increasing, and the bottom four are the variation trend of ensemble accuracy. Figure 2 shows that FP significantly reduces forests size (almost 40%~60% of original ensemble) and significantly improves their accuracy. However, the performance of FP is almost stable after two iterations. Therefore, we set iteration number to be 2 in the following experiments.
Figure 2

Results on data sets. (a) Forest size (node number) versus the times of running FP. (b) Forest accuracy versus the times of running FP.

The second experiment aims at investigating the performance of FP on pruning forests with different scales. The number of decision trees grows gradually from 10 to 200. More experimental setup refers to Section 5.1. The experimental results are shown in Figure 3, where the top four subfigures are the comparison between pruned and unpruned ensembles with the growth of the number of decision trees, and the bottom four are the comparison of ensemble accuracy. As shown in Figure 3, for each data set, the rate of forest nodes pruned by FP keeps stable and forests accuracy improved by FP is also basically unchanged, no matter how many decision trees are constructed.
Figure 3

Results on data sets. (a) Forest size (node number) versus the number of decision trees. (b) Forest accuracy versus the number of decision trees. Solid curves and dash curves represent the performance of FP and bagging, respectively.

The third experiment is to evaluate the performance of FP on pruning the ensemble constructed by ensemble learning method. The setup details are shown in Section 5.1. Tables 2, 3, 4, and 5 show the experimental results of compared methods, respectively, where Table 2 reports the mean accuracy and the ranks of algorithms, Table 3 reports the average ranks using nonparameter Friedman test [32] (using STAC Web Platform [33]), Table 4 reports the comparing results using post hoc with Bonferroni-Dunn (using STAC Web Platform [33]) of 0.05 significance level, and Table 5 reports the mean node number and standard deviations. Standard deviations are not provided in Table 2 for clarity. The column of “FP” of Table 2 is the results of pruned forest and, “bagging” and “random forest” are the results of unpruned forests constructed by bagging and random forest, respectively. In Tables 3 and 4, Alg1, Alg2, Alg3, Alg4, Alg5, and Alg6 indicate PF pruning bagging with unpruned C4.5, bagging with unpruned C4.5, PF pruning bagging with pruned C4.5, bagging with pruned C4.5, PF pruning random forest, and random forest. From Table 2, FP significantly improves ensemble accuracy in most of the 19 data sets, no matter whether the individual classifiers are pruned or unpruned, no matter whether the ensemble is constructed by bagging or random forest. Besides, Table 2 shows that the ranks of FP always take place of best three methods in these data sets. Tables 3 and 4 validate the results in Table 2, where Table 3 shows that the average rank of PF is much small than other methods and Table 4 shows that, compared with other methods, PF shows significant better performance. Table 5 shows FP is significantly smaller than bagging and random forest, no matter whether the individual classifier is pruned or not.
Table 2

The accuracy of FP, bagging, and random forest. ∙ represents that FP outperforms bagging in pairwise t-tests at 95% significance level and denotes that FP is outperformed by bagging.

Dataset Unpruned C4.5Pruned C4.5PFRF
PFBaggingPFBagging
Australian87.14 (2.0)86.09 (5.0)86.80 (3.0)85.86 (6.0)87.21 (1.0)86.14 (4.0)
Autos74.40 (2.0)73.30 (4.0)74.20 (3.0)73.20 (5.0)74.72 (1.0)73.10 (6.0)
Backache85.07 (3.0)83.17 (5.5)85.89 (1.0)83.17 (5.5)85.21 (2.0)83.22 (4.0)
Balance-scale78.89 (3.0)75.07 (6.0)79.79 (1.0)76.64 (4.0)79.65 (2.0)76.32 (5.0)
Breast-cancer69.98 (2.0)67.10 (5.0)69.97 (3.0)66.58 (6.0)70.11 (1.0)68.88 (4.0)
Cars86.51 (4.0)86.78 (2.0)86.88 (1.0)86.28 (5.0)86.55 (3.0)86.11 (6.0)
Credit-rating86.44 (2.0)85.54 (4.0)86.34 (3.0)85.43 (5.0)86.82 (1.0)85.42 (6.0)
German-credit75.33 (1.0)73.83 (4.0)74.86 (3.0)73.11 (6.0)75.22 (2.0)73.18 (5.0)
Ecoli84.47 (2.0)83.32 (6.0)84.20 (3.0)83.40 (5.0)84.52 (1.0)83.89 (4.0)
Hayes-roth78.75 (3.0)78.63 (5.0)78.77 (1.0)76.31 (6.0)78.76 (2.0)77.77 (4.0)
Heart-c80.94 (2.0)80.34 (5.0)81.01 (1.0)80.27 (6.0)80.90 (3.0)80.87 (4.0)
Horse-colic84.52 (1.0)83.29 (6.0)84.33 (2.0)83.42 (5.0)84.31 (3.0)83.99 (4.0)
Ionosphere93.99 (1.0)93.93 (2.0)93.59 (6.0)93.71 (4.0)93.87 (3.0)93.56 (5.0)
Iris93.55 (6.0)94.24 (4.0)94.52 (3.0)94.53 (2.0)94.21 (5.0)94.62 (1.0)
Lymphography83.81 (5.0)83.43 (6.0)84.55 (2.0)84.53 (3.0)84.38 (4.0)84.82 (1.0)
Page-blocks97.03 (4.5)97.04 (2.5)97.04 (2.5)97.06 (1.0)97.03 (4.5)97.01 (6.0)
Pima75.09 (3.0)74.27 (4.0)75.46 (1.0)74.06 (5.0)75.43 (2.0)73.21 (6.0)
prnn-fglass78.14 (4.0)78.46 (1.0)77.62 (6.0)77.84 (5.0)78.18 (3.0)78.32 (2.0)
Vote95.77 (1.0)95.13 (6.0)95.67 (3.0)95.33 (4.0)95.72 (2.0)95.31 (5.0)
Table 3

The ranks of algorithms using Friedman test, where Alg1, Alg2, Alg3, Alg4, Alg5, and Alg6 indicate PF pruning bagging with unpruned C4.5, bagging with unpruned C4.5, PF pruning bagging with pruned C4.5, bagging with pruned C4.5, PF pruning random forest, and random forest.

AlgorithmAlg5Alg3Alg1Alg2Alg6Alg4
Ranks2.392.502.714.324.424.66
Table 4

The testing results using post hoc, Alg1, Alg2, Alg3, Alg4, Alg5, and Alg6 indicate PF pruning bagging with unpruned C4.5, bagging with unpruned C4.5, PF pruning bagging with pruned C4.5, bagging with pruned C4.5, PF pruning random forest, and random forest.

ComparisonStatistic p value
Alg1 versus Alg22.644690.04088
Alg3 versus Alg43.555150.00189
Alg5 versus Alg63.338370.01264
Table 5

The size (node number) of PF and bagging. ∙ denotes that the size of PF is significantly smaller than the corresponding comparing method.

Dataset Unpruned C4.5Pruned C4.5PF-RFRF
PFBaggingPFBagging
Australian4440.82 ± 223.245950.06 ± 210.532194.71 ± 99.652897.88 ± 98.661989.67 ± 99.652653.88 ± 99.61
Autos1134.83 ± 193.451813.19 ± 183.49987.82 ± 198.221523.32 ± 193.22954.26 ± 198.221429.12 ± 182.21
Backache1162.79 ± 96.581592.80 ± 75.97518.77 ± 40.49764.24 ± 37.78522.74 ± 40.49789.23 ± 45.62
Balance-scale3458.52 ± 74.554620.58 ± 78.203000.44 ± 71.763762.60 ± 65.552967.44 ± 71.763763.19 ± 79.46
Breast-cancer2164.64 ± 156.413194.20 ± 144.95843.96 ± 129.441189.33 ± 154.08886.66 ± 129.441011.21 ± 148.92
Cars1741.68 ± 60.592092.20 ± 144.951569.11 ± 57.551834.91 ± 46.801421.32 ± 56.651899.92 ± 68.88
Credit-rating4370.65 ± 219.275940.51 ± 223.512168.11 ± 121.512904.40 ± 99.732015.21 ± 140.582650.40 ± 102.13
German-credit9270.75 ± 197.6211464.19 ± 168.634410.11 ± 114.945421.60 ± 107.244311.54 ± 124.685340.60 ± 217.48
Ecoli1366.62 ± 61.681736.52 ± 64.911304.30 ± 54.391611.02 ± 56.311324.30 ± 54.421820.02 ± 88.74
Hayes-roth498.65 ± 28.99697.58 ± 40.87272.30 ± 45.11308.48 ± 53.86264.24 ± 46.46299.48 ± 63.84
Heart-c1503.46 ± 65.471946.94 ± 62.52647.89 ± 102.15974.93 ± 129.83647.89 ± 102.151032.93 ± 111.57
Horse-colic2307.67 ± 106.993625.23 ± 116.63684.29 ± 106.35974.93 ± 129.83647.89 ± 102.15743.25 ± 120.43
Ionosphere552.49 ± 61.41680.43 ± 69.95521.83 ± 58.01634.73 ± 64.44542.58 ± 96.02665.84 ± 66.44
Iris168.46 ± 111.12222.66 ± 150.42144.52 ± 97.26191.84 ± 133.12133.24 ± 98.32212.55 ± 129.47
Lymphography1089.87 ± 67.161394.37 ± 61.85711.62 ± 37.61856.44 ± 30.83724.53 ± 37.61924.33 ± 50.78
Page-blocks1420.05 ± 278.512187.45 ± 555.021394.11 ± 600.062092.93 ± 403.791401.11 ± 588.032134.40 ± 534.97
Pima2202.41 ± 674.182776.77 ± 852.952021.19 ± 698.022481.64 ± 747.191927.67 ± 625.272521.43 ± 699.82
prnn-fglass1219.98 ± 39.851398.62 ± 36.291145.20 ± 39.761269.28 ± 35.521098.18 ± 34.261314.05 ± 60.97
Vote303.06 ± 124.00527.80 ± 225.05174.04 ± 77.61276.00 ± 127.46182.14 ± 76.21288.33 ± 113.76
The last experiment is to evaluate the performance of FP on pruning subensembles selected by ensemble selection method EPIC. Table 6 shows the results on the 19 data sets, where left and right are the accuracy and size, respectively. As shown in Table 6, FP can further significantly improve the accuracy of subensembles selected by EPIC and reduce the size of the subensembles.
Table 6

The performance of FP on pruning subensemble obtained by FP on bagging. ∙ represents that FP is significantly better (or smaller) than EPIC in pairwise t-tests at 95% significance level and denotes that FP is significantly worse (or larger) than EPIC.

DatasetError rateSize
PFEPICPFEIPC
Australian86.83 ± 3.7286.22 ± 3.692447.50 ± 123.933246.16 ± 116.07
Autos84.83 ± 4.4682.11 ± 5.89708.01 ± 54.55931.44 ± 51.16
Backache84.83 ± 4.4682.11 ± 5.89708.01 ± 54.55931.44 ± 51.16
Balance-scale79.74 ± 3.6978.57 ± 3.823277.76 ± 85.074030.82 ± 94.67
Breast-cancer70.26 ± 7.2467.16 ± 8.36843.96 ± 129.441189.33 ± 154.08
Cars87.02 ± 5.0686.83 ± 5.04178.32 ± 60.442022.81 ± 53.19
Credit-rating86.13 ± 3.9285.61 ± 3.952414.60 ± 123.663226.25 ± 131.46
German-credit74.98 ± 3.6373.13 ± 4.004410.11 ± 114.946007.28 ± 124.30
Ecoli83.77 ± 5.9683.24 ± 5.981498.86 ± 62.271806.26 ± 70.98
Hayes-roth78.75 ± 9.5776.81 ± 9.16275.09 ± 47.90311.32 ± 57.05
Heart-c81.21 ± 6.3779.99 ± 6.651230.14 ± 54.801510.57 ± 52.56
Horse-colic84.53 ± 5.3083.80 ± 6.11940.07 ± 66.641337.60 ± 75.73
Ionosphere93.90 ± 4.0594.02 ± 3.83590.63 ± 65.62706.79 ± 73.17
Iris94.47 ± 5.1194.47 ± 5.02152.58 ± 108.04197.80 ± 141.31
Lymphography81.65 ± 9.4581.46 ± 9.39858.42 ± 46.501022.67 ± 39.68
Page-blocks97.02 ± 0.7497.07 ± 0.691396.63 ± 237.032086.89 ± 399.10
Pima74.92 ± 3.9474.03 ± 3.582391.95 ± 764.162910.31 ± 936.70
prnn-fglass78.13 ± 8.0677.99 ± 8.441280.14 ± 43.851410.84 ± 39.59
Vote95.70 ± 2.8695.33 ± 2.97177.36 ± 86.10281.62 ± 140.60

6. Conclusion

An ensemble with decision trees is also called forest. This paper proposes a novel ensemble pruning method called forest pruning (FP). FP prunes trees' branches based on the proposed metric called branch importance, which indicates the importance of a branch (or a node) with respect to the whole ensemble. In this way, FP achieves reducing ensemble size and improving the ensemble accuracy. The experimental results on 19 data sets show that FP significantly reduces forest size and improves its accuracy in most of the data sets, no matter whether the forests are the ensembles constructed by some algorithm or the subensembles selected by some ensemble selection method, no matter whether each forest member is a pruned decision tree or an unpruned one.
  3 in total

1.  Ensemble learning via negative correlation.

Authors:  Y Liu; X Yao
Journal:  Neural Netw       Date:  1999-12

2.  Rotation forest: A new classifier ensemble method.

Authors:  Juan J Rodríguez; Ludmila I Kuncheva; Carlos J Alonso
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2006-10       Impact factor: 6.226

3.  An analysis of ensemble pruning techniques based on ordered aggregation.

Authors:  Gonzalo Martínez-Muñoz; Daniel Hernández-Lobato; Alberto Suárez
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2009-02       Impact factor: 6.226

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.