Literature DB >> 33267251

Discriminatory Target Learning: Mining Significant Dependence Relationships from Labeled and Unlabeled Data.

Zhi-Yi Duan1, Li-Min Wang1, Musa Mammadov2, Hua Lou3, Ming-Hui Sun4.   

Abstract

Machine learning techniques have shown superior predictive power, among which Bayesian network classifiers (BNCs) have remained of great interest due to its capacity to demonstrate complex dependence relationships. Most traditional BNCs tend to build only one model to fit training instances by analyzing independence between attributes using conditional mutual information. However, for different class labels, the conditional dependence relationships may be different rather than invariant when attributes take different values, which may result in classification bias. To address this issue, we propose a novel framework, called discriminatory target learning, which can be regarded as a tradeoff between probabilistic model learned from unlabeled instance at the uncertain end and that learned from labeled training data at the certain end. The final model can discriminately represent the dependence relationships hidden in unlabeled instance with respect to different possible class labels. Taking k-dependence Bayesian classifier as an example, experimental comparison on 42 publicly available datasets indicated that the final model achieved competitive classification performance compared to state-of-the-art learners such as Random forest and averaged one-dependence estimators.

Entities:  

Keywords:  Bayesian network; discriminatory target learning; unlabeled instance

Year:  2019        PMID: 33267251      PMCID: PMC7515026          DOI: 10.3390/e21050537

Source DB:  PubMed          Journal:  Entropy (Basel)        ISSN: 1099-4300            Impact factor:   2.524


1. Introduction

With the rapid development of computer technologies, business and government organizations create large amounts of data, which need to be processed and analyzed. Over the past decade, to satisfy the urgent need of mining knowledge hidden in the data, numerous machine learning models [1,2] (e.g., decision tree [3], Bayesian network [4,5], support vector machine [6] and Neural network [7]) have been proposed. To mine all “right” knowledge that exist in a database, researchers mainly proposed two kinds of learning strategies to address this issue. (1) Increase structure complexity to represent more dependence relationships, e.g., convolutional neural network [8] and k-dependence Bayesian classifier (KDB) [9]. However, as structure complexity grows overfitting will inevitably appear, which will result in redundant dependencies and performance degradation. Sometimes the overly complex structures hide the internal working mechanism and make them criticized for being used as “black box”. (2) Build ensemble of several individual members having relatively simple network structure, e.g., Random forest [10] and averaged one-dependence estimators (AODE) [11]. Ensembles can generally perform better than any individual member. However, it is difficult or even impossible to give a clear semantic explanation of the combined result since the working mechanisms of individual members may differ greatly. In practice, people would rather use models with simple and easy-to-explain structures, e.g., decision tree [12] and Naive Bayes (NB) [13,14,15], although they may perform poorer. Bayesian networks (BNs) have long been a popular medium for graphically representing the probabilistic dependencies, which exist in a domain. Recently, work in Bayesian methods for classification has grown enormously. Numerous Bayesian network classifiers (BNCs) [9,16,17,18,19,20] have been proposed to mine the significant dependence relationships implicated in training data. With solid theoretic support, they have strong potential to be effective for practical application in a number of massive and complex data-intensive fields such as medicine [21], astronomy [22], biology [23], and so on. A central concern for BNC is to learn conditional dependence relationships encoded in the network structure. Some BNCs, e.g., KDB, use conditional mutual information to measure the conditional dependence relationships between and , which is defined as follows [24], For example, indicates that attributes and are conditionally independent. However, in practice, for any specific event or data point, the situation will be much more complex. Taking Waveform dataset as an example, attributes and are conditionally dependent, since always holds. Figure 1 shows the distributions of , where . As can be seen, there exist some positive values of and . However, for the class label , the negative or zero values of have a high proportion among all values. That is, for different class labels, the conditional dependence relationships may be different rather than invariant when attributes take different values. We argue that most BNCs (e.g., NB and KDB), which build only one model to fit training instances, cannot capture this difference and cannot represent the dependence relationships flexibly, especially hidden in unlabeled instances.
Figure 1

The distributions of on Waveform dataset, where . The x-axis represents the index of each instance, the y-axis represents the value of .

The scientific data can be massive, and labeled training data may account for only a small portion. In this paper, we propose a novel learning framework, called discriminatory target learning, for achieving better classification performance and high-level of dependence relationships while not increasing structure complexity. KDB is taken as an example to illustrate the basic idea and prove the feasibility of discriminatory target learning. By redefining mutual information and conditional mutual information, we build a “precise” model kdb for each unlabeled instance with respect to class label . The ensemble of kdb, i.e., kdb, can finely describe the dependency relationships hidden in . The final ensemble of kdb and regular KDB can fully and discriminately describe the dependence relationships in training data and unlabeled instance. The rest of the paper is organized as follows: Section 2 introduces some state-of-the-art BNCs. Section 3 introduces the basic idea of discriminatory target learning. Experimental study on 42 UCI machine learning datasets is presented in Section 4, including a comparison with seven algorithms. The final section draws conclusions and outlines some directions for further research.

2. Bayesian Network Classifiers

The structure of a BN on the random variables is a directed acyclic graph (DAG), which represents each attribute in a given domain as a node in the graph and dependencies between these attributes as arcs connecting the respective nodes. Thus, independencies are represented by the lack of arcs connecting particular nodes. BNs are powerful tools for knowledge representation and inference under conditions of uncertainty. BNs were considered as classifiers only after the discovery of NB, a very simple kind of BN on the basis of conditional independence assumption. It is surprisingly effective and efficient for inference [5]. The success of NB has led to the research of Bayesian network classifiers (BNCs), including tree-augmented naive Bayes (TAN) [16], averaged one-dependence estimators (AODE) [18] and k-dependence Bayesian classifier (KDB) [9,17]. Let each instance x be characterized with n values for attributes , and class label is the value of class variable Y. NB assumes that the predictive attributes are conditional independent of each other given the class label, that is Correspondingly for any value pair of arbitrary two attributes and , always holds. From Equation (1) there will be and this can explain why there exist no arc between attributes for NB. However, in the real world, it will be much more complex when considering different specific event or data point. We now formalize our notion of the spectrum of point dependency relationship in Bayesian classification. For unlabeled data point Equation (2) is a modified version of pointwise conditional mutual information that is applicable to labeled data point [25]. By comparing Equations (1) and (2), is a summation of expected values of given all possible values of and Y. The traditional BNCs, e.g., TAN and KDB, use to roughly measure the conditional dependence between and . is non-negative, iff and are conditionally dependent given Y. However, only considering as the criterion for identifying the conditional independent relationship is too strict for BN learning, which may lead to classification bias, since may hold for specific data point . That may be the main reason why NB performs better in some research domains. To address this issue, in this paper is applied to measure the extent to which and are relatively conditionally dependent when or relatively conditionally independent or irrelevant when , respectively. For unlabeled data point TAN maintains the structure of NB and allows each attribute to have at most one parent. Then, the number of arcs encoded in TAN is . During the constructing procedure of maximum weighted spanning tree, TAN sorts the arcs between arbitrary attributes and by comparing , and adds them in turn to the network structure if no cycle appears. KDB further relaxes NB’s independence assumption and can represent arbitrary degree of dependence while capturing much of the computational efficiency of NB. KDB first sorts attributes by comparing mutual information , which is defined as follows [24], Suppose the attribute order is . By comparing , select its parents, e.g., , from attributes that ranks before it in the order. KDB requires that must have parents and there will exist arcs between and its parents. The number of arcs encoded in KDB is and will grow as k grows. Thus, KDB can represent more dependency relationships than TAN. For TAN or KDB, they do not evaluate the extent to which the conditional dependencies are weak enough and should be neglected. They simply specify the maximum number of parents that attribute can have before structure learning. Some arcs corresponding to weak conditional dependencies will inevitably be added to the network structure. The prior and joint probabilities in Equations (1) and (3) will be estimated from training data as follows: where N is the number of training instances. Then, and in Equations (1) and (3) can be computed as follows: Sahami [9] suggested that, if k is large enough to capture all “right” conditional dependencies that exist in a database, then a classifier would be expected to achieve optimal Bayesian accuracy. However, as k grows, KDB will encode more weak dependency relationships, which correspond to smaller value of . That increases the risk of occurrence of negative values of and may introduce redundant dependencies, which will mitigate the positive effect from significant dependencies that correspond to positive values of . On the other hand, conditional mutual information cannot finely measure the conditional dependencies hidden in different data points. The arc in BNC learned from training data corresponds to positive value of and represents strong conditional dependence between and . However, for specific labeled instance , may hold. Then, and are -conditionally independent or irrelevant on point and the arc should be removed. For unlabeled instance, the possible dependency relationships between nodes may differ greatly with respect to different class labels. Thus, BNCs with highly complex network structure do not necessarily beat those with simple ones. The conditional dependencies hold for training data in general do not necessarily hold for each instance. BNCs should discriminate between conditionally dependent and irrelevant relationship for different data points. Besides, BNC should represent all possible spectrums of point dependency relationship that correspond to different class labels for dependence analysis.

3. Discriminatory Target Learning

In probabilistic classification, Bayes optimal classification suggests that, if we can determine the conditional probability distribution with true distribution available, where y is one of the m class labels and is the n-dimensional data point that represents an observed instance, then we could achieve the theoretically optimal classification. can be described in an unrestricted Bayesian network, as shown in Figure 2a. By applying arc reversal, Shachter [26] proposed to produce the equivalent dependence structure, as shown in Figure 2b. The problem is reduced to estimating the conditional probability . Figure 2a,b represents two inference processes that run in the opposite directions. Figure 2a indicates the causality that runs from the state of (the cause) to the state of Y (the effect). In contrast, if the causality runs in the opposite direction as shown in Figure 2b and the state of Y (the effect) is uncertain, the dependencies between predictive attributes (the causes) should be tuned to match with different states of Y. That is, the restricted BNC shown in Figure 2b presupposes the class label first and then the conditional dependencies between attributes can verify the presupposition.
Figure 2

Example of (a) unrestricted BNC, and (b) restricted BNC.

For different class labels or presuppositions, the conditional dependencies should be different. It is not reasonable that, no matter what the effect (class label) is, the relationships between causes (predictive attributes) remain the same. Consider an unlabeled instance ; if , then the conditional dependence between and on data point with respect to class label y is reasonable, otherwise it should be neglected. Since the class label for is uncertain and there are m labels available, we take as the target and learn an ensemble of m micro BNCs, i.e., bnc = {bnc1, ⋯, bnc}, each of them fully describes the conditional dependencies between attribute values in with respect to different class labels. The linear combiner is used for models that output real-valued numbers, thus is applicable for bnc. The ensemble probability estimate for bnc is, bnc may overfit the unlabeled instance and underfit training data. In contrast, regular BNC learned from training data may underfit the unlabeled instance. Thus, they are complementary in nature. After training bnc and regular BNC, the final ensemble that estimates the class membership probabilities by averaging both predictions will be generated. The framework of discriminatory target learning is shown in Figure 3.
Figure 3

The framework of discriminatory target learning.

Because in practice it is hardly possible to find the true distribution of from data, KDB approximates the estimation of by allowing for the modeling of arbitrarily complex dependencies between attributes. The pseudocode of KDB is shown in Algorithm 1. From the definition of in Equation (3), we can have For unlabeled data point Equation (8) is a modified version of pointwise mutual information that is applicable to labeled data point [25]. The prior and joint probabilities in Equations (2) and (8) will be estimated as follows Conditional probabilities in Equations (2) and (8) can be estimated by: Similar to the Laplace correction [27], the main idea behind Equation (9) is equivalent to creating a “pseudo” training set by adding to the training data a new instance with multi-label by assuming that the probability that this new instance is in class y is for each For unlabeled data point KDB uses to sort the attributes and to measure the conditional dependence. Similarly, for unlabeled instance , the corresponding micro KDB with respect to class label , called kdb, uses (see Equation (8)) to sort the attribute values and (see Equation (2)) to measure the conditional dependence. The learning procedure of kdb is shown in Algorithm 2. Breiman [28] revealed that ensemble learning brings improvement in accuracy only to those “unstable” learning algorithms, in the sense that small variations in the training set would lead them to produce very different models. bnc is obviously an example of such learners. For individual members of kdb, the difference in network structure is the result of change of or , or, more precisely, the conditional probability defined in Equations (2) and (8). Given unlabeled instance and binary class labels and , if , i.e., , then is -dependent on . Because and we have and Thus, is -irrelevant on . plays totally different roles in the relationships with different class labels on the same instance. Supposing that before small variations in the training set and after that , the attribute values will be resorted and correspondingly the network structures of kdb1 and kdb2 for will change greatly. The sensitivity to the variation makes kdb finely describe the dependencies hidden in . Figure 4 shows examples of kdb1 and kdb2 corresponding to class labels and , respectively. If the decision of the final ensemble is , then we will use Figure 4a for dependence analysis. Otherwise, we will use Figure 4b instead. The attribute values annotated in black correspond to positive values of or 2) and they should be focused on.
Figure 4

Example of (a) kdb1, and (b) kdb2.

KDB requires training time complexity of (dominated by the calculations of ) and classification time complexity of [9] for classifying a single unlabeled instance, where n is the number of attributes, N is the number of data instances, m is the number of class labels, and v is the maximum number of discrete values that an attribute may take. Discriminatory target learning requires no additional training time, thus the training time complexity of final ensemble is the same as that of regular KDB. At classification time it requires to calculate , and the same time complexity for classifying a single unlabeled instance.

4. Experiments and Results

We compared the performance of our proposed methods kdb and KDB with several state-of-the-art classifiers. We analyzed the performance in terms of zero-one loss, root mean square error (RMSE), bias and variance on 42 natural domains from the UCI Machine Learning Repository [29]. These datasets are described in Table 1, in ascending order of number of instances. The structure of this section is as follows: we discuss our experimental methodology and evaluation function in details in Section 4.1. Section 4.2 includes comparisons with three classic single-structure BNCs, namely NB, TAN and KDB, as well as one ensemble BNC: AODE. Then, in Section 4.3, KDB is compared with Random Forest with 100 decision trees. Section 4.4 presents a global comparison of all learners considered by applying the Friedman and Nemenyi tests.
Table 1

Datasets. Imbalanced datasets are annotated with the symbol “*”.

IndexDatasetInstanceAttributeClassIndexDatasetInstanceAttributeClass
1Contact-lenses244322Kr-vs-kp3196362
2Labor5716223Dis *3772292
3Echocardiogram1316224Hypo3772294
4Lymphography14818425Sick *3772292
5Sonar20860226Abalone *417783
6Glass-id2149327Waveform-50005000403
7New-thyroid *2155328Phoneme5438750
8Heart-disease-c30313229Wall-following5456244
9Soybean-large307351930Page-blocks5473105
10Ionosphere *35134231Satellite *6435366
11Dermatology36634632Thyroid91692920
12House-votes-84 *43516233Pendigits10,9921610
13Chess *55139234Sign12,54683
14Soybean *683351935Nursery12,96085
15Breast-cancer-w6999236Magic19,020102
16Tic-tac-toe9589237Letter-recog20,0001626
17Vowel990131138Adult *48,842142
18Car *17286439Shuttle *58,00097
19Mfeat-mor200061040Connect-467,557423
20Segment231019741Waveform *100,000213
21Hypothyroid *316325242Localization164,860511

4.1. Experimental Methodology and Evaluation Function

The experiments for all BNCs used C++ software (NetBeans 8.0.2) specially designed to deal with classification problems. Each algorithm was tested on each dataset using 10-fold cross validation. All experiments were conducted on a desktop computer with an Intel(R) Core(TM) i3-6100 CPU @ 3.70 GHz, 64 bits and 4096 MB of memory(Dell Vostro 2667, Changchun, China). Win/Draw/Lose (W/D/L) Record: When two algorithms were compared, we counted the number of datasets for which one algorithm performed better, equally well or worse than the other on a given measure. We considered there exists a significant difference if the output of a one-tailed binomial sign test was less than 0.05. Missing Values: Missing values for qualitative attributes were replaced with modes, and those for quantitative attributes were replaced with means from the training data. Numeric Attributes: For each dataset, we used MDL (Minimum Description Length) discretization [30] to discretize numeric attributes. Dataset Sizes: Datasets were categorized in terms of their sizes. That is, datasets with instances <1000, ≥1000 and <10,000, ≥10,000 were denoted as small size, medium size and large size, respectively. We report results on these sets to discuss suitability of a classifier for datasets of different sizes. Zero-one loss: Zero-one loss can be used to measure the extent to which a learner correctly identifies the class label of an unlabeled instance. Supposing y and are the true class label and that generated by a learning algorithm, respectively, given M unlabeled test instances, the zero-one loss function is defined as where if and 0 otherwise. Bias and variance: The bias-variance decomposition proposed by Kohavi and Wolpert [31] provides valuable insights into the components of the zero-one loss of learned classifiers. Bias measures how closely the classifier can describe the decision boundary, which is defined as where is the combination of any attribute value. Variance measures the sensitivity of the classifier to variations in the training data, which is defined as RMSE: For each instance, RMSE accumulates the squared error, where the error is the difference between 1.0 and the probability estimated by the classifier for the true class for the instance, and then computes the squared root of the mean of the sum, which is defined as where s is the sum of training instances.

4.2. KDBe Versus Classic BNCs

We compared KDB with several classic BNCs, namely NB, TAN, KDB and AODE. Sahami [9] proposed the notion of k-dependence BNC, which allows each attribute to have a maximum of k attributes as parents. NB and TAN are, respectively, 0-dependence and 1-dependence BNCs. To clarify the effect of dependence complexity, we set for both KDB and KDB.

4.2.1. Zero-One Loss and RMSE Results

The detailed results in terms of zero-one loss and RMSE are shown in Table A1 and Table A2 in Appendix A, respectively. Table 2 and Table 3 show W/D/L records summarizing the relative zero-one loss and RMSE of different BNCs. When , NB, TAN and KDB can, respectively, represent 0, and conditional dependencies, where n is the number of predictive attributes. As shown in Table 1, since holds for all datasets, also holds. Thus, KDB can represent the largest number of dependencies among all. With respect to zero-one loss, NB represents no conditional dependencies due to its independence assumption and performed the worst in general. As the dependence degree or structure complexity increased, KDB was competitive compared to NB and TAN. AODE performed better than the other single-structure BNCs due to its ensemble mechanism. Surprisingly, kdb had significantly better zero-one loss performance than NB, TAN and KDB. When discriminatory target learning was introduced for discovery of dependencies that exist in different unlabeled instances, the final ensemble KDB could possess significant advantage over other classifiers. For example, KDB beat KDB in 26 domains and lost only in three in terms of zero-one loss. RMSE-wise, KDB still performed the best. For instance, KDB enjoyed a significant advantage over TAN (20/19/3). When compared to KDB, KDB also achieved superior performance, with 17 wins and 5 losses.
Table A1

Experimental results of average zero-one loss.

DatasetNBTANKDBAODERFkdbeKDBe
Contact-lenses0.37500.3750 0.2500 0.37500.34380.37500.2917
Labor 0.0351 0.0526 0.0351 0.05260.0939 0.0351 0.0351
Echocardiogram0.33590.32820.34350.32060.3489 0.3130 0.3130
Lymphography 0.1486 0.17570.23650.16890.21320.17570.2095
Sonar0.23080.22120.24520.2260 0.2067 0.24520.2308
Glass-id0.26170.21960.21960.2523 0.2132 0.22430.2150
New-thyroid0.05120.06510.0698 0.0465 0.08160.06050.0605
Heart-disease-c 0.1815 0.20790.22440.20130.22120.19470.2079
Soybean-large0.12380.11070.0879 0.0782 0.11070.12700.0814
Ionosphere0.10540.06840.07410.07410.07660.0912 0.0655
Dermatology0.01910.03280.0656 0.0164 0.03670.05460.0519
House-votes-840.09430.05520.05060.0529 0.0416 0.05750.0437
Chess0.1125 0.0926 0.09980.09980.1074 0.0926 0.0926
Soybean0.0893 0.0469 0.0556 0.0469 0.07030.05420.0527
Breast-cancer-w 0.0258 0.04150.07440.03580.03860.04010.0629
Tic-tac-toe0.30690.22860.20350.26510.2115 0.1931 0.2004
Vowel0.4242 0.1303 0.18180.14950.16740.17880.1626
Car0.14000.0567 0.0382 0.08160.07720.05960.0411
Mfeat-mor0.3140 0.2970 0.30600.31450.30000.30150.3035
Segment0.07880.03900.0472 0.0342 0.04130.03550.0433
Hypothyroid0.01490.01040.01070.01360.0122 0.0092 0.0095
Kr-vs-kp0.12140.07760.04160.0842 0.0128 0.04600.0382
Dis0.01590.01590.01380.01300.01330.0127 0.0122
Hypo0.01380.01410.0114 0.0095 0.01220.00980.0098
Sick0.03080.0257 0.0223 0.02730.02630.02700.0233
Abalone0.47620.45870.4563 0.4472 0.48230.45340.4484
Waveform-50000.20060.18440.2000 0.1462 0.15580.17820.1756
Phoneme0.26150.27330.19840.2392 0.1789 0.31390.1931
Wall-following0.10540.05540.04010.0370 0.0216 0.03980.0387
Page-blocks0.06190.04150.03910.0338 0.0309 0.03230.0322
Satellite0.18060.12140.10800.11480.10850.1265 0.1052
Thyroid0.11110.07200.07060.07010.0750 0.0586 0.0642
Pendigits0.11810.03210.0294 0.0200 0.03390.02020.0248
Sign0.35860.27550.25390.2821 0.2038 0.26850.2419
Nursery0.09730.06540.02890.0730 0.0248 0.05090.0356
Magic0.22390.16750.16370.17520.16740.1716 0.1598
Letter-recog0.25250.13000.09860.08830.0902 0.0675 0.0861
Adult0.15920.13800.13830.1493 0.1204 0.13150.1316
Shuttle0.00390.00150.00090.0008 0.0005 0.00060.0007
Connect-40.27830.23540.22830.2420 0.1875 0.23370.2268
Waveform0.02200.02020.0256 0.0180 0.15580.01940.0193
Localization0.49550.35750.29640.35960.2976 0.2659 0.2743
Table A2

Experimental results of average RMSE.

DatasetNBTANKDBAODERFkdbeKDBe
Contact-lenses0.37780.4496 0.3639 0.40660.40980.40860.3825
Labor 0.1420 0.21850.16850.19000.28240.36470.2271
Echocardiogram0.48960.48860.48890.4903 0.4574 0.47820.4813
Lymphography 0.2446 0.26840.30310.24780.27010.27290.2680
Sonar0.44210.41310.40840.4285 0.3518 0.40710.3959
Glass-id0.35400.33320.33950.3439 0.3146 0.33110.3275
New-thyroid 0.1544 0.17310.17970.16140.15600.16890.1714
Heart-disease-c0.37430.37750.39630.36590.3696 0.3572 0.3690
Soybean-large0.10320.09630.08580.08580.11430.1051 0.0856
Ionosphere0.31570.26150.27140.2506 0.2403 0.28220.2523
Dermatology 0.0631 0.08510.12060.06920.13030.18570.1313
House-votes-840.29970.21810.19690.1994 0.1846 0.19620.1847
Chess0.2944 0.2594 0.26150.27250.27710.29370.2642
Soybean0.0933 0.0642 0.06540.06560.09220.07540.0643
Breast-cancer-w 0.1570 0.19280.24970.18480.17960.21940.2137
Tic-tac-toe0.43090.40230.37720.3995 0.2916 0.38300.3693
Vowel0.2270 0.1271 0.15820.14250.15810.16850.1516
Car0.22520.1617 0.1379 0.20050.17820.17490.1505
Mfeat-mor0.2086 0.1940 0.19740.19850.20740.19480.1954
Segment0.13980.09670.1034 0.0879 0.10610.09570.0919
Hypothyroid0.11380.09550.09370.1036 0.0770 0.09790.0913
Kr-vs-kp0.30220.23580.18690.2638 0.1268 0.26260.2091
Dis0.11770.11030.10240.1080 0.1011 0.10740.1021
Hypo0.07660.07380.06710.06500.07150.0719 0.0635
Sick0.17000.1434 0.1382 0.15720.14870.14890.1394
Abalone0.46300.42500.4277 0.4193 0.45390.42200.4220
Waveform-50000.33480.29470.3149 0.2659 0.30360.29500.2869
Phoneme0.08800.09020.07840.0885 0.0731 0.09520.0783
Wall-following0.21770.15860.13630.1292 0.1206 0.13150.1210
Page-blocks0.14500.11870.11280.10210.0974 0.0972 0.0991
Satellite0.24000.18510.17770.18000.16820.1865 0.1644
Thyroid0.09670.07460.07440.07450.0770 0.0674 0.0679
Pendigits0.14270.07250.0687 0.0568 0.09790.07930.0646
Sign0.39840.35050.33340.3524 0.3104 0.34680.3300
Nursery0.17660.13850.11210.1571 0.1010 0.13720.1217
Magic0.39740.34610.34700.35410.35710.3514 0.3411
Letter-recog0.11840.08600.07680.07070.08960.0756 0.0685
Adult0.34090.30760.30890.32450.32740.3021 0.3015
Shuttle0.02980.01820.01460.01260.0142 0.0121 0.0125
Connect-40.35870.33150.32470.3370 0.3057 0.34090.3279
Waveform0.11760.09510.11450.0860 0.0799 0.09990.0901
Localization0.23900.20950.19600.20950.1939 0.1834 0.1846
Table 2

W/D/L comparison results of zero-one loss on all datasets.

NBTANKDBAODEkdbe
TAN29/7/6----
KDB30/5/720/9/13---
AODE33/5/416/14/1220/6/16--
kdbe30/5/717/18/720/11/1113/15/14-
KDBe34/3/523/13/626/13/322/10/1014/20/8
Table 3

W/D/L comparison results of RMSE on all datasets.

NBTANKDBAODEkdbe
TAN32/4/6----
KDB32/4/616/19/7---
AODE29/9/416/19/715/15/12--
kdbe30/5/79/21/1211/17/147/19/16-
KDBe34/3/520/19/317/20/517/17/821/21/0
To make the experimental results more intuitive, from the viewpoints of the ensemble mechanism and structure complexity, Figure 5a,c shows the comparisons of KDB, KDB and AODE in terms of zero-one loss, whereas Figure 5b,d shows the comparisons for RMSE. The red squared symbols are used to indicate significant advantages of KDB over the other BNCs. In Figure 5a,b, only two points are far above the diagonal line, thus the negative effect caused by discriminatory target learning was negligible. In contrast, many more points are below the diagonal line, which means that discriminatory target learning worked effectively in most cases. A notable case is Waveform dataset, where discriminatory target learning helped to substantially reduce classification error, such as the reduction from 0.0256 to 0.0193 for zero-one loss and from 0.1145 to 0.0901 for RMSE. When comparing KDB with AODE, it can be seen in Figure 5c,d that there are still many points below the diagonal line, which means that KDB enjoyed a significant advantage over AODE. For example, a notable case is our largest dataset Localization, where the zero-one loss of KDB (0.2743) was much lower than that of AODE (0.3596).
Figure 5

Scatter plot of zero-one loss and RMSE comparisons for KDB, KDB and AODE.

4.2.2. Bias and Variance Results

The detailed results in terms of bias and variance are shown in Table A3 and Table A4 in Appendix A, respectively. The W/D/L records with respect to bias and variance results are shown in Table 4 and Table 5, respectively. We can observe in Table 4 that ensemble classifiers, i.e., AODE and kdb, performed better than TAN but worse than KDB, although these results were not always statistically significant. NB still performed the worst. High-dependence structure or ensemble construction strategy could help reduce the bias. Jointly applying both helped KDB reduce bias significantly. For example, KDB performed better than TAN (26/9/7) and KDB (11/27/4).
Table A3

Experimental results of average bias.

DatasetNBTANKDBAODERFkdbeKDBe
Contact-lenses0.21630.18250.31750.2850 0.1748 0.18630.2850
Labor0.02890.02110.02790.03470.0409 0.0184 0.0279
Echocardiogram0.28440.26420.30650.2751 0.2256 0.26020.2686
Lymphography 0.0902 0.10270.10410.09330.12880.09510.0996
Sonar0.16720.16460.16860.1696 0.1045 0.18290.1762
Glass-id0.29010.27560.27130.2785 0.1348 0.27300.2732
New-thyroid0.0290 0.0277 0.0348 0.0277 0.02850.02790.0396
Heart-disease-c0.12970.12630.12990.11380.1304 0.1128 0.1274
Soybean-large0.10700.14220.1086 0.0648 0.12130.17170.1112
Ionosphere0.12200.08040.08550.0744 0.0624 0.09120.0862
Dermatology0.00790.02740.0489 0.0055 0.01900.05410.0451
House-votes-840.08990.0410 0.0258 0.04300.03270.04570.0301
Chess0.14130.14370.11190.1290 0.0548 0.12650.1192
Soybean0.10150.0522 0.0491 0.05240.05860.09710.0502
Breast-cancer-w 0.0187 0.03840.04490.03380.03010.02210.0348
Tic-tac-toe0.26140.1746 0.1367 0.20050.02700.14340.1390
Vowel0.33010.19420.17450.1895 0.0756 0.18450.1736
Car0.09370.04780.03870.05560.03890.0389 0.0374
Mfeat-mor0.2624 0.2077 0.21420.24770.23110.22230.2166
Segment0.07850.04910.04530.0367 0.0253 0.03870.0419
Hypothyroid0.01160.01040.00960.00940.0516 0.0090 0.0094
Kr-vs-kp0.11070.07020.04170.0747 0.0063 0.04340.0407
Dis 0.0165 0.01930.01910.01700.02030.01920.0191
Hypo0.00920.01240.0077 0.0071 0.00830.00980.0073
Sick0.02460.02070.01980.0224 0.0194 0.02540.0196
Abalone0.41800.3126 0.3033 0.32010.32570.31950.3132
Waveform-50000.17620.12320.11570.1235 0.1114 0.12190.1147
Phoneme0.22160.23940.15720.2207 0.1102 0.29270.1551
Wall-following0.09510.04910.02570.0251 0.0122 0.02960.0245
Page-blocks0.04510.03080.02800.0251 0.0217 0.02770.0264
Satellite0.17460.09500.08080.09020.08740.1011 0.0802
Thyroid0.09940.05870.05530.06110.0516 0.0493 0.0531
Pendigits0.10950.03140.02070.02280.02160.0196 0.0189
Sign0.32570.24200.21610.2531 0.1540 0.23220.2132
Nursery0.09280.05210.02810.0651 0.0086 0.04000.0322
Magic0.21110.1252 0.1241 0.16000.12440.13230.1265
Letter-recog0.22070.10320.08060.0876 0.0490 0.07000.0732
Adult0.16490.13120.12200.1437 0.1109 0.12400.1226
Shuttle0.00400.00080.0007 0.0006 0.0006 0.0006 0.0006
Connect-40.26600.22530.20220.2264 0.1427 0.21690.2075
Waveform0.0219 0.0152 0.02100.01560.01580.01720.0161
Localization0.45230.31060.21340.31290.2047 0.2027 0.2038
Table A4

Experimental results of average variance.

DatasetNBTANKDBAODERFkdbeKDBe
Contact-lenses0.17130.19250.1700 0.1275 0.20130.21380.1775
Labor0.03950.06320.0721 0.0179 0.07580.06050.0721
Echocardiogram0.1272 0.1265 0.14000.13190.14690.13740.1337
Lymphography 0.0343 0.11160.14080.04760.13520.09270.1249
Sonar 0.0907 0.11650.11990.09420.11890.09830.1107
Glass-id 0.0930 0.10750.11890.10040.10890.11010.1099
New-thyroid 0.0161 0.02720.03850.02300.03650.02850.0351
Heart-disease-c 0.0248 0.04790.05820.03570.07180.04660.0498
Soybean-large 0.0783 0.11760.09820.08420.13730.09210.0947
Ionosphere 0.0242 0.04010.05810.03850.05820.03440.0497
Dermatology0.02160.05130.0684 0.0199 0.06850.07460.0648
House-votes-84 0.0066 0.01700.01970.00940.01790.01640.0168
Chess 0.0401 0.04860.05310.04150.06260.04230.0447
Soybean 0.0302 0.06540.04390.03260.06060.05090.0406
Breast-cancer-w 0.0010 0.03370.05040.01340.01010.01990.0425
Tic-tac-toe 0.0455 0.08240.11250.05130.05900.08130.0951
Vowel0.25420.24450.23250.2344 0.1093 0.23370.2255
Car0.0520 0.0376 0.04340.04380.04560.04470.0379
Mfeat-mor 0.0622 0.10200.10310.06770.13510.08820.0960
Segment0.02590.02940.03810.0255 0.0191 0.02910.0344
Hypothyroid0.00310.0034 0.0024 0.00340.02790.0034 0.0024
Kr-vs-kp0.01860.01520.01110.01860.0077 0.0076 0.0077
Dis0.00690.00050.00110.00710.00210.0005 0.0003
Hypo0.00510.00710.00690.0049 0.0046 0.00780.0060
Sick0.00470.00510.00430.00420.00820.0052 0.0035
Abalone 0.0682 0.16930.17690.15440.18650.15110.1633
Waveform-5000 0.0259 0.06900.08430.04100.05280.06250.0666
Phoneme0.12150.18280.10640.1343 0.0818 0.18500.1052
Wall-following0.02110.02880.02940.0242 0.0112 0.02660.0278
Page-blocks0.01350.01430.01770.0124 0.0110 0.01150.0146
Satellite 0.0139 0.03670.04550.03630.02510.03880.0406
Thyroid 0.0205 0.02570.02720.02350.02790.02200.0235
Pendigits0.01570.02000.0236 0.0127 0.01480.01570.0198
Sign 0.0313 0.03860.05960.03780.05930.05720.0488
Nursery 0.0085 0.01680.01950.01050.01930.01790.0168
Magic 0.0174 0.04900.04910.02970.05120.04070.0440
Letter-recog0.04710.05910.07090.04480.0492 0.0440 0.0619
Adult 0.0069 0.01650.02850.01160.04250.01410.0185
Shuttle0.00090.0004 0.0003 0.00040.00040.0004 0.0003
Connect-40.0156 0.0149 0.03090.02220.05340.02150.0222
Waveform 0.0009 0.00530.00370.00250.00680.00210.0035
Localization 0.0460 0.05940.10990.05800.11060.08970.0955
Table 4

W/D/L comparison results of bias on all datasets.

NBTANKDBAODEkdbe
TAN30/5/7----
KDB30/5/725/9/8---
AODE32/7/318/14/1015/4/23--
kdbe31/3/820/10/1215/8/1916/11/15-
KDBe32/3/726/9/711/27/421/13/817/18/7
Table 5

W/D/L comparison results of variance on all datasets.

NBTANKDBAODEkdbe
TAN4/3/35----
KDB8/1/339/7/26---
AODE9/8/2530/8/434/3/5--
kdbe7/1/3419/13/1030/4/86/10/26-
KDBe8/2/3216/12/1434/8/07/4/3112/9/21
In terms of variance, since the network structures of NB and AODE are definite and irrelevant to the variation of the training data, the independence assumption helped reduce the variance significantly. KDB was the most sensitive to the variation in training data among all classifiers. As discussed in Section 3, discriminatory target learning made kdb underfit training data and overfit the unlabeled instance. When kdb was integrated with regular KDB, discriminatory target learning helped to reduce the variance and the final ensemble classifier, i.e., KDB, performed the best only after NB and AODE.

4.2.3. Time Comparison

We compared KDB with the other classic BNCs in terms of training and classification time. Since kdb is a part of KDB, we removed it in this experiment. Figure 6a,b shows the training and classification time comparisons for all BNCs. Each bar represents the sum of time on 42 datasets in a 10-fold cross-validation experiment. No parallelization techniques were used in any case. As discussed in Section 3, discriminatory target learning requires no additional training time, thus the training time complexity of KDB was the same as that of regular KDB. Due to the structure complexity, KDB and KDB required a bit more time for training than the other BNCs. With respect to classification time, KDB took a little more time than the other BNCs. The reason lies in that KDB learned kdb for each unlabeled test instance, while the other BNCs only needed to directly calculate the joint probabilities. In general, discriminatory target learning helped to significantly improve the classification performance of its base classifier at the cost of a small increase in time consumption, which is perfectly acceptable.
Figure 6

Training and classification time comparisons for BNCs.

4.3. KDBe Versus Random Forest

To further illustrate the performance of our proposed discriminatory target learning framework, we compared KDB with a powerful learner, i.e., Random forest.Random forest (RF) is a combination of decision tree predictors, where each tree is trained on data selected at random but with replacement from the original data [10]. As the number of trees in the forest becomes large, the classification error for forests tends to converge to a limit. RF is an effective tool in prediction. RF can process high-dimensional data (that is, data with a lot of features) without making feature selection. Furthermore, due to the random mechanism, RF has the capacity to deal with imbalanced datasets or data with numerous missing values. Moreover, the framework in terms of strength of the individual predictors and their correlations gives insight into the ability of the RF to predict [10]. Because of its high classification accuracy, RF has been applied to many scientific fields, e.g., ecology and agriculture [32]. In our experiment, RF with 100 decision trees was used. The detailed results of RF in terms of zero-one loss, RMSE, bias and variance can be found in Table A1, Table A2, Table A3 and Table A4 in Appendix A, respectively. Table 6 shows the W/D/L records with different dataset sizes. When zero-one loss was compared, KDB won more frequently than RF, especially on small and medium datasets. The results indicate 10/4/3 on small datasets and 7/4/4 on medium datasets. The reason may lie in that 100 decision trees are complex and tend to overfit the training data. RMSE-wise, KDB also performed better than RF, which is shown as 16 wins and 11 losses. Bias and variance comparison of KDB and RF (Table 6) suggested that KDB is a low variance and high bias classifier. One can expect it to work extremely well on small and medium datasets. This is evident in Table 6 showing the zero-one loss and RMSE comparisons. KDB beat RF on 26 datasets and lost on 12 datasets with respect to variance. Thus, the advantages of KDB over RF in terms of zero-one loss and RMSE could be attributed to the change in variance. Since the variance term increased as the algorithm became more sensitive to the change in labeled training data, obviously, discriminatory target learning helped to alleviate the negative effect caused by overfitting.
Table 6

W/D/L records between KDB and RF.

AllSmallMediumLarge
Zero-one loss20/10/1210/4/37/4/43/2/5
RMSE16/15/114/9/48/4/34/2/4
Bias11/11/205/1/115/5/51/5/4
Variance26/4/1211/3/37/1/78/0/2
Besides, we display the time comparisons between KDB and RF in Figure 7. It is obvious that KDB enjoyed a great advantage over RF in terms of training time on datasets of all sizes. This advantage could be attributed to that KDB only learned a regular KDB for every dataset during the training phase while RF needed to train 100 decision trees. When comparing classification time, the performance of KDB and RF showed a slight reversal. Learning kdb for each unlabeled test instance made KDB take a bit more time than RF. However, when comparing on small and medium datasets, the advantage of RF over KDB was not significant. To conclude, on small and medium datasets, KDB had a significantly better zero-one loss performance and better RMSE than RF. This was packaged with KDB’s far superior training times and competitive classification times over RF, which makes KDB an excellent alternative to RF, especially for dealing with small and medium datasets.
Figure 7

Training and classification time comparisons between KDB and RF.

4.3.1. Discussion

RF has been applied to several scientific fields and associated research areas [32], because of its high classification accuracy. However, RF is more negatively affected in terms of computation consumption (memory and time) by dataset sizes than BNCs [19]. Furthermore, due to the random mechanism, RF is sometimes criticized for difficulty giving a clear semantic explanation of the combined result that is outputted by numerous decision trees. In contrast, our proposed discriminatory target learning framework considers not only the dependence relationships that exist in the training data, but also that hidden in unlabeled test instances, which makes the final model highly interpretable. KDB outperformed RF in terms of zero-one loss, RMSE and variance, especially on small and medium size datasets, while RF beat KDB in terms of bias. Moreover, RF required substantially more time for training and KDB took a bit more time for classifying. To illustrate the better interpretability of KDB than that of RF, we took medical diagnostic application as an example. The Heart-disease-c dataset (http://archive.ics.uci.edu/ml/datasets/Heart+Disease) from UCI Machine Learning Repository was collected from Cleveland Clinic Foundation, containing 13 attributes and two class labels. The detailed description of this dataset is shown in Table 7. The zero-one loss results of KDB, RF and KDB are 0.2244, 0.2212 and 0.2079, respectively. KDB learned from training data can describe the general conditional dependencies, while for a certain instance some of dependence relationships may hold instead of all the dependencies shown in KDB. In contrast, kdb can encode the most possible local conditional dependencies hidden in one single test instance. We argue that an ideal phenomenon is that KDB and kdb are complementary to each other for classification and they may focus on different key points. To prove this, randomly taking an instance from Heart-disease-c dataset as an example, the detail of this instance is shown as, . Figure 8 and Figure 9 show the structural difference between KDB and the submodels of kdb. For KDB, by comparing mutual information , are the first three key attributes for this dataset. There are 23 arcs in the structure of KDB which represent the conditional dependencies between predictive attributes. However, the values of , , and are all 0. For the instance , in Figure 9, we can easily find that the structure of kdb differed greatly from that of KDB. The true class label for is . KDB misclassified , while KDB correctly classified the instance. Thus, we can use Figure 9a for dependence analysis. By comparing the pointwise -mutual information, are the first three key attribute values for . It is worth mentioning that ranked second in KDB, whereas ranked last in kdb. Furthermore, there were only 15 arcs in kdb, which means that some redundant dependencies were eliminated. In general, KDB could utilize the knowledge learned from the training data and unlabeled test instances by building different models, which is obviously suitable for precision medical diagnosis.
Table 7

Description of Heart-disease-c dataset.

AttributeDescriptionSymbol
agereal value X0
sexmale or female, {0,1} X1
cpchest pain type (angina, abnang, notang, asympt), {1,2,3,4} X2
trestbpsresting blood pressure, real value X3
cholcholesterol, real value X4
fbsfasting blood sugar < 120 (true or false), {0,1} X5
restecgresting electrocardiographic results (norm, abn, hyper), {0,1,2} X6
thalachmaximum heart rate achieved, real value X7
exangexercise induced angina (true or false), {0,1} X8
oldpeakST depression induced by exercise relative to rest, real value X9
slopethe slope of the peak exercise ST segment (up, flat, down), {1,2,3} X10
canumber of vessels colored, real value X11
thalthal (norm, fixed, rever), {3,6,7} X12
class0 for health, 1 for sick Y
Figure 8

The structure of KDB on Heart-disease-c dataset.

Figure 9

The structure of submodels of kdb.

4.3.2. Imbalanced Datasets

There are 15 imbalanced datasets in our experiments, which are annotated with the symbol “*” in Table 1. To prove that KDB has the capacity to deal with imbalanced datasets, we conducted a set of experiments to compare the performance of KDB with RF in terms of extended Matthews correlation coefficient (MCC). The MCC provides a balanced measure for skewed datasets by taking into account the class distribution [33]. The classification results can be shown in the form of a confusion matrix as follows: Each entry of the matrix gives the number of instances, whose true class was that were actually assigned to , where . Each entry of the matrix gives the number of instances, whose true class was that were actually assigned to , where and . Given the confusion matrix, the extended MCC can be calculated as follow, Note that the MCC reaches its best value at 1, which represents a perfect prediction, and worst value at −1, which indicates a total disagreement between the predicted and observed classifications. Figure 10 shows the scatter plot of KDB and RF in terms of MCC. We can see that many points fall close to the diagonal line, which means that KDB achieved competitive results compared with RF. Furthermore, there are three points far above the diagonal line, which means KDB enjoys significant advantages on these datasets. A notable case is Dis dataset annotated with red color, where the MCC of KDB (0.4714) was much higher than that of RF (0.3710). In general, KDB had the capacity to handle the imbalanced datasets.
Figure 10

The scatter plot of KDB and RF in terms of MCC. Dis dataset is annotated with red color, which is a notable case where KDB enjoys significant advantages.

4.4. Global Comparison of All Classifiers

In this section, to assess whether the overall differences in performance of these learners was statistically significant, we employed the Friedman test [34] and the post-hoc Nemenyi test, as recommended by Demšar [35]. The Friedman test is a non-parametric test for multiple hypotheses testing. It ranks the algorithms for each dataset separately: the best performing algorithm getting the rank of 1, the second best ranking 2, and so on. In case of ties, average ranks are assigned. The null-hypothesis is that all of the algorithms perform almost equivalently and there is no significant difference in terms of average ranks. The Friedman statistic can be computed as follows: where and is the rank of the jth of t algorithms on the ith of N datasets. The Friedman statistic is distributed according to with degrees of freedom. Thus, for any pre-determined level of significance , the null hypothesis will be rejected if . The critical value of for with six degrees of freedom is 12.592. The Friedman statistics of zero-one loss and RMSE were 53.65 and 60.49, which were both larger than 12.592. Hence, the null-hypotheses was rejected. According to the detailed results of rank shown in Table A5 and Table A6 in Appendix A, Figure 11 plots the average ranks across all datasets, along with the standard deviation for each learner. When assessing the calibration of the probability estimates using zero-one loss, KDB obtained the lowest average rank of 2.5952, followed by kdb with 3.5595 and RF with 3.7024 (very close to those for AODE). When assessing performance using RMSE, KDB still performed the best, followed by RF with 3.4285 and AODE with 3.7500. We found NB at the other extreme on both measures, with average ranks 5.8690 and 5.9523 out of a total of seven learners.
Table A5

Ranks in terms of zero-one loss of different learners.

DatasetNBTANKDBAODERFkdbeKDBe
Contact-lenses5.05.0 1.0 7.03.05.02.0
Labor 2.5 5.0 2.5 6.07.0 2.5 2.5
Echocardiogram5.04.06.03.07.0 1.5 1.5
Lymphography 1.0 3.57.02.06.03.55.0
Sonar4.52.06.53.0 1.0 6.54.5
Glass-id7.03.53.56.0 1.0 5.02.0
New-thyroid2.05.06.0 1.0 7.03.53.5
Heart-disease-c 1.0 4.57.03.06.02.04.5
Soybean-large6.04.53.0 1.0 4.57.02.0
Ionosphere7.02.03.04.05.06.0 1.0
Dermatology2.03.07.0 1.0 4.06.05.0
House-votes-847.05.03.04.0 1.0 6.02.0
Chess7.02.04.05.06.0 2.0 2.0
Soybean7.0 1.0 5.02.06.04.03.0
Breast-cancer-w 1.0 5.07.02.03.04.06.0
Tic-tac-toe7.05.03.06.04.0 1.0 2.0
Vowel7.0 1.0 6.02.04.05.03.0
Car7.03.0 1.0 6.05.04.02.0
Mfeat-mor6.0 1.0 5.07.02.03.04.0
Segment7.03.06.0 1.0 4.02.05.0
Hypothyroid7.03.04.06.05.0 1.0 2.0
Kr-vs-kp7.05.03.06.0 1.0 4.02.0
Dis6.56.55.03.04.02.0 1.0
Hypo6.07.04.0 1.0 5.02.52.5
Sick7.03.0 1.0 6.04.05.02.0
Abalone6.05.04.02.07.03.0 1.0
Waveform-50007.05.06.0 1.0 2.04.03.0
Phoneme5.06.03.04.0 1.0 7.02.0
Wall-following7.06.05.02.0 1.0 4.03.0
Page-blocks7.06.05.04.0 1.0 3.02.0
Satellite7.05.02.04.03.06.0 1.0
Thyroid7.05.03.04.06.0 1.0 2.0
Pendigits7.05.04.0 1.5 6.0 1.5 3.0
Sign7.05.03.06.0 1.0 4.02.0
Nursery7.05.02.06.0 1.0 4.03.0
Magic7.04.02.06.03.05.0 1.0
Letter-recog7.06.05.03.04.0 1.0 2.0
Adult7.04.05.06.0 1.0 2.03.0
Shuttle7.06.05.04.0 1.0 2.03.0
Connect-47.05.03.06.0 1.0 4.02.0
Waveform5.04.06.0 1.0 7.03.02.0
Localization7.05.03.06.04.0 1.0 2.0
Sum of ranks246.5179.5175.5160.5155.5149.5 109.0
Table A6

Ranks in terms of RMSE of different learners.

DatasetNBTANKDBAODERFkdbeKDBe
Contact-lenses2.07.0 1.0 4.06.05.03.0
Labor 1.0 4.02.03.06.07.05.0
Echocardiogram5.04.07.06.0 1.0 2.03.0
Lymphography 1.0 4.07.02.05.06.03.0
Sonar7.05.04.06.0 1.0 3.02.0
Glass-id7.04.05.06.0 1.0 3.02.0
New-thyroid 1.0 6.07.03.02.04.05.0
Heart-disease-c5.06.07.02.04.0 1.0 3.0
Soybean-large5.04.03.02.07.06.0 1.0
Ionosphere7.04.05.02.0 1.0 6.03.0
Dermatology 1.0 3.04.02.05.07.06.0
House-votes-847.06.04.05.0 1.0 3.02.0
Chess7.0 1.0 2.04.05.06.03.0
Soybean7.0 1.0 4.03.06.05.02.0
Breast-cancer-w 1.0 4.07.03.02.06.05.0
Tic-tac-toe7.06.03.05.0 1.0 4.02.0
Vowel7.0 1.0 5.02.04.06.03.0
Car7.03.0 1.0 6.05.04.02.0
Mfeat-mor7.0 1.0 5.04.06.02.03.0
Segment7.04.05.0 1.0 6.03.02.0
Hypothyroid7.04.03.06.0 1.0 5.02.0
Kr-vs-kp7.04.02.06.0 1.0 5.03.0
Dis7.06.03.05.0 1.0 4.02.0
Hypo7.06.03.02.04.05.0 1.0
Sick7.03.02.06.04.05.0 1.0
Abalone7.04.05.0 1.0 6.02.52.5
Waveform-50007.03.06.0 1.0 5.04.02.0
Phoneme4.06.03.05.0 1.0 7.02.0
Wall-following7.06.05.03.0 1.0 4.02.0
Page-blocks7.06.05.04.02.0 1.0 3.0
Satellite7.05.03.04.02.06.0 1.0
Thyroid7.04.05.03.06.0 1.0 2.0
Pendigits7.04.03.0 1.0 6.05.02.0
Sign7.05.03.06.0 1.0 4.02.0
Nursery7.05.02.06.0 1.0 4.03.0
Magic7.02.03.05.06.04.0 1.0
Letter-recog7.05.04.02.06.03.0 1.0
Adult7.03.04.05.06.02.0 1.0
Shuttle7.06.05.03.04.0 1.0 2.0
Connect-47.04.03.05.0 1.0 6.02.0
Waveform7.04.06.02.0 1.0 5.03.0
Localization7.05.54.05.53.0 1.0 2.0
Sum of ranks250.0178.5170.0157.5144.0173.5 102.5
Figure 11

Average ranks in terms of zero-one loss and RMSE for all learners.

Since we rejected the null-hypotheses, Nemenyi test was used to further analyze which pairs of algorithms were significantly different in terms of average ranks of the Friedman test. The performance of two classifiers is significantly different if their corresponding average ranks of the Friedman test differ by at least the critical difference (): where the critical value for and is 2.949. Given seven algorithms and 42 datasets, we used Equation (16) to calculate and the result is 1.3902. The learners in Figure 12 are plotted on the red line on the basis of their average ranks, corresponding to the nodes on the top black line. If two algorithms had no significant difference, they were connected by a line. As shown in Figure 12a, we easily found that KDB had a significantly lower average zero-one loss rank than NB, TAN and KDB. KDB also achieved lower average zero-one loss rank than kdb, RF and AODE, but not significantly so. When RMSE was considered, KDB still performed the best and the rank of KDB was significantly lower than that of KDB, providing solid evidence for the effectiveness of our proposed discriminatory target learning framework.
Figure 12

Nemenyi test in terms of zero-one loss and RMSE for all learners.

5. Conclusions

Lack of explanatory insight into the relative influence of the random variables greatly restricts the application domain of machine learning techniques. By redefining mutual information and conditional information, the framework of discriminatory target learning can help fully and discriminately describe the dependency relationships in unlabeled instance and labeled training data. The kdb learned from unlabeled instance and regular KDB learned from training data are different but complementary in nature, which will help further improve the classification performance. Discriminatory target learning can be expected to play for different types of BNCs with different dependency complexities. Exploration of application of discriminatory target learning in other kinds of machine learning techniques, e.g., decision tree or support vector machine, is a further area for future work.
  3 in total

1.  [Spectral classification based on Bayes decision].

Authors:  Rong Liu; Hong-Mei Jin; Fu-Qing Duan
Journal:  Guang Pu Xue Yu Guang Pu Fen Xi       Date:  2010-03       Impact factor: 0.589

2.  Comparing two K-category assignments by a K-category correlation coefficient.

Authors:  J Gorodkin
Journal:  Comput Biol Chem       Date:  2004-12       Impact factor: 2.877

3.  Identification of differentially expressed molecular functions associated with breast cancer using Gibbs sampling.

Authors:  Gang Zhou; Ming-Qian Lu; Dao-Jun Li; Bao-An Gao; Rong Guo
Journal:  Oncol Lett       Date:  2017-10-10       Impact factor: 2.967

  3 in total
  1 in total

1.  Information Theoretic Measures and Their Applications.

Authors:  Osvaldo A Rosso; Fernando Montani
Journal:  Entropy (Basel)       Date:  2020-12-07       Impact factor: 2.524

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.