Literature DB >> 35573261

Multi-Source Selection Transfer Learning with Privacy-Preserving.

Weifei Wu1.   

Abstract

Transfer learning has ability to create learning task of weakly labeled or unlabeled target domain by using knowledge of source domain to help, which can effectively improve the performance of target learning task. At present, the increased awareness of privacy protection restricts access to data sources and poses new challenges to the development of transfer learning. However, the research on privacy protection in transfer learning is very rare. The existing work mainly uses differential privacy technology and does not consider the distribution difference between data sources, or does not consider the conditional probability distribution of data, which causes negative transfer to harm the effect of algorithm. Therefore, this paper proposes multi-source selection transfer learning algorithm with privacy-preserving MultiSTLP, which is used in scenarios where target domain contains unlabeled data sets with only a small amount of group probability information and multiple source domains with a large number of labeled data sets. Group probability means that the class label of each sample in target data set is unknown, but the probability of each class in a given data group is available, and multiple source domains indicate that there are more than two source domains. The number of data set contains more than two data sets of source domain and one data set of target domain. The algorithm adapts to the marginal probability distribution and conditional probability distribution differences between domains, and can protect the privacy of target data and improve classification accuracy by fusing the idea of multi-source transfer learning and group probability into support vector machine. At the same time, it can select the representative dataset in source domains to improve efficiency relied on speeding up the training process of algorithm. Experimental results on several real datasets show the effectiveness of MultiSTLP, and it also has some advantages compared with the state-of-the-art transfer learning algorithm.
© The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022.

Entities:  

Keywords:  Group probabilities; Multi-source transfer learning; Privacy-preserving

Year:  2022        PMID: 35573261      PMCID: PMC9077647          DOI: 10.1007/s11063-022-10841-6

Source DB:  PubMed          Journal:  Neural Process Lett        ISSN: 1370-4621            Impact factor:   2.565


Introduction

Machine learning has progressed dramatically over the past two decades, from laboratory curiosity to a practical technology in widespread commercial use [1]. A prominent aspect of machine learning is the ability to deal with large amounts of unorganized information problems by learning models from labeled data in domain, so sufficiently available labeled data is the basis for reliable results from machine learning models. Currently, machine learning has been widely used in computer vision [2], intrusion detection [3], speech emotion recognition [4, 5], natural language processing [6] and text classification [7]. At present, in order to obtain better accuracy and reliability, the existing traditional machine learning models usually need to meet two basic assumptions: there are enough available data samples in training data set; the training and test data come from the same feature space and distribution [8]. However, in practical applications, training and test data usually come from different fields, and it is difficult to ensure that the data distribution is consistent. In addition, labeled data is scarce in some areas. When the distribution changes, the machine learning algorithm needs to re-collect and re-label training data. In many real-world applications, the cost of re-collecting training data and reconstructing model is very expensive, or even impossible [9]. In this case, transfer learning between learning task domains is desirable, the motivation is that people can use the previously learned knowledge to better solve new problems, and the purpose is to use the label information of another related domain (source domain) for building the model of target domain [10-14]. Unlike traditional machine learning algorithms, them assume that training and test data have the same distribution, transfer learning can use knowledge from different distributions of data. In view of the advantages of transfer learning, a lot of research on it has been launched [15-18]. On the other hand, the smooth implementation of transfer learning usually requires source and target domains to directly share the original data, which cannot be met in some cases, especially when it involves some confidential or sensitive data. So, the protection of data privacy in transfer learning is becoming an important issue that people pay attention to. The research on privacy protection in transfer learning is very rare. The most recent related work is the differential privacy hypothesis transfer learning method for logistic regression proposed by Wang et al. [19], which uses the public unlabeled source data set to measure the relationship between source and target domain with the hypothesis trained on source domain to improve the learning of target hypothesis. Other related researches focus on variants of transfer learning, such as iterative differential privacy multi-task learning [20], distributed training data aggregation that considers covariate shift (covariate shift) [21], these works either did not consider the distribution differences between data sources, or did not consider the conditional distribution of data. Recently, a class of machine learning methods that use information in group probabilities to train classifier provides an effective way to protect data privacy, this is a type of semi-supervised learning method between supervised learning and unsupervised learning [21-24]. As shown in Fig. 1, for a set of training data without class labels, if you only know the probability of belonging to a certain class label in each data group, that is, the group probability, then use the group probability information to obtain a classifier that can effectively label the data. A typical application related to group probability is in political election events, where the number of voters in a constituency is known. In order to protect the privacy of each voter, only the number of candidates votes will be provided, and the specific ballot information for each voter is unknown. It can be seen that the group probability information provides an effective means for protecting the privacy of data.
Fig. 1

Leaning from group probability

Leaning from group probability In view of the advantages of group probability to protect the privacy of training data in data classification, using group probability to solve the privacy protection problem in transfer learning has attracted the attention of researchers [25, 26]. The existing algorithms only consider the knowledge in a source domain to target learning task, and the marginal probability difference. However, there is more than one source domain related to target domain. Therefore, it is natural that many transfer learning algorithms related to multiple source domains are proposed [27-29]. The multi-source transfer method extracts knowledge from data sets of two or more source domains for the learning task of target domain. Compared with the transfer learning method that only uses one source domain, it can increase the chance that transferring relevant knowledge from source domains to target domain and improve learning result. Today, transfer learning has been applied to COVID-19 recognition [30], law article prediction [31], the classification of histological images of colorectal cancer [32], human action recognition [33], cross-domain recommendations [34] and EEG signal analysis [35]. In this paper, we utilize the group probability and multi-source transfer learning theory, in the case of the application scenarios that the target domain has only group probability data with only a small amount of unlabeled data, and multiple source domains contain large amount of labeled data in each source domain, a new multi-source selection transfer learning algorithm with privacy-preserving (MultiSTLP) is proposed. The idea of MultiSTLP algorithm is to use the knowledge of group probability in target domain and the knowledge of labeled data in multiple source domains into the framework of support vector machine structure risk minimization, by constructing the similar distance term between target and each source domains. During the process, considering the marginal and conditional probability differences the knowledge of the representative dataset which is selected form source domains is transferred into target domain, and then the optimizable objective function is constructed. The theoretical proof of the objective function shows that the solution process is a quadratic programming problem with optimal solution. In the algorithm, the group probability protects the privacy of the target data, and the representative dataset of sources domain helps to reduce the size of training samples and improve the efficiency of algorithm training. Compared with the previous works, the contributions of this paper are: The rest of the paper is organized as follows. In Sect. 2, the related works of selective transfer learning support vector machine and group probability are briefly reviewed. The MultiSTLP is proposed in Sect. 3. In Sect. 4, we verify the effectiveness of MultiSTLP on four real-world datasets, and the experimental results are analyzed. The last section summarizes the conclusions of this paper and researches in the future. A new multi-source selection transfer learning algorithm with privacy-preserving MultiSTLP is proposed, which utilizes samples of the representative dataset that is selected from multiple sources and unlabeled group probability samples in target domain. MultiSTLP not only improves training efficiency, but also protects data privacy. The objective function of MultiSTLP can be transformed into a traditional standard quadratic programming problem and proved to have global optimal values through rigorous mathematical proof. By reducing the marginal and conditional probability differences, the knowledge of each source domain with similarity to target domain are transferred to the greatest extent, which effectively solves the negative transfer and improves the effect of algorithm. In addition, the representative dataset in source domains can make full use of high-quality samples, reducing the number of training samples and speeding up the algorithm training process. Extensive experiments have been carried out on real datasets, the experimental results show that the result of MultiSTLP is better than the state-of-art algorithms or at least comparable to them.

Brief Review of Related Works

We briefly introduce the selective transfer learning support vector machine and group probabilities in this section. In the group probability introduction, we focus on the IC technology and the IC-SVM algorithm. First of all, we start with the variable definitions of terminologies. For clarity, Table 1 lists the frequently used notations.
Table 1

Notations and descriptions

NotationDescription
D\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${}_S$$\end{document}S/D \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$_T$$\end{document}TSource/target domain
D\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${}_{S_{i}}$$\end{document}SiS\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$_i$$\end{document}i-th Source domain
X\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${}_S$$\end{document}S/X\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$_T$$\end{document}TSource/target sample set
Y\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${}_S$$\end{document}S/Y\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$_T$$\end{document}TSource/target class label set
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${n_{{S_{i}}}}'$$\end{document}nSiNumber of labeled \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${S_i}$$\end{document}Si-th source domain
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${M}$$\end{document}MNumber of source domain
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${n}$$\end{document}nNumber of target domain samples
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${w_s}/{b_s}$$\end{document}ws/bsParameters of source linear classifier
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${w_t}/{b_t}$$\end{document}wt/btParameters of target linear classifier
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${p_k}$$\end{document}pkGroup probability
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${G_k}$$\end{document}Gkk-th Groups
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${K}$$\end{document}KNumber of group
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\gamma $$\end{document}γWeight of source domains
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\beta ^{{S_{i}}}}$$\end{document}βSiWeight of representative data set in \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${S_i}$$\end{document}Si-th source domain
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${v}$$\end{document}vWeight of samples
Notations and descriptions

Selective Transfer Learning Support Vector Machine

For SVM, a lot of training samples is a prerequisite for achieving better training results. This not only requires a lot of manpower to label, but also a lot of time is consumed in the training phase, so the training efficiency of SVM is not satisfactory. In order to improve the efficiency of SVM training, a method of using training samples near the largest hyperplane to train SVM approximate the extreme point support vector machine (AESVM) was proposed in [36]. AESVM no longer needs all training samples to train the learning model. The training sample size can be greatly reduced, so that the training cost of the learning model is reduced. On this basis, Li et al. [10] proposed a selective transfer learning support vector machine algorithm (STL-SVM), which uses AESVM to select representative dataset from source domain. STL-SVM first utilizes an improved maximum mean discrepancy (MMD) to calculate the weight vector of the importance of the sample in source domain relative to target domain; then AESVM method is applied to select a representative dataset and the weight of samples; finally, combined with the support vector to construct an objective function with the ability the transfer learning. Given a source domain containing n sample data, , , , . Similarly, for a target domain with m samples, . The objective function of STL-SVM is shown in Eq. (1):In Eq. (1), and represent the parameters in target domain, and represent the parameters in source domain, these parameters include knowledge in domains; and represent the knowledge obtained by SVM training only on dataset in target domain; is non-linear mapping function; () and () are the slack variables in target and source domains, respectively; n is the number of samples in source domain, is the number of samples in representative dataset calculated by AESVM; m is the number of samples in target domain; represents the weight value corresponding to each sample in representative data set; () and () are the degree of penalty error of the regularization coefficient in target and source domain, respectively; T represents the transposition of the matrix; is the decision function of SVM classifier in target domain. Solve the Eq. (1) to obtain the model parameters and, substitute them into the decision function Eq. (2) of STL-SVM:On the one hand, STL-SVM reduces the size of training samples in source domain through AESVM and accelerates the learning progress; on the other hand, it uses MMD and objective function construction principles to effectively solve the negative transfer problem that is easy to occur in transfer learning. Therefore, STL-SVM completes the knowledge transfer by effectively fusing the knowledge of the source and target domains, so as to obtain a better classification effect. Experiments on artificial and real datasets show the effectiveness of STL-SVM. Compared with previous research work on transfer learning algorithms, the STL-SVM algorithm can better solve the negative transfer situation and the long training time of the classifier due to too many training samples in source domain. However, the problem with STL-SVM is that it only considers the marginal probability between domains and does not discuss the conditional probability; it only uses the knowledge in one source domain, when there are multiple source domains related to the target domain that will cause a waste of data resources. In addition, it does not have the ability to protect privacy of target data.

Group Probabilities

Given a dataset , is -th sample, N represents the number of samples, and the class label of samples is unknown. The group probability is defined as follows: Assuming that the dataset is divided into groups, , represents the total number of samples in each group, and the group probability of each group is known as , which represents the probability that the sample is positive class in the group. In each group, we know the probability that the sample is a positive class, however the class label of each sample is unknown. is called the group probability of dataset , which can effectively protect the privacy of dataset . For the purpose of solving the difficulty of applying the traditional classification model directly to group probability, IC-SVM [22] first labels the group-probability data based on platt model of the inverse calibration technique (IC), then uses these labeled data to train the SVM. IC-SVM utilizes the sigmod function as an estimated SVM posterior probability output method:In Eq. (3), the parameters A and B are obtained by the minimum cross entropy, x is the sample feature vector, y is a class label, indicates the probability that the sample is positive. Setting and , the Eq. (3) can be converted into the following Eq. (4):Further deformation is as follows:In practice, it is difficult to obtain the class label of each sample, so the average values of the class label estimated in each group is approximated as the predicted value of sample, as in Eq. (6):In Eq. (6), is number of group , and are the parameters in classification hyperplane of SVM, which sets up a bridge between the group probability information and SVM. The optimization problem of the IC combined with the SVM theory can be expressed as follows:In Eq. (7), K denotes the number of group, is defined minimum required precision of the estimate. Equation (7) uses SVM classifier to conveniently process group probability information, which also provides theoretical support for the proposed multi-source transfer learning algorithm.

Implementation of MultiSTLP

This section describes the multi-source transfer algorithm with group probability in detail. The algorithm framework is shown in Fig. 2. As shown in Fig. 2, the input information of MultiSTLP framework consists of two parts: label samples in target domains which contains unlabeled samples with only group probability information. For convenience, we only consider the binary classification problem (Fig. 3).
Fig. 2

Framework of MultiSTLP

Fig. 3

Flowchart of MultiSTLP

source domains are defined as: , denotes -th sample of -th source domain, the corresponding class label is , is the number of samples in source domain, means joint distribution. Analogously, target domain is , is the number of group and joint distribution is . As a class label, equals to that is estimating probability of class label. and , and are the marginal probability and conditional probability of source and target domains, respectively. Normally, and . MultiSTLP not only considers the differences between source and target domains, but also transferability of source domain by simultaneously reducing the differences between the marginal and conditional probabilities. Framework of MultiSTLP Flowchart of MultiSTLP

Select Representative Data Set and Adapt the Probability of Partial Difference

Refer to reference [10], using the AESVM method to calculate the representative data set in source domain and its corresponding weight vector , the number of samples in representative data set is. In order to effectively transfer knowledge from source domain that is similar to target domain, this paper adapts both marginal and conditional probability differences. we use MMD to calculate the weights of samples in source domain on the marginal probability difference. denotes that feature is mapped to a regenerative kernel Hilbert space , is the number of the representative data set in source domain , the number of group in target domain is , is also the dimension of . The minimization problem of Eq. (8) is a standard quadratic programming problem and can be solved using many existing solvers. When constructing the objective function MultiSTLP, by using samples in each source domain we add the corresponding weights as in Eq. (9).In Eq. (9), represents the sample vector after weighting samples of source domain . For the convenience of subsequent calculations, set , that is, the samples of source domain are weighted with . On the basis of the above calculation of marginal probability difference, calculate of source domain , which reflects similarity between source and target domains. First, we learn the classifier , this ensures that the classifier learned on source domain with similar marginal probability distributions. Then, using predicts unlabeled samples in target domain. denotes classifiers, is corresponding weight vector. Therefore, the goal of Eq. (10) is to find the optimal weight by minimizing the difference in prediction labels between two neighboring points in target domain. is the predicting result of -th sample using is a similarity parameter between two data samples of target domain. Equation (10) can be rewritten as the form of Eq. (11):In Eq. (11), is the graph Laplacian associated with the data of target domain, is the similarity matrix, is the diagonal matrix given by . The minimization problem of Eq. (11) is also a standard quadratic programming problem, which can be calculated using many existing solvers.

Construction of Object Function

On the basis of 3.1, we combine the structural risk minimization theory and the similarity distance minimization to construct the objective function of MultiSTLP as follows: is decision function vector of source domains, is decision function in target domain. and are the structure risk terms controlling the complexity of the classifier in the source domain and the target domain, respectively. indicates L2-norm. and are the regularization coefficients in source domain and target domain. is convex non-negative loss function. is used to quantify the diversity between source and target domains. is the trade-off parameter. Equation (12) consists of three items, the first term refers to the knowledge learning from source domains. The second term denotes the knowledge learning from target domain. The third term is that guarantees good generalization performance by minimizing the differences between each source and target domains. In further, is utilized to quantify the diversity between domains. So, Eq. (12) can be rewritten into Eq. (13).In Eq. (13), we chose two different hinge loss functions in source and target domains: and . Therefore, Eq. (13) can be formulated as an optimization problem:In Eq. (14), , and are slack variables; the first constraint guarantees that each source domain is classified as accurately as possible; the second and three constraints control estimating class probability of in target domain to approximate . is the estimated minimum precision of that satisfies the following function:According to [22], is set to be , is the group probability , is a very small positive constant.

Theorems Related to the Objective Function

Theorem 1

The dual problem of Eq. (14) is a QP problem as shown in Eq. (16). The proof of Theorem 1 can be seen in “Appendix 1”.

Theorem 2

The quadratic form of the optimization problem of Eq. (16) is a standard convex quadratic programming problem. The proof of Theorem 1 can be seen in “Appendix 2”. It is clear from the above results that the optimization problem in Eq. (16) for training can be transformed into a convex QP problem and can be directly solved by the traditional SVM solutions. Simultaneously, Eq. (16) is a convex quadratic programming problem, the KKT condition is also a sufficient condition, and thus the obtained solution is the optimal solution. According to the results obtained by Eq. (16), the results of the optimal solution are as follows:Finally, the decision function of MultiSTLP is expressed as follows:As we can see from Eqs. (17) and (18), the results contain both information of source and target domains: such as , is the knowledge that is learned from source domains, otherwise the knowledge from target domain is .

Training MultiSTLP

According to Sects. 3.1–3.3, the training process of MultiSTLP is now summarized and described in Table 2.
Table 2

MultiSTLP algorithm training

Steps of MultiSTLP Algorithm
Input: Labeled source domains \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${D_S} = \{ {D_{{S_i}}} = (x_j^{{S_i}},y_j^{{S_i}})_{j = 1}^{{n_{{s_i}}}},i = 1,\ldots ,M\} $$\end{document}DS={DSi=(xjSi,yjSi)j=1nsi,i=1,,M},
the number of sample in \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${D_{{S_i}}}$$\end{document}DSi is \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${n_{{S_i}}}$$\end{document}nSi.
An unlabeled training dataset \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${D_T} = {({x_i})_{i = 1,..,d}}$$\end{document}DT=(xi)i=1,..,d in the target domain,
the group probability is \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\{ ({G_k},{p_k})\} _{k = 1}^d$$\end{document}{(Gk,pk)}k=1d.
MultiSTLP algorithm training

Experimental Results

In this section, for the purpose of testing the generalization performance of MultiSTLP algorithm, we compare MultiSTLP with the benchmark algorithms on four real-world datasets 20-Newsgroups, TRECVID video detection, sentiment analysis, and Email spam. In the experiments, without loss of generality, we only consider the binary classification problem.

Experimental Environment and Evaluation Criteria

For the fairness of experiments, a 5-fold cross-validation strategy is selected, and we repeat the strategy twice as the final comparison results. In the experiments, we will run 10 times, the average value of classification accuracy, recall, precision, training time with their standard deviations are recorded. The representation of classification accuracy is as follows: represents datasets in the target domain, is true tag category, is the result of classifying using the learned classifier. The recall is expressed as follows:The precision is:TP represents the number of positive class samples that are accurately classified as positive classes by the classifier; FP is the number of negative class samples that are incorrectly classified as positive classes; and FN is the number that indicates that the positive class samples are incorrectly classified as negative classes. For each algorithm, a Gaussian kernel function is selected in the form . The parameters , and of the proposed MultiSTLP are determined by searching the grid . For baseline algorithms, the default parameter settings in their literatures are adopted in our experiments. The hardware setting of all experiments are as follows: Intel Core (TM), 3.6 GHz, 8 GB, Windows 10 operating system. The following state-of-the-art baseline algorithms are selected as the comparison algorithms for MultiSTLP. TrGNB [25] integrates the transfer learning and group probability information into naive Bayesian framework, the knowledge transfer is completed in the process of solving the naive Bayes by using the maximum posterior probability method. Compared with TrGNB, the advantages of our model are as follows: the marginal and conditional probability between source and target domains are considered; the knowledge in more than one source domains is transferred. IC-SVM [22] is based on the framework of traditional SVM classifier, combined with Inverse Calibration technology (IC) to construct an optimization function of classifier for class labels, which have no the ability of privacy protection and transfer knowledge compared with MultiSTLP. ARTL [17] learns the adaptive classifier by simultaneously optimizing the structural risk function, the joint distribution matching between domains and the manifold consistency behind the marginal distribution. The differences from the proposed method is that only one source domain can be used. STL-SVM [10] compared with the proposed method, which have no the ability of privacy protection and transfer knowledge of multi-source. TSVM-GP [26] integrates transfer term and group probability information into a support vector machine (SVM) to improve the classification accuracy. This method considers only a single source domain and marginal probability compared to the proposed method. SVM [35] is traditional support vector model, which has no ability to learn across domains.

Datasets

20-Newsgroups [8], TRECVID 2005 [24], Sentiment analysis [20], and Email spam [21] are commonly used in transfer learning applications, so all experiments in this paper are performed on these datasets. The 20-Newsgroups dataset are divided into 4 top categories: comp, rec, sci and talk, which contains about 20,000 documents, and each top category can also be divided into four sub-categories with detailed information as shown in Table 3. According to the construction of task group in [8, 17], two top categories are randomly selected from top categories, one of which is positive class and the other is negative class. Each task group is specifically: comp vs rec, comp vs sci, comp vs talk, rec vs sci, rec vs talk, and sci vs talk.TRCVID 2005 contains approximately 86 h of video programs and consists of 74,523 video shots. Each shot is represented by a video frame as a keyframe, and each keyframe is depicted by a 273-dimensional feature vector. All shots are manually labeled with 39 semantic categories. These semantics cover a variety of types, including outdoor scenes, indoor scenes, news types, and generally common objects. TRCVID video are from CNN_ENG, NBC_ENG, MSNBC_ENG, CCTV_CHN, NTDTV_CHN and LBC_ARB 6 channels, 13 news programs. Each channel represents a domain, except LBC containing 3 news programs, the other channels contain 2 news programs. The source domains datasets are selected from 3 English channels and 2 Chinese channels, and the target domain dataset is selected from the Arabic dataset. The sentiment analysis dataset contains four different comments of Amazon products: books, DVDs, electronics, and kitchen, which represent four domains Books (B), DVDs (D), Electronics (E), and Kitchen (K). Each comment contains product name, comment title, date, location, and comment content. We will evaluate the product with a rating of 3 stars (0–5 stars) or more as a positive example, a product with a rating of less 3 stars as a negative example, and discard if a fuzzy evaluation is found. In the every domain, there are 2000 labeled instances, and about 4000 unlabeled instances, where the number of positive and negative instances is substantially the same. The dataset details are shown in Table 4.
Table 3

The statistics of 20-Newsgroups

DomainsCommentsTrainingTestingPositive (%)Feature
Books (B)6465200044655030,000
DVDs (D)5586200035865030,000
Electronics (E)7681200056815030,000
Kitchen (K)7945200059455030,000
Table 4

The statistics of sentiment analysis dataset

DomainsCommentsTrainingTestingPositive (%)Feature
Books (B)6465200044655030,000
DVDs (D)5586200035865030,000
Electronics (E)7681200056815030,000
Kitchen (K)7945200059455030,000
20-Newsgroups The statistics of 20-Newsgroups TRECVID 2005 Sentiment analysis The statistics of sentiment analysis dataset Email spam Email spam dataset was released by ECML/PKDD 2006,see Table 5 for details. It contains a set of 4000 publicly available labeled emails (U4) as well as three email sets annotated by three different users . Therefore, the data distributions of the three user-annotated email sets and the publicly available email set are different from each other, in which one half of the emails are non-spam and the other half are spam. Since the spam and non-spam in four email subsets have been differentiated, their distributions are relevant but different.
Table 5

The statistics of spam dataset

DomainNumberPositiveNegativeFeature
U1400020002000206,908
U2250012501250206,908
U3250012501250206,908
U4250012501250206,908
The statistics of spam dataset

Analysis of Experimental Results

In this section, the experimental results of MultiSTLP and six benchmark algorithms on real datasets are analyzed and compared. TRECVID 2005 dataset We utilize two Chinese channels CCTV_CHN, three English channels CNN_ENG, NBC_ENG, MSNBC_ENG and NTDTV_CHN as the source domains, and LBC_ARB as the target domain. The details are shown in Table 6. The four transfer learning methods in the benchmarks can only use one source domain, so in the experiment, one of these source domains is randomly selected as the training dataset, and the MultiSTLP algorithm uses the datasets in all source domains simultaneously.
Table 6

Description of source and target domains on TRECVID dataset

DomainSource domainsTarget domain
ChannelCNN_ENGMSNBS_ENGNBC_ENGCCTV_CHNNTDTV_CHNLBC_ARB
# Keyframes11,0258905932210,896648115,272
Description of source and target domains on TRECVID dataset 20-Newsgroups dataset: in the experiment, we constructed 3 source domains and one target domain. The details are shown in Table 7. For the four single source domain transfer learning algorithms TrGNB, ARTL, STL-SVM and TSVM-GP randomly select one of the source domains for training, and MultiSTLP can simultaneously use the datasets of the three source domains for training.
Table 7

Description of source and target domains on 20-Newsgroups dataset

DomainSource domainsTarget domain
rec vs sci(r vs s)rec.autos & sci.cryptrec.sport.hockey & sci.space
rec.motorcycles & sci.electronics
rec.sport.baseball & sci.med
com vs sci(c vs s)comp.graphics & rec.autoscomp.sys.mac.hardware & rec.sport.hockey
comp.os.ms-windows.misc & rec.motorcycles
comp.sys.ibm.pc.hardware & rec.sport.baseball
sci vs com(s vs c)sci.crypt & comp.graphicssci.space & comp.sys.mac.hardware
sci.electronics & comp.os.ms-windows.misc
sci.med & comp.sys.ibm.pc.hardware
Description of source and target domains on 20-Newsgroups dataset Sentiment analysis dataset: in this dataset, we use Books, DVDs, and Electronics to construct three source domains, and Kitchen as the target domain. The details of source and target domains are shown in Table 8.
Table 8

Description of source and target domains on Sentiment analysis datase

DomainSource domainsTarget domain
Sentiment datasetBooksDVDsElectronics (E)Kitchen (K)
# Sentiment6465558676817945
Description of source and target domains on Sentiment analysis datase Email Spam dataset: for the MultiSTLP algorithm, three personal email datasets are used as three source domains, and the public email dataset is used as target domain; the other four single source domain transfer learning algorithms randomly select one of the three personal email. The detailed information is shown in Table 9.
Table 9

Description of source and target domains on email spam dataset

DomainSource domainsTarget domain
Emails datasetU1U2U3U4
#emails2500250025002500
Description of source and target domains on email spam dataset Comparison of average classification accuracy with standard deviation on real-world four transfer datasets Bold represents the results of algorithm proposed in this paper Comparison of average recall with standard deviation on four real-world transfer datasets Bold represents the results of algorithm proposed in this paper Comparison of average precision with standard deviation on four real-world transfer datasets Bold represents the results of algorithm proposed in this paper Comparison of average training time (s) with standard deviation on four real-world transfer datasets Bold represents the results of algorithm proposed in this paper Tables 10, 11, 12 and 13 show the average classification accuracies, average recall and average precision with their standard deviations of all the benchmarking classifiers on different transfer learning tasks. From these results, we can draw the following conclusions: In addition, in the transfer learning algorithms TrGNB, ARTL, STL-SVM and TSVM-GP only transfer the knowledge in one source domain, and only consider the marginal probability difference between data between domains, without considering the conditional probability difference. On the one hand, this resulted in insufficient knowledge to be transferred. On the other hand, the large difference between the transferred knowledge and the data in target domain resulted in a negative transfer phenomenon, which harmed the learning effect.
Table 10

Comparison of average classification accuracy with standard deviation on real-world four transfer datasets

DatasetsSVMIC-SVMTrGNBARTLSTL-SVMTSVM-GPMultiSTLP
20-Newsgroupsr vs s80.14(2.81)81.66(2.82)91.45(1.32)86.35(1.55)88.87(1.47)90.67(1.56)94.12(1.23)
c vs s74.24(1.98)79.24(2.01)92.28(1.11)82.65(1.32)81.86(1.36)88.82(1.45)95.32(1.11)
s vs c75.35(2.48)77.73(2.25)90.11(1.56)85.38(1.48)86.15(1.55)87.97(1.58)93.22(1.01)
Average76.58(2.42)79.54(2.36)91.48(1.33)84.79(1.45)85.63(1.46)89.15(1.53)92.45(1.30)
TRECVID 2005CN vs L79.16(2.87)80.29(2.65)91.53(1.98)86.91(2.17)87.32(2.51)93.15(1.69)94.32(1.52)
MS vs L77.87(2.13)78.26(2.27)89.35(1.89)83.88(2.12)84.62(2.42)90.55(1.76)92.57(1.64)
NB vs L73.97(2.49)74.65(2.59)86.53(1.85)80.27(1.96)82.74(2.01)88.71(1.81)89.76(1.59)
CC vs L74.53(2.01)75.72(1.96)87.62(1.76)81.32(2.78)83.79(1.81)90.95(1.53)91.54(1.44)
NT vs L69.69(1.93)70.66(1.87)84.81(1.65)78.58(1.76)79.85(1.73)86.98(1.58)87.62(1.32)
Average75.04(2.29)75.91(2.27)87.97(1.83)82.19(2.16)83.66(2.10)90.27(1.68)91.16(1.50)
Sentiment analysisB vs K78.46(1.77)79.37(1.83)90.21(1.87)88.75(2.01)91.85(1.93)89.87(1.71)92.53(1.64)
D vs K75.18(1.69)77.25(1.72)86.37(1.75)84.66(2.46)87.95(1.98)84.54(1.68)88.65(1.55)
E vs K77.59(2.11)78.27(2.05)84.56(1.98)80.93(2.15)85.11(2.26)81.75(1.85)86.58(1.76)
Average77.08(1.86)72.30(1.87)87.05(1.87)84.78(2.21)88.30(2.06)85.39(1.75)89.25(1.65)

Bold represents the results of algorithm proposed in this paper

Table 11

Comparison of average recall with standard deviation on four real-world transfer datasets

DatasetsSVMIC-SVMTrGNBARTLSTL-SVMTSVM-GPMultiSTLP
20-Newsgroupsr vs s70.87(2.73)71.22(2.61)75.12(2.42)74.24(2.54)71.66(3.43)76.01(2.31)78.54(2.15)
c vs s62.52(3.32)63.68(3.26)70.43(2.65)69.12(2.87)68.12(3.54)71.11(2.45)73.22(2.02)
s vs c74.24(4.05)75.78(3.97)72.81(2.73)71.67(2.95)72.33(3.65)73.64(2.65)75.66(2.42)
Average69.21(3.37)70.23(3.28)72.79(2.60)71.68(2.79)70.70(3.54)73.57(2.47)75.81(2.20)
TRECVID 2005CN vs L60.34(3.65)61.32(3.32)71.33(2.85)66.15(3.25)70.26(3.17)71.45(2.73)72.86(2.66)
MS vs L70.27(2.64)71.12(2.34)75.45(2.42)74.23(2.53)73.52(2.92)75.46(2.42)76.65(2.36)
NB vs L75.68(3.86)75.97(3.75)73.57(2.53)70.57(2.76)72.62(2.58)73.73(2.51)74.53(2.32)
CC vs L73.53(3.19)74.15(3.04)78.23(2.35)74.15(3.05)75.36(2.44)77.28(2.47)79.87(2.23)
NT vs L78.65(2.48)79.26(2.35)81.89(2.12)79.85(2.24)80.14(2.23)82.57(2.11)83.75(2.02)
Average71.69(3.16)72.36(2.96)76.09(2.45)72.99(2.77)74.38(3.25)76.10(2.45)77.53(2.32)
Sentiment analysisB vs K62.45(3.45)63.98(3.23)73.23(2.43)71.65(2.66)72.49(2.79)72.98(2.47)74.78(2.32)
D vs K61.98(3.11)62.76(3.08)70.36(2.03)70.25(2.19)68.17(3.11)70.01(2.12)72.63(1.98)
E vs K50.76(3.85)51.52(3.63)68.65(2.42)67.37(2.76)64.56(2.92)67.95(2.53)70.21(2.25)
Average58.40(3.47)59.42(3.31)70.75(2.29)69.76(2.54)68.41(2.94)70.31(2.37)72.54(2.18)

Bold represents the results of algorithm proposed in this paper

Table 12

Comparison of average precision with standard deviation on four real-world transfer datasets

DatasetsSVMIC-SVMTrGNBARTLSTL-SVMTSVM-GPMultiSTLP
20-Newsgroupsr vs s76.87(2.71)77.18(2.59)82.66(2.12)80.12(2.33)81.61(2.43)81.96(2.11)83.76(2.01)
c vs s69.25(2.26)70.12(2.31)83.45(2.25)78.84(1.78)77.66(2.85)82.84(2.65)84.94(2.11)
s vs c72.64(3.23)73.76(3.35)82.46(2.32)77.75(2.05)80.51(2.74)80.14(2.13)83.76(2.02)
Average72.92(2.40)73.69(2.75)82.86(2.23)78.90(2.05)79.93(2.67)81.65(2.30)84.15(2.05)
TRECVID 2005CN vs L65.32(3.12)66.21(2.92)77.75(1.45)73.65(2.02)75.84(1.96)76.23(1.53)79.57(1.33)
MS vs L74.64(2.55)75.36(2.53)83.42(1.31)80.82(2.11)83.41(2.13)82.56(1.36)84.74(1.23)
NB vs L71.25(2.68)72.52(2.43)79.65(1.26)77.57(1.99)78.87(1.97)80.67(1.37)80.36(1.22)
CC vs L78.38(2.56)79.73(2.45)86.43(1.42)83.76(2.15)83.86(1.92)85.32(1.55)87.21(1.31)
NT vs L84.26(2.44)85.32(2.36)89.65(1.15)85.95(1.93)85.49(1.81)86.46(1.25)89.82(1.12)
Average74.77(2.67)75.83(2.54)83.38(1.32)80.35(2.04)81.49(1.96)82.25(1.41)84.34(1.24)
Sentiment analysisB vs K65.65(2.21)66.36(2.32)76.56(1.32)73.75(1.68)74.29(2.11)75.35(1.38)77.75(1.28)
D vs K64.52(2.38)65.55(2.23)72.45(1.28)70.72(2.11)69.25(1.87)71.84(1.42)73.54(1.22)
E vs K54.88(2.55)55.48(2.45)68.72(1.24)69.18(1.87)67.68(2.32)67.27(1.51)69.67(1.12)
Average61.68(2.38)62.46(2.33)72.58(1.28)71.22(1.89)70.41(2.10)71.49(1.47)73.65(1.25)

Bold represents the results of algorithm proposed in this paper

Table 13

Comparison of average training time (s) with standard deviation on four real-world transfer datasets

DatasetsSVMIC-SVMTrGNBARTLSTL-SVMTSVM-GPMultiSTLP
20-Newsgroupsr vs s1.22(0.14)0.04(0.07)3.29(1.13)8.75(1.32)9.25(1.28)3.26(1.14)3.21(1.16)
c vs s1.18(0.15)0.06(0.09)3.16(1.14)8.65(1.33)9.11(1.22)3.02(1.03)2.15(1.17)
s vs c1.46(0.15)0.08(0.11)3.38(1.12)8.84(1.34)9.36(1.25)3.47(1.13)3.32(1.18)
TRECVID 2005CN vs L24.67(1.22)1.16(0.82)86.57(1.35)97.46(1.98)100.57(2.11)86.58(1.43)85.75(1.42)
MS vs L20.86(1.19)1.12(0.78)82.84(1.33)93.25(1.89)94.23(1.96)80.47(1.39)76.74(1.23)
NB vs L21.35(1.21)1.13(0.81)84.45(1.33)94.33(1.92)96.54(1.98)82.36(1.41)81.27(1.19)
CC vs L23.47(1.25)1.14(0.81)85.86(1.34)95.45(1.97)98.76(2.08)84.54(1.42)82.53(1.22)
NT vs L19.43(1.17)1.09(0.77)81.53(1.32)91.86(1.88)92.52(1.94)79.55(1.38)79.28(1.17)
Sentiment analysisB vs K13.34(0.88)0.43(0.12)18.57(1.15)21.37(1.25)23.46(1.34)18.44(1.18)18.15(1.24)
D vs K13.16(0.92)0.41(0.13)18.36(1.13)20.87(1.28)21.75(1.32)18.27(1.16)17.86(1.23)
E vs K13.53(1.02)0.45(0.15)18.63(1.16)21.65(1.31)23.66(1.33)18.68(1.15)18.26(1.25)
Email spamU1vsU41.73(0.12)0.06(0.02)4.25(1.23)6.25(1.67)7.87(1.86)3.86(1.12)3.24(2.11)
U2vsU41.75(0.13)0.05(0.02)4.37(1.25)6.56(1.75)8.11(2.15)3.87(1.15)3.62(2.25)
U3vsU41.72(0.11)0.06(0.03)4.36(1.24)6.47(1.73)7.94(2.01)3.65(1.03)3.34(2.18)

Bold represents the results of algorithm proposed in this paper

In terms of the average classification accuracy, it can be seen from Table 10 that the transfer learning algorithms TrGNB, ARTL, STL-SVM, TSVM-GP and MultiSTLP have better classification results than the non-transfer learning algorithms SVM and IC-SVM. This is because only a small amount of data set with probability information in target domain is not enough to train a reliable learning model, and the transfer learning algorithms can use the knowledge in a large amount of labeled data in source domain to assist target domain to create classification task, so the trained model is better. The MultiSTLP algorithm proposed in this paper makes up for the above problems. It not only transfers the knowledge of multiple source domain, but also adapts the marginal probability and conditional probability, and the classification effect is also better. Therefore, the average accuracy of the MultiSTLP algorithm proposed in this article on the four data sets 20-Newsgroups, TRECVID 2005, sentiment analysis and email spam is better than the comparison algorithms, which are 92.45%, 91.16%, 89.25% and 95.05%, respectively. In summary, through the comparative analysis of experimental results, we can see that the algorithm proposed in this paper is effective and efficient. It also shows the rationality and effectiveness of the proposed algorithm. The average recall in Table 11 show that MultiSTLP has certain advantages compared with non-transfer learning algorithms (SVM, IC-SVM) and single source transfer learning algorithms (TrGNB, ARTL, STL-SVM and TSVM-GP) on all transfer learning tasks. The average precision in Table 12, it is can be seen that MultiSTLP is better than benchmark algorithms. On the four data sets 20-Newsgroups, TRECVID 2005, sentiment analysis and email spam, the average precisions are 84.15%, 84.34%, 73.65% and 92.97%, respectively. In terms of training time shown in Table 13, it can also be seen that MultiSTLP has obvious advantage over transfer learning algorithms TrGNB, ARTL, STL-SVM and TSVM-GP in training time, because of selecting representative data set from source domain. IC-SVM needs less training time than other six classifiers in our experiments. This is because its training data only contains the group probabilities constructed from the 5% randomly selected unlabeled target data which is much less than the size of the training data for the other classifiers. Even though, SVM is effective in training time, its classification accuracy is not prominent on transfer learning problems. Finally, to test the differences between MultiSTLP and benchmark algorithms with similar classification results, the Wilcoxon signed rank test is applied to these methods. According to the contents of Table 10, the average classification accuracy of all algorithms is shown. In Table 10, we can see the average classification accuracy of all algorithms on real datasets. The results of the Wilcoxon test on real-world datasets 20-Newsgroups, TRECVID 2005, Sentiment analysis and Email spam are discussed below. 20-Newsgroups: the classification accuracy of MultiSTLP is only 0.97% higher than TrGNB; therefore, when using MultiSTLP and TrGNB to classify three cross-domain tasks, each task is repeated 10 times, the values of and are and , respectively. For the bilateral test of , when , by querying the distribution table of the Wilcoxon signed rank test, . Because of , was accepted: there was no significant difference in the classification results between the two methods. TRECVID 2005: the classification accuracy of MultiSTLP increases 0.89% by comparison with TSVM-GP; therefore, when using MultiSTLP and TSVM-GP to classify 5 cross-domain tasks, each task is repeated 10 times, the values of and are and , respectively. For the bilateral test of , when , by querying the distribution table of the Wilcoxon signed rank test, . Because of , was accepted: there was no significant difference in the classification results between the two methods. Sentiment analysis The classification accuracy of MultiSTLP is only 0.95% higher than STL-SVM; therefore, when using MultiSTLP and STL-SVM to classify three cross-domain tasks, each task is repeated 10 times, the values of and are and , respectively. For the bilateral test of , when , by querying the distribution table of the Wilcoxon signed rank test, . Because of , was accepted: there was no significant difference in the classification results between the two methods. Email spam: compared with TrGNB and ARTL, the classification accuracy of MultiSTLP increases 0.83 and 0.36%. When MultiSTLP, TrGNB and ARTL are used to classify three cross-domain tasks, each task was repeated 10 times. For MultiSTLP and TrGNB, the values of and are and , respectively. For the bilateral test of , when , by querying the distribution table of the Wilcoxon signed rank test, . Because of , was accepted: there was no significant difference in the classification results between the two methods. Similarly, for MultiSTLP and TrGNB, the values of and are 182 and , , is accepted: there is no significant difference in the classification results of the two methods.

Parameter Sensitivity Analysis

In TrGNB, ARTL and TSVM-GP, they performed sensitivity analysis of parameters with large influence on performance of algorithm: TrGNB analyzed the difference parameter between source and target domains; MMD regularization parameter and manifold regularization parameter were analyzed to determine the influence of parameters on classification performance for ARTL; TSVM-GP analyzed the regularization coefficient of source domain, the regularization coefficient of source of target domains, and the trade-off term. Like them, we analyze the sensitivity of three parameters: regularization coefficient of target domain , regularization coefficient of source domain and trade-off coefficient in objective function of MultiSTLP, which illustrates their influence on the classification performance in this section. For each parameter, we fix the other two parameters at the optimal values determined by cross-validation, and then observe the effect of parameter with different values on the classification result. The experimental results are shown in Figs. 4, 5 and 6.
Fig. 4

Sensitivity of parameter for MultiSTLP

Fig. 5

Sensitivity of parameter for MultiSTLP

Fig. 6

Sensitivity of parameter for MultiSTLP

Sensitivity of parameter for MultiSTLP Sensitivity of parameter for MultiSTLP Sensitivity of parameter for MultiSTLP From the results of Figs. 4, 5 and 6, the following conclusions can be drawn: From Figs. 4 and 5, MultiSTLP is considerably sensitive to regularization parameters and with a wide range. This denotes that it is critical to determine the value of parameter by some effective strategies. In Fig. 6, we can see that shows that MultiSTLP is sensitive to . When approaches 1, MultiSTLP achieves the best classification performance. As is too small, the distribution difference between source and target domains is ignored, so the classification performance will be poor. When it is too large, the distribution difference between source and target domains will theoretically be larger, but this will also reduce the knowledge of source domains that can be transferred to target domain, and the classification performance is also poor.

Conclusion

Aiming at the current hot data privacy protection problem in machine learning, we propose a MultiSTLP by combining group probability information with transfer learning. MultiSTLP first uses AESVM to select representative dataset in each source domain; secondly, based on minimizing the marginal probability difference calculate the weight of samples of representative dataset; then, according to conditional probability difference calculates the weight of source domains; finally, the group probability knowledge in the target domain and the weighted knowledge of representative dataset from multiple source domains are combined into the support vector machine structure risk minimization framework, the objective function of MultiSTLP is proposed and proved theoretically. MultiSTLP not only improves training efficiency and result, but also protects data privacy. The effectiveness of the classifier obtained by training MultiSTLP is demonstrated on experiments utilizing four real-world datasets 20-Newsgroups, TRECVID 2005, Sentiment analysis and Email spam. Although the experimental results show that the MultiSTLP algorithm has advantages over the benchmark algorithms, it is still a problem worthy of further study in terms of training efficiency and domain similarity.
  6 in total

Review 1.  Machine learning: Trends, perspectives, and prospects.

Authors:  M I Jordan; T M Mitchell
Journal:  Science       Date:  2015-07-17       Impact factor: 47.728

2.  Feature Selection with Annealing for Computer Vision and Big Data Learning.

Authors:  Adrian Barbu; Yiyuan She; Liangjing Ding; Gary Gramajo
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2016-03-21       Impact factor: 6.226

3.  Unsupervised multi-sense language models for natural language processing tasks.

Authors:  Jihyeon Roh; Sungjin Park; Bo-Kyeong Kim; Sang-Hoon Oh; Soo-Young Lee
Journal:  Neural Netw       Date:  2021-05-25

4.  Transfer learning for establishment of recognition of COVID-19 on CT imaging using small-sized training datasets.

Authors:  Chun Li; Yunyun Yang; Hui Liang; Boying Wu
Journal:  Knowl Based Syst       Date:  2021-02-06       Impact factor: 8.038

5.  Multi-source fast transfer learning algorithm based on support vector machine.

Authors:  Peng Gao; Weifei Wu; Jingmei Li
Journal:  Appl Intell (Dordr)       Date:  2021-04-06       Impact factor: 5.019

6.  Multi-Source Deep Transfer Neural Network Algorithm.

Authors:  Jingmei Li; Weifei Wu; Di Xue; Peng Gao
Journal:  Sensors (Basel)       Date:  2019-09-16       Impact factor: 3.576

  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.