Literature DB >> 34603538

Iterative joint classifier and domain adaptation for visual transfer learning.

Shiva Noori Saray1, Jafar Tahmoresnezhad1.   

Abstract

Current available supervised classifiers cannot generalize across various domains due to distribution mismatch among them. Domain adaptation and transfer learning algorithms are proposed to tackle domain shift problem that originates from different data collection conditions. In this paper, we propose a transfer learning framework called iterative joint classifier and domain adaptation for visual transfer learning (ICDAV), which utilizes the balanced maximum mean discrepancy to better transfer knowledge across domains. Also, for learning a robust classifier against domain shift, a set of graph manifold regularizer and modified joint probability maximum mean discrepancy are simultaneously exploited to capture the domain structures and adapt the distribution of projected samples during the model learning process. Variety of experiments on several public datasets indicates that our approach achieves remarkable performance on visual domain adaptation and transfer learning tasks.
© The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature 2021.

Entities:  

Keywords:  Domain adaptation; Manifold regularization; Transfer learning; Visual classification

Year:  2021        PMID: 34603538      PMCID: PMC8479271          DOI: 10.1007/s13042-021-01428-z

Source DB:  PubMed          Journal:  Int J Mach Learn Cybern        ISSN: 1868-8071            Impact factor:   4.377


Introduction

In today’s world where the information is derived from classified data, and given that in any specialized fields, the data (text, audio and video) classification is one of the most important issues. Artificial intelligence is one of the most important solutions to this issue. Data classification is widely used in many specialized fields including medicine, military, traffic guidance and weather forecasting [4, 32, 45, 48]. One example of medical application is the diagnosis of disease. Now that the world is facing the Covid-19 problem, the robust diagnostic models are needed to identify the symptoms of Covid-19 [13]. Supervised learning as a well-studied machine learning method needs massive amounts of labeled data to create the robust models [1]. Since the manual labeling of training data to learn a model is prohibitive and the input dataset may be bias, the learned model most probably will perform poorly on new data. For example, the image properties of people can be vary in different countries, which causes the distribution shift across domains. Transfer learning is one of the promising solutions to solve the label famine and dataset bias problems [28, 39, 51]. Transfer learning benefits from abundant labeled samples of source domain to reduce the dataset bias in target data labeling. Auxiliary source domain often has different distribution from target domain that makes traditional machine learning algorithms useless, since they assume that the source and target data are under the same distributions. In such a circumstance, transfer learning methods are used to reduce the distribution divergence where they can be divided into following three categories: (a) instance-based methods [47], which reuse the samples from source domain according to re-weighting techniques, (b) feature-based methods [46, 54], which learn the subspace with shared features to represent the source and target data under common conditions or perform distribution alignment to minimize the marginal or conditional distribution divergences between domains, and (c) model-based methods [31], which transfer parameters of source model to improve the performance of target model. One of the effective factors in domain discrepancy is the marginal and conditional distribution discrepancies. Joint marginal and conditional distribution adaptation is a useful method in transfer learning, which measures the distribution shift between domains by a metric such as maximum mean discrepancy (MMD) [42]. In many joint distribution adaptation based approaches, the marginal and conditional distributions are often treated equally, which may not be optimal [9, 34, 43]. In the other words, if two domains are very dissimilar, it means that the marginal distributions have more discrepancy and need more notice to align; while if two domains are similar (i.e., the marginal distributions are close), it means that the conditional distributions needs more attention and should be given more weight. Hence, considering the distributions with equal importance will degrade the efficiency of methods. To this end, in the first step, we use the dynamic balanced distribution adaptation to specify the relative importance of marginal and conditional distributions by -distance and learning a domain-invariant representation. Also, in the second step to have a domain-invariant transfer classifier, the joint probability distribution discrepancy is utilized. Most of the available joint adaptation methods compute the sum of marginal and conditional distribution discrepancies, while the joint probability distribution discrepancy uses the product of class-conditional probability and the class prior probability to compute the discrepancy across domains. In fact, the calculation of joint probability distribution discrepancy is simpler and more accurate. Also, joint probability distribution discrepancy increases both the consistency of the same class samples from various domains (i.e., inter-domain transferability maximization) and the discrepancy of various classes (i.e., inter-class discriminability maximization) to facilitate the classification. In addition, for obtaining a robust transfer classifier with superior performance, the principle of structural risk minimization (SRM) [22] and a group of graph manifold regularizer are exploited. SRM consists of loss function and regularizer term to minimize the loss on predicted labels and preserve the structure of domains in transfer learning process. In summary, our proposed method brings the following contributions. (1) We introduce a modified visual domain adaptation (MVDA) to learn a new representation to avoid feature distortion in original space. MVDA, in addition exploits the domain-invariant clusters to discriminate between different classes, and considers the relative importance of marginal and conditional distributions, quantitatively, for domain adaptation. (2) We use two forms of MMD criteria to benefit the advantages of it in learning the new subspace and improving the classifier efficiency. We utilize the dynamic balanced MMD form in a novel representation learning step via proposed MVDA to perform an optimal distribution adaptation and obtain a new subspace. Dynamic balanced MMD evaluates the importance of marginal and conditional distributions to perform the optimal domain adaptation. Also, we benefit the joint probability MMD to boost the transferability and discriminability of classifier by reducing the discrepancy of the same class samples of various domains and increasing the discrepancy between different classes’ samples of various domains. (3) We introduce the projected joint probability distribution discrepancy, which consists of embedded classifier to increase the between-class discriminability and between-domain transferability by participating the prediction function to improve the transfer learning performance. (4) For increasing the compatibility between the intrinsic manifold structure of data and classifier structure, two graph-based regularizers (i.e., Hessian and Laplacian graphs) are utilized by embedded classifier. (5) We conduct extensive experiments on real-world benchmark datasets to demonstrate the advantage of our proposed ICDAV over state-of-the-art traditional transfer learning and domain adaptation methods. The results illustrate a noteworthy improvement of ICDAV in term of average classification accuracy compared to the other transfer learning methods on most datasets. The remainder of this paper is organized as follows. We review the related work in Sect. 2. In Sect. 3, we describe our proposed ICDAV method and its components. Extensive experiments are shown in Sect. 4, where it consists of dataset definition and performance results. Moreover, we extensively analyze the performance of ICDAV in Sects. 5 and 6. At the end, Section 7 concludes the paper.

Related work

The scope of transfer learning is very wide, including homogeneous and heterogeneous transfer learning [19], visual domain adaptation [15, 36, 40] and cross-dataset recognition [50], which is applicable in various areas such as object recognition [8], text classification [16], speech recognition [12] and face recognition [49]. Transfer learning methods are divided into the following three general categories based on the type of transferred information between source and target data [29]: instance-based transfer learning, feature-based transfer learning and model-based transfer learning. Instance-based transfer learning: In this perspective, some of the labeled source data that are more similar to target data achieve more weight to be used in training of a target classifier. Hasse et al. [11] introduced “Instance-weighted transfer learning of active appearance models” approach, which integrates the training data from an auxiliary source domain via predicting the sample-specific weights and selects similar but still informative samples that improves the degree of transfer. Chen et al. [5] introduced “Instance weighting framework for genetic programming (GP) based symbolic regression for transfer learning”, which utilizes the differential evolution to achieve the optimal weights during the evolutionary process of GP. This framework helps GP to recognize and remove the less useful source domain instances and learn from more useful source domain instances. Feature-based transfer learning: This consists of subspace and manifold learning methods that map data from source and target domains into a new subspace with a better representation. Also, the distribution alignment methods, which they reduce the marginal, conditional or both distribution discrepancies across domains falls into this perspective. Principal component analysis (PCA) [14] is one of the most widely used methods, which reduces the dimensionality of data and finds a new subspace, while preserving the variability as much as possible. Kute et al. [17] proposed “Cross domain association using transfer subspace learning” that uses Bregman divergence regularization to reduce the distribution divergence across training and test domains to map into a common subspace. Also, it utilizes Fisher linear discriminant analysis subspace learning algorithm [37] to obtain the projection matrix to discriminate samples within the individual domains. Li et al. [18] proposed “Manifold alignment and balanced distribution adaptation (MBDA)” approach, which uses the balance factor to weigh the marginal and conditional distribution adaptation. Also, MBDA uses the marginal distribution to preserve the neighboring structures of domains, and reduce the dimension as much as possible. Manifold transfer learning via discriminant regression analysis (MTL-DRA) [24] is also one of the newest feature-based methods. MTL-DRA utilizes within- and between-class graphs to transfer the local geometry structure information across the source and target domains. Moreover, it provides a transform matrix, which is robust and sparse to deal with the noise and the negative transfer learning. Model-based transfer learning: In this perspective, the goal is to adapt the trained classifier or regression model via the source domain to use in target domain. A novel shape deviation modeling scheme is proposed in [6], which models the dimensional error of the product in a parameter-based transfer learning approach. In particular, the shape deviation is decomposed into the following two components: i) the shape-independent error, which is modeled by multiple linear regression models in a Cartesian coordinate system, and ii) the shape-specific error, which is separated from the shape-independent error by analyzing the error generating mechanism with a particularly designed experiment. Thus, for prediction of the shape independent error of any new shape and minimization of the need in creating more test parts, the earned knowledge is transferred. In this paper, our proposed method is a combination of both feature- and model-based transfer learning perspectives. In this approach, the feature-based learning perspective considers the adaptation of source and target domains using dynamic balanced distribution adaptation and dense clusters. Whereas, in the perspective of model-based transfer learning, the domain-invariant optimal classifier learning is desired in a new subspace. In the classifier learning phase, in addition to minimizing the loss function, the joint probability distribution discrepancy adaptation is used to reduce the difference between conditional and marginal distributions, jointly. Also, a set of graph-based regularizer is utilized in classifier learning process, which benefits from the manifold structures.

Proposed method

Let domain D is composed of two components, the feature space X and the marginal probability distribution P(X). The source domain and target domain are defined as and , respectively, which and illustrate the data from source and target datasets with the same class space, and induces the source labels. The data in domain is unlabeled. The number of samples in and are and , in turn. Also, each domain has another distribution called conditional distribution, i.e., P(Y|X), which is equivalent to the classifier f(X). In fact, the conditional distribution indicates the probability of occurring label for input . Accordingly, ICDAV first learns the transferable features to represent data with the best properties. Then, ICDAV performs the adaptive classification to classify target data. These two steps are iteratively optimized via several iterations. Therefore, we consider and , separately, to distribution adaptation. After that, we assume , directly, to find the better domain-invariant classifier.

Modified visual domain adaptation

In this step, we seek to introduce the modified visual domain adaptation to eliminate domain shifts. For this, we first briefly define VDA, which is the basis of MVDA. The goal of VDA is to obtain an adaptation matrix , which projects the source and target domains from m-dimensional feature subspace into a k-dimensional invariant feature subspace where the sample properties in the new representation is preserved. For finding the projection matrix A, Eq. 1 is utilized:where the first and second components introduce the marginal and class-conditional distribution distances across domains, respectively, where it uses the maximum mean discrepancy as a distance metric. The third component also shows the distance of each sample from its mean () in class c. Moreover, A, and denote the adaptation matrix, the set of source and target data and the trace of matrix, in turn. Also, is MMD coefficient matrix where , and are computed through , and , respectively. As the same way, matrix are calculated by , and , respectively, where and represent in order the number of samples in source and target domains related to class c. In VDA, PCA is used to reconstruct data such that is maximized subject to where induces the covariance matrix of input data. Also, is the centering matrix where I and 1 are the identity matrix and the ones vector, in turn. Thus, the optimization problem is achieved as follow:where and are the Frobenius norm and the regularization parameter, respectively. We propose the modified VDA in which the importance of aligning marginal (i.e., P(X)) and conditional (i.e., P(Y|X)) distributions in domain adaptation is considered differently due to domain shift scale. To this end, we refine Eq. 2 as follow:where is the adaptive factor that is estimated through Eq. 4:where and are the marginal and conditional –distances and are introduced as the errors of a linear classifier () in discrimination of source and target domains asAccordingly, the Lagrange function for Eq. 3 can be written aswhere illustrates the Lagrange multipliers. By setting , the generalized eigen-decomposition is obtained asThe adaptation matrix A is then formed by the k trailing eigen-vectors. At the end, the new representation of source and target data can be achieved as and . Note that due to the lack of label in target domain in the first iteration, the marginal adaptation only is participated in distribution adaptation. After that, both the marginal and conditional distributions are utilized together in distribution adaptation and the labels of target domain is refined in later iterations.

Adaptive classifier

The aim of this step is to train a transferable classifier f, which has the least error on target domain.

Structural risk minimization

For obtaining the prediction function f, we use the structural risk minimization principle [41], which is formulated aswhere represents the loss function and and y are the predicted label and true label of sample z, in turn. is the reproduced kernel Hilbert space (RKHS) [3] induced by Mercer kernel [26] function K(., .). Also, represents the regularization term, which can be composed of several components. We describe the regularization as follows:where the definition of different components of Eq. 9 are explained in the rest. The first term induces the modified joint probability MMD that performs the joint distribution alignment, which uses the maximum mean discrepancy measurement to compute the joint distribution shift across domains. This term performs the discriminability concurrently with transferability. Also, and are graph manifold regularizer based on Hessian and Laplacian regularizations, respectively. Moreover, , , are the regularization coefficients. Thus, the embedding of Eq. 9 in Eq. 8 results the Eq. 10 as follows:Classifier f can be defined as Eq. 11 due to the representer theorem [2]:where denotes the coefficients vector.

Loss function

For minimizing the error of classifier , the square loss function is adopted where z is the projected samples of source and target domains. By considering Eq. 11, the first term of Eq. 10 becomeswhere is the squared norm of classifier f in RKHS while indicates the Frobenious norm. Coefficient denotes the regularization parameter. Also, induces the kernel matrix whose entries are defined as , and is a diagonal domain indicator that has value 1 if and value 0 if . Moreover, denotes the source and target label matrix. In the rest, we introduce the components of regularization term in details.

Modified joint probability MMD adaptation

In this step, instead of considering (i.e., the marginal distribution discrepancy) and (i.e., the conditional distribution discrepancy), separately, (i.e., joint marginal and conditional distribution discrepancy) is jointly considered to distribution adaptation. Joint probability MMD [52] is defined as Eq. 13, which calculates the joint probability discrepancy directly, and increases both the class discriminability and the domain transferability, simultaneously, as follows:where is a trade-off coefficient, while , and are kernel matrices of projected samples of source, target and total matrices, respectively. Also, and are defined as and where , . such that and are one-hot coding label matrices of source and target domains. Also, and are defined to calculate where and demonstrate the column of and repeat time of , respectively. In the same way, illustrates the to columns of matrix . Therefore, the modified joint probability MMD is achieved in Eq. 14 by embedded classifier f (using Eq. 11) as the following feature map function:

Set of graph manifold regularizer

The fourth and fifth terms of Eq. 10 induce the second-order Hessian energy regularizer and the Laplacian regularizer, respectively. The set of both graph regularizer based on the Hessian and Laplacian graphs are utilized to preserve the manifold structure for global label consistency. The second-order Hessian energy regularizer performs extrapolation better than the Laplacian regularizer if the number of labeled data is low. The definition of Hessian energy is aswhere is the squared norm of second covariant derivative of f in data point , which corresponds to the Frobenius norm of Hessian of f at normal coordinates. Parameters denotes the set of source and target data and d illustrates the PCA eigenvectors of p nearest neighbors space , which forms the orthogonal basis of local tangent space . Also, and are variables of function f. Moreover, and induce the neighboring point and function value at point , in turn. Moreover, the symbol denotes the Hessian energy matrix, which is the sum up of all matrices. The definition of is given as Eq. 16 where H is an operator to estimate the Hessian of f at . Due to each sample is only associated with its neighbors, the Hessian energy matrix is sparse as follows:In fact, the Laplacian graph is utilized to preserve the label information in built manifold on training data, which is defined aswhere represents the Laplacian graph matrix, in which shows a diagonal matrix that is calculated through . Also, illustrates the affinity matrix whose entries are described aswhere is the local set consisting of the p-nearest neighbors of .

Overall reformulation

Thus, the combination of Eqs. (12), (14), (15) and (17) results the Eq. (19) as follow:By setting the derivative , the solution for Eq. (19) is obtained as follow:In this way, we introduce a two-step framework for image classification that has considerable performance compared to other state-of-the-art transfer learning and domain adaptation methods. In the next section, to evaluate the performance of the proposed method, we provide the classification results on benchmark visual datasets.

Experiments

Here, we describe the used benchmark visual datasets in our experiments. We utilize following five datasets to evaluate the prediction accuracy of the proposed method compared to the other methods. I. Office + Caltech-10 dataset [10, 35] Office + Caltech-10 dataset contains 4 domains, in which 3 domains (i.e., Amazon (A), DSLR (D), Webcam (W)) are from the Office-31 and the another one (i.e., Caltech10 (C)) is from Caltech-256. In experiments, the samples number of Amazon, DSLR, Webcam and Caltech10 domains are 958, 157, 295 and 1123, respectively. The common 10 classes (i.e., projector, calculator, head-phones, keyboard, bike, mouse, monitor, backpack, mug, and laptop) among the Office-31 and Caltech-256 are exploited to perform experiments across 4 domains. In evaluation, 12 tasks for 4 domains are considered as follows: AD, AC, AW, DA, DC, DW, CA, CD, CW, WA, WC and WD. II. Digits dataset [7] MNIST (M) and USPS (U) are two datasets containing the images of 0 to 9 digits with different distributions. The MNIST consists of 60,000 training images and 10,000 test images. The USPS dataset consists of 7291 training images and 2007 test images. In experiments, 2000 and 1800 images are randomly sampled from MNIST and USPS, respectively. Also, each image are described with a 256-dimensional feature vector. In evaluation, two cross-domain tasks, USPSMNIST and MNISTUSPS are examined. III. Multi-PIE dataset [38] Multi-PIE consists of 41,368 face images of 68 different identities with different poses, illuminations and expressions. In experiments, cross pose face recognition is performed on five different face orientations, including PIE05 (P1): left pose, PIE07 (P2): upward pose, PIE09 (P3): downward pose, PIE27 (P4): front pose and PIE29 (P5): right pose. Totally, 3332, 1629, 1632, 3329, and 1632 facial images are contained in P1, P2, P3, P4 and P5, respectively. Therefore, 20 tasks were evaluated, like P1P2, P1P3, P1P4, P1P5. IV. ImageNet-VOC 2007 [44] ImageNet (I) and VOC2007 (V) are large-scale image classification datasets. ImageNet-VOC 2007 couple is much larger in diversity and much more accurate compared to the current image datasets. However, five shared classes (i.e., chair, cat, bird, person and dog) are selected from both datasets. To verify the effectiveness of proposed approach, we use two tasks ImageNet-VOC 2007 (I-V) and VOC 2007-ImageNet (V-I), which each dataset is considered as one domain. V. COIL20 [33] The Columbia Object Image Library (COIL-20) dataset consists of 1440 gray scale images with 20 classes (objects such as duck) and 72 images per class (object). Different images of the same class are captured in different angles (i.e., at the interval of five degrees), which makes the images of the same class to have different distribution. We use two subsets C1 (all images taken in the directions of [0, 85] [180, 265]) and C2 (all images taken in the directions of [90, 175] [270, 355]) to construct two tasks C1-C2 and C2-C1.

Method evaluation

In this section, we present the performance of our ICDAV algorithm on five widely used object recognition datasets (Table 1). A comparison of the proposed method with other DA methods including TCA [30], JDA [23], CDDA [25], BDA (Wang et al. 2018), CLGA [20], GSL [52], JPDA [53], UCGS [21] and CIDA [27] is also provided.
Table 1

Classification accuracy (%) on Office + Caltech-10, Multi-PIE, MNIST + USPS, ImageNet-VOC and COIL20 datasets

DatasetsourcetargetTCA (2011)JDA (2013)CDDA (2017)BDA (2018)CLGA (2018)GSL (2019)JPDA (2020)UCGS (2021)CIDA (2021)ICDAV
 Office-Caltech-10 CA38.2044.7848.3344.5748.0258.748.5441.7659.19
W38.6441.6944.7540.3442.3759.745.7633.4857.97
D41.445.2248.4145.2249.0449.745.8633.0452.23
 AC37.7639.3642.1239.2742.34642.2152.3048.98
W37.6337.9741.6937.9741.3644.142.0337.6851.86
D33.1239.4937.5840.7636.3147.136.9436.3348.41
 WC29.3031.1731.9731.4332.9537.935.1747.831.61
A30.0632.7837.2732.4634.5741.833.8239.3242.28
D87.2689.1787.989.1792.3688.589.1780.6893.63
 DC31.7031.5234.6431.1733.6635.334.4646.5036.15
A32.1533.0933.5133.1935.9940.634.3443.9541.75
W86.1089.4990.5189.4989.8385.891.1972.6190.51
 Multi-PIE P1P240.7658.8160.2258.2067.8350.1558.2065.1238.6071.21
P341.7954.2358.752.8263.8559.6866.5462.8144.4865.69
P459.6384.5083.4883.0388.9584.8182.8879.6972.0990.18
P529.3549.7554.1749.1461.7654.7249.7551.2939.8361.64
 P2P141.8157.6262.3357.3571.448.0263.3662.6137.6373.74
P351.4762.9364.6462.7572.9842.1660.4863.9140.774.75
P464.7375.8279.975.7686.2473.3677.5380.8465.5691.44
P533.7039.894439.7151.2337.5047.7955.2734.6260.72
 P3P134.6950.9658.4651.3570.1755.3459.0360.0247.5575.93
P247.7057.9559.7356.4173.4853.1061.5162.4944.9170.60
P456.2368.4677.267.8689.3173.2774.878.1372.7991.98
P533.1539.9547.2442.4055.5155.8251.1658.2750.6761.03
 P4P155.6480.5883.180.5289.5686.4684.2182.6269.1893.52
P267.8382.6382.2683.0692.9478.9483.1884.8469.8493.5
P375.8687.2586.6487.2593.0881.0786.7683.0969.4290.07
P540.2654.6658.3354.5371.6372.6764.7168.9262.0677.45
 P5P126.9846.4648.0247.9957.6845.6253.3958.0736.0964.98
P229.9042.0545.6143.2255.4339.4749.8554.9234.0161.76
P329.9053.3152.0247.9258.0350.9857.3561.6946.8169.79
P433.6457.0155.9957.1071.8566.9659.8470.7159.2576.48
DigitsUM51.0559.6562.0559.9058.3528.3659.3566.10
MU56.2867.2876.2267.3971.2851.2369.17--81.11
ImageNet-Voc 2007IV63.763.465.0268.1661.2662.7168.7270.41
VI64.970.274.0679.2972.4672.3578.8982.75
Coil20C1C288.4789.3191.5389.4496.8192.992.0899.72
C2C185.8388.4793.8988.3391.1189.389.86-99.58
Avg. Office-Caltech-1043.6146.3148.2246.2548.2352.948.2947.1254.55
Avg. Multi-PIE44.7560.2463.1059.9272.1560.5164.6267.2751.8175.82
Avg. Digits53.6763.4769.1463.6564.8239.864.2673.61
Avg. Image-Voc64.366.8-69.5473.7366.8667.5373.8176.58
Avg. Coil2087.1588.8992.7188.8993.9691.190.9799.65
Overall Avg.58.765.1465.6570.5862.2367.1376.04
All of the reported results are the classification accuracy on target domain, which is defined as follows:where and y(x) are the predicted label of target dataset and the real label vector. In the rest, we compare our ICDAV with other state-of-the-art methods. Transfer component analysis (TCA) seeks to obtain transfer components across domains and find a new latent subspace for marginal distribution adaptation by maximum mean discrepancy (MMD). While ICDAV uses the balanced marginal and conditional distribution adaptation to improve the transfer learning between the source and target domains. Our results show that ICDAV gets (10.94%), (19.94%), (31.07%), (12.28%) and (12.5%) significant classification accuracy improvement compared to TCA on Office-Caltech-10, Digits, PIE, ImageNet-VOC and COIL20 datasets, respectively. Joint distribution adaptation (JDA) treats with both of the marginal and conditional distributions equally to adapt domains. Also, JDA measures the distribution shift across domains by maximum mean discrepancy. However, ICDAV adapts both distributions by considering different significance for each distribution to optimally adapt domains. As is clear from the results, ICDAV achieves (8.24%), (10.14%), (15.58%), (9.78%) and (10.76%) improvement in classification accuracy compared to JDA on Office-Caltech-10, Digits, PIE, ImageNet-VOC and COIL20 datasets, respectively. Close yet discriminative domain adaptation (CDDA) obtains a latent feature representation by minimizing both marginal and conditional probability distribution discrepancies of source and target data via MMD and maximizing the distances between classes using repulsive force term. However, ICDAV increases the dependence between the labels and projected features alongside reducing the difference between the marginal and conditional distributions. ICDAV obtains (6.33%), (4.47%), (12.72%) and (6.94%) improvement against CDDA in prediction accuracy on Office-Caltech-10, Digits, PIE and COIL20 datasets, respectively. Classification accuracy (%) on Office + Caltech-10, Multi-PIE, MNIST + USPS, ImageNet-VOC and COIL20 datasets Balanced domain adaptation (BDA) is an extended version of JDA, which allocates different weights to marginal and conditional distributions by several methods. BDA considers domain adaptation as the sum of the marginal and conditional distribution discrepancies, while ICDAV utilizes both the sum and joint probability forms of distribution discrepancies for domain alignment. ICDAV on Office-Caltech-10 datasets obtains (8.3%) improvement, and in Digits, PIE, ImageNet-VOC and COIL20 datasets obtains (9.96%), (15.9%), (7.04%) and (10.76%) performance improvement against BDA, respectively. Coupled local–global adaptation (CLGA) is an unsupervised multi-source domain adaptation approach, which jointly reduces both the marginal and conditional distribution shifts to maximize the adaptation ability between source and target dataset (i.e., global level). Moreover, CLGA maximizes the discriminative ability (i.e., local level) to obtain a shared subspace by exploiting both class and domain manifold structures of data samples. However, ICDAV finds domain-invariant subspace for source and target domains and minimizes the distributional and geometrical discrepancies across domains. According to the results, ICDAV achieves (6.32%), (8.79%), (3.67%), (2.85%) and (5.69%) improvement in average accuracy compared to CLGA on Office-Caltech-10, Digits, PIE, ImageNet-VOC and COIL20 datasets, respectively. Guide subspace learning for unsupervised domain adaptation (GSL) integrates three guidance losses including subspace guidance for distribution adaptation, data guidance for domain adaptation, and label guidance for label prediction to obtain target subspace. While our ICDAV benefits from joint MMD-based distribution adaptation and clustering to adapt domains. Also, unlike GSL which uses the basic classifiers for data classification, ICDAV proposes a novel classifier, which is more accurate and robust against noise. Moreover, ICDAV illustrates further improvement and achieves the state-of-the-art performance (1.65%), (33.81%), (15.31%), (9.72%) and (8.55%) on Office-Caltech-10, Digits, PIE, ImageNet-VOC and COIL20 datasets, respectively, against GSL. Joint probability domain adaptation (JPDA) reduces the joint probability distribution disparity of the same class across domains for transferability, and increases the joint probability distribution disparity between different classes of different domains for discriminability, simultaneously. However, ICDAV uses the balanced and joint probability of distribution disparities, jointly for effective transfer across domains. Compared to JPDA, the average performance improvement of ICDAV is (6.26%), (9.35%), (11.2%), (9.05%) and (8.68%) on Office-Caltech-10, Digits, PIE, ImageNet-VOC and COIL20 datasets, respectively. Unified cross-domain classification method via geometric and statistical adaptations (UCGS) unifies the structural risk minimization, frequently-used joint MMD adaptation, and the Nyström method for learning an adaptive classifier by considering the statistical and geometric adaptations. While ICDAV uses discriminative JPMMD, which is more accurate and enhances the discriminability and transferability at the same time. Also, ICDAV utilizes a set of graph manifold regularizer (i.e., Hessian and Laplacian) that explores better geometric structure across domains compared to the Nyström method. The performance improvement of ICDAV is (8.55%) and (2.77%) on PIE and ImageNet-VOC datasets, respectively, compared to the method UCGS. Cross- and multiple-domains visual transfer learning via iterative Fisher linear discriminant analysis (CIDA) uses modified Fisher linear discriminant analysis to adapt domains. In the other words, CIDA maximizes the marginal distribution discrepancy of different classes of source and target domains and minimizes the difference between marginal distribution of the same class samples and the amount of variance between the various classes. Our ICDAV utilizes the clustering and balanced distribution adaptation for domain adaptation. Also, CIDA uses the empirical risk minimization and Laplacian matrix to construct a classifier, which does not consider the distribution adaptation. While ICDAV exploits the joint probability distribution adaptation and coupled Laplacian and Hessian energy regulizers to construct a robust adaptive classifier. ICDAV gains (7.43%) and (24.01%) on Office-Caltech-10 and PIE datasets, respectively, compared to CIDA in terms of the average classification accuracy.

Parameter evaluation

In this section, for performance evaluation of our proposed method in different conditions, we provide the parameter evaluation on tasks of following five benchmark datasets Office-Caltech-10, Multi-PIE, Digits, ImageNet-VOC 2007 and COIL20. Therefore, the classification accuracy changes according to the different values of parameters , k, , , , p, and ζ on five selected tasks C-A (from Office-Caltech-10), P1-P2 (from Multi-PIE), U-M (from Digits), I-V (from ImageNet-VOC 2007) and C2-C1 (from COIL20) as shown in Fig. 1.
Fig. 1

The performance analysis of ICDAV on three tasks C-A (from Office-Caltech-10), P1-P2 (from Multi-PIE), U-M (from Digits), I-V (from ImageNet-VOC) and C2-C1 (from COIL20) with respect to the parameters , k, , , , p, and ζ

Figure 1. (a) shows the classification accuracy of parameter [0.0 1.0] on five tasks C-A, P1-P2, U-M, I-V and C2-C1. The parameter is the regularization parameter of modified joint probability MMD. As is clear from the plot, ICDAV has smooth flow performance in different values of parameter on two tasks (i.e., I–V and C2–C1), which it means that the joint probability MMD has no effect on I–V and C2–C1 tasks because there is not much distribution difference between domains, therefore, this component does not help in transferability and discriminability between domain samples. As the same way, the prediction accuracy in two other tasks (i.e., P1-P2 and C-A) increases with a low slope, which means that the domain data in two tasks are very similar, and the component has a constructive effect on transferability and separability of the classifier. By increasing the value of , the performance of our method on U-M task increases at first and then decreases, which indicates a high distribution difference between domains. In the other words, when the difference of domain distributions is large, the joint probability MMD component can reasonably improve the transferability of the same class between different domains and discriminability between different classes of different domains, but increasing the impact of this component decreases the classifier performance. Figure 1. (b) induces the experiments for evaluating the parameter k {5,10,...,60}. The parameter k illustrates the dimensionality of latent subspace. The plot indicates that in all five cases, increasing the value of k increases the ICDAV performance. In fact, ICDAV determines better performance in subspaces with high dimension because in this case the novel latent subspace is closer to the original subspace and has less information loss, so we will have a higher classification accuracy with high values of k. Figure 1. (c) indicates the experiments of parameter on tasks of Office-Caltech-10, Multi-PIE, Digits, ImageNet-VOC 2007 and COIL20 datasets. The evaluation range of parameter is 0.0 to 1.0. The classification accuracy on task P1-P2 has negative slope after = 0.1 while on tasks U-M, I-V and C2-C1, it almost behaves smoothly. Also, the classification accuracy has partially uniform manner on C-A task. The descending manner on P1-P2 shows that ICDAV cannot learn a robust representation for cross-domain classification in Multi-PIE dataset at this interval. While, uniform slope of prediction accuracy on other tasks for the values more than 0.1 indicates that the robust representation is constructed in and more without considering the value of . Figure 1. (d) illustrates the classification accuracy of ICDAV with respect to [0.0 1.0] where is the trade-off parameter between transferability and discriminability of different classes and different domains. ICDAV on tasks U-M and I-V has partially uniform trend, which induces the minor role of discriminability between different classes of different domains. In the other words, in these tasks, various classes’ samples in various domains have enough distribution differences, and discriminability component cannot further increase the distribution difference. ICDAV behaves differently on Office-Caltech-10 (C-A), Multi-PIE (P1-P2) and COIL20 (C2-C1) datasets and has a completely downward trend. The downward trend on both tasks C-A and P1-P2 induces that the cross-domain transferability component contribute most to the discrepancy and increasing the coefficient of cross-class discriminability component reduces the performance of ICDAV. However, there is a sudden decrease on task C2-C1, which it means that the distributions of the same class samples between different domains is similar and the transferability should be given more weight than the discriminability. Therefore, allocating more weight in the discriminability component decreases the efficiency of the method, suddenly. The performance analysis of ICDAV on three tasks C-A (from Office-Caltech-10), P1-P2 (from Multi-PIE), U-M (from Digits), I-V (from ImageNet-VOC) and C2-C1 (from COIL20) with respect to the parameters , k, , , , p, and ζ Figure 1. (e) presents the results for parameter on five tasks (i.e., C-A, P1-P2, U-M, I-V and C2-C1) where is the regularization parameter of second-order Hessian energy regularizer, which preserves the manifold structure in domain adaptation. ICDAV behaves unbalanced but ascending on both tasks P1-P2 and U-M while it has mild descending manner on tasks C-A, I-V and C2-C1. This behavior shows that the similarity of domain pairs structures in P1-P2 and U-M tasks is high, and with increasing the value of parameter , the effect of the Hessian component on the domain adaptation increases, and has a positive effect on the transfer of inter-domain knowledge. While in other three tasks (i.e., C-A, I-V and C2-C1), domain structures are not very similar and for high values of parameter, the effect of the mentioned component in domain adaptation is negative. Figure 1. (f) shows the effect of varying values of p on ICDAV performance. Therefore, p {2, 4, ..., 20} induces p-nearest neighbors in the second-order Hessian energy regularizer and Laplacian regularizer. ICDAV has stable behavior during the value changes of parameter p on four tasks (i.e., C-A, U-M, I-V and C2-C1) while the inconstant trend is observed in the middle of the chart on P1-P2. In four tasks C-A, U-M, I-V and C2-C1, the best geometric structure of the domain is obtained in p with values much more than 60, so these tasks experience a constant slope at low values p. Whereas in task P1-P2, the best geometric structure of domains is obtained at several points, even at low values of p. Figure 1. (g) demonstrates the performance of ICDAV relation to the change in values of [0.0 1.0]. Parameter indicates the importance of Laplacian regularizer in optimization problem. As seen in the plot, by increasing the value of the parameter , the classification accuracy fluctuates slightly on tasks P1-P2 and U-M. Also, the performance is declining by increasing the parameter on C-A, I-V and C2-C1 tasks. In fact, the Laplacian is the neighborhood graph of samples and the best Laplacian graph gives the best data structure when the similar data place in the same neighborhood graph as much as possible. The best Laplacian structure of domains on tasks C-A, I-V and C2-C1 is achieved in low values of because the dissimilarity of domain samples in these tasks causes that the good graphs are not formed. Therefore, by increasing the value, the accuracy of classification decreases. While on tasks P1-P2 and U-M, the optimal Laplacian structures of domains are obtained in different points. Figure 1. (h) displays the performance sensitivity of ICDAV to parameter [0.0 1.0], which is the regularization parameter for squared norm of classifier f in RKHS. In fact, supervises the complexity of the prediction function. Therefore, the accuracy has stable manner in four tasks C-A, U-M, I-V and C2-C1 for values more than 0.1 while it is a bit non-stable on task P1-P2. In fact, 0.1 is a critical value for parameter on all tasks. When 0, the classifier performance decreases and the classifier overfits. While, by , the classifier is not fitted on input samples and underfitting occurs where the diagram displays the constant slope.

Ablation study

In this section, we evaluate the impact of each component on prediction accuracy of proposed method. Therefore, we provide Table 2, which shows the effect of each component by setting its coefficient to zero.
Table 2

Impacts of ICDAV’s parameters on recognition accuracies on Office-Caltech-10, Multi-PIE, Digits, ImageNet-VOC and COIL20 datasets

DatasetSourceTargetWithout \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\eta$$\end{document}ηWithout \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\lambda$$\end{document}λWithout \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\gamma$$\end{document}γWithout \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\mu$$\end{document}μWithout \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\omega$$\end{document}ωWithout \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\rho$$\end{document}ρWithout \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\zeta$$\end{document}ζICDAV
 Office-Caltech-10 CA59.0854.9156.0559.1959.9259.609.6059.19
W58.3153.9057.6357.9752.8858.3149.4957.97
D52.2351.5950.3252.2345.8653.5045.2252.23
 AC48.1747.8246.4848.1749.7848.446.5948.98
W50.1746.1050.1750.5146.7852.5412.5451.86
D49.0444.5945.8648.4140.1347.1314.0148.41
 WC32.7728.5034.2831.4332.7731.2626.8031.61
A42.0728.7133.6142.0739.8742.389.6042.28
D93.6388.5492.9993.6388.5494.9094.9093.63
 DC35.8034.1136.9536.1535.9834.5510.4236.15
A41.6533.8242.9041.7543.9540.929.6041.75
W90.5189.1588.4790.5190.5190.5189.8390.51
 Multi-PIE P1P269.3120.1469.6770.1062.8671.3354.2171.21
P365.635.7065.7565.5659.9969.3058.5265.69
P488.7714.6089.3789.8289.6488.9894.4790.18
P562.8721.3257.4861.4063.6063.9732.9061.64
 P2P170.4432.3578.3670.0277.1665.5265.4073.74
P361.4632.0559.2573.4762.8157.6067.9574.75
P491.501.4791.4791.5090.7891.5090.0691.44
P562.8737.3862.6862.7550.5562.8746.9460.72
 P3P174.4323.6274.5873.8967.8973.5656.9975.93
P269.3156.2363.7270.6056.7271.2168.0270.60
P492.0478.8291.8691.9289.8593.4290.0991.98
P561.5254.1760.1160.3657.1763.9756.1961.03
 P4P192.651.4792.4492.0291.2793.1990.1993.52
P293.7482.4491.9693.6891.2294.0591.1093.49
P390.3885.6090.6390.4490.3289.9589.0990.07
P577.2166.6179.1778.3176.2379.9065.5077.45
 P5P163.901.4765.9763.8467.1465.9429.0264.98
P256.059.8266.8556.2951.8160.1048.3161.76
P370.8327.7070.5368.6965.3270.5350.0669.79
P472.7859.2773.1874.1172.0673.0971.7976.48
DigitsUM65.8048.3065.7065.8564.5565.6035.2066.10
MU81.1128.3979.5679.8373.4481.6161.6781.11
ImageNet-voc 2007IV70.4168.3967.6568.9968.9362.5013.6370.41
VI82.7579.4976.8879.6178.9370.3218.5382.75
Coil20C1C299.72599.7296.1193.7596.1166.5399.72
C2C199.58599.1797.597.0897.574.8699.58
Avg. Office-Caltech-1054.4550.1452.9854.3352.2554.5031.5554.55
Avg. Multi-PIE74.3835.6174.7574.9471.727565.8475.82
Avg. Digits73.4638.3472.6372.846973.6148.4373.61
Avg. Image-Voc76.5873.9472.2774.3073.9366.4116.0876.58
Avg. Coil2099.65599.4496.8195.4296.8170.6999.65
Overall Avg.75.7040.6174.4174.6472.4673.2746.5276.04
Parameter is the balance parameter of the marginal and conditional distributions in Eq. 7. Our ICDAV by considering the parameter, has a high performance in 9 out of 12 tasks in Office-Caltech dataset, 12 out of 20 tasks in PIE dataset, and 1 out of 2 tasks in Digits dataset. This result indicates that the balance parameter improves the prediction accuracy in more than half of tasks. Also, the average accuracy is improved 0.1% on Office- Caltech dataset, 1.44% on PIE dataset and 0.15% on Digits dataset, by using the parameter . While the parameter has no effect on ImageNet-VOC and COIL20 datasets because the marginal and conditional distributions are with equal importance on these datasets. The parameter is the regulizer parameter in Eq. 7. As can be seen from the results, the average accuracy of the label prediction is significantly improved by considering parameter . In the other words, the effect of this parameter on improving the prediction accuracy is 4.41% on Office-Caltech dataset, 40.21% on PIE dataset, 35.27% on Digits dataset, 2.64% on ImageNet-VOC dataset and 94.65% on COIL20 dataset. Parameter is the coefficient of influence of joint probability distribution in Eq. 20. According to Table 2, the mentioned component in ICDAV results 1.57%, 1.07%, 0.98%, 4.31% and 0.21% classification accuracy improvement on the Office-Caltech, PIE, Digits, ImageNet-VOC and COIL20 datasets, respectively. Also, the ICDAV method has a superior performance in 26 out of 38 tasks by considering parameter . Impacts of ICDAV’s parameters on recognition accuracies on Office-Caltech-10, Multi-PIE, Digits, ImageNet-VOC and COIL20 datasets Parameter is the trade-off parameter of joint probability distribution in Eq. 20. It establishes the balance between the and matrices. According to the results, parameter helps ICDAV to perform the same or better in 33 out of 38 tasks. In addition, the effect of parameter on prediction accuracy is 0.22% improvement on Office-Caltech dataset, 0.88% improvement on PIE dataset, 0.77% improvement on Digits dataset, 2.28% improvement on ImageNet-VOC dataset and 2.84% improvement on COIL20 dataset Parameter is the regularization coefficient of the second-order Hessian energy regulizer in Eq. 20. ICDAV obtains 3.58% prediction accuracy improvement and achieves high efficiency in 30 out of 38 tasks by considering the second-order Hessian energy regulizer. Parameter is the regularization parameter of Laplacian regulizer in Eq. 20. Parameter has the greatest impact on ImageNet-VOC dataset with 10.17% prediction accuracy improvement. Also, the effect of Laplacian regulizer is about 0.05%, 0.82% and 2.84% improvement on the Office-Caltech, PIE and COIL20 datasets, while in the Digits dataset is not improved. This indicates that the effect of parameter depends on the type of dataset. Parameter is the adjustment parameter of squared norm of function f in Hilbert space, which has a significant effect on improving the classification prediction accuracy. The mentioned component increases the efficiency of ICDAV in 36 out of 38 tasks and improves (29.52%) the average prediction accuracy of 38 tasks. In the other words, the presence of in objective function results 23%, 9.98%, 25.18%, 60.5% and 28.96% improvement on Office-Caltech, PIE, Digits, ImageNet-VOC and COIL20 datasets, respectively. As a result, ICDAV can have the best performance in domain adaptation by setting the optimal values of parameters. Also, the all parameters together have significant improvement on prediction accuracy of ICDAV.

Conclusion

As is from the results, our proposed ICDAV can cope with the distribution disparities across domains, with a successful approach. ICDAV minimizes data discrepancies through the modified visual domain adaptation method and obtains an adaptive classifier by maintaining the data structure and reducing the domain distribution discrepancies. To maintain the data structures, a set of graph manifold regularizer is exploited while the discrepancy reduction between domains is achieved using modified joint probability MMD. ICDAV uses two different forms to calculate MMD between data distribution, which has a tremendous impact on the quality of transfer learning across domains. Also our experiments induce the competitive results on all considered benchmark datasets. In future, we aim to perform some developments on proposed method to achieve more acceptable results in transfer learning. The developments can include multi-source experiments and embedded deep learning.
  6 in total

1.  Domain adaptation via transfer component analysis.

Authors:  Sinno Jialin Pan; Ivor W Tsang; James T Kwok; Qiang Yang
Journal:  IEEE Trans Neural Netw       Date:  2010-11-18

2.  Scatter Component Analysis: A Unified Framework for Domain Adaptation and Domain Generalization.

Authors:  Muhammad Ghifary; David Balduzzi; W Bastiaan Kleijn
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2016-08-11       Impact factor: 6.226

Review 3.  Principal component analysis: a review and recent developments.

Authors:  Ian T Jolliffe; Jorge Cadima
Journal:  Philos Trans A Math Phys Eng Sci       Date:  2016-04-13       Impact factor: 4.226

4.  Heterogeneous Domain Adaptation: An Unsupervised Approach.

Authors:  Feng Liu; Guangquan Zhang; Jie Lu
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2020-03-03       Impact factor: 10.451

5.  Guide Subspace Learning for Unsupervised Domain Adaptation.

Authors:  Lei Zhang; Jingru Fu; Shanshan Wang; David Zhang; Zhaoyang Dong; C L Philip Chen
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2019-11-01       Impact factor: 10.451

6.  Classification of the COVID-19 infected patients using DenseNet201 based deep transfer learning.

Authors:  Aayush Jaiswal; Neha Gianchandani; Dilbag Singh; Vijay Kumar; Manjit Kaur
Journal:  J Biomol Struct Dyn       Date:  2020-07-03
  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.