Next Article in Journal
Efficient Noisy Sound-Event Mixture Classification Using Adaptive-Sparse Complex-Valued Matrix Factorization and OvsO SVM
Previous Article in Journal
Earthquake Probability Assessment for the Indian Subcontinent Using Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Subspace Based Transfer Joint Matching with Laplacian Regularization for Visual Domain Adaptation

Department of Electrical Engineering, National Taipei University of Technology, Taipei 10608, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(16), 4367; https://doi.org/10.3390/s20164367
Submission received: 13 July 2020 / Revised: 30 July 2020 / Accepted: 3 August 2020 / Published: 5 August 2020
(This article belongs to the Section Sensing and Imaging)

Abstract

:
In a real-world application, the images taken by different cameras with different conditions often incur illumination variation, low-resolution, different poses, blur, etc., which leads to a large distribution difference or gap between training (source) and test (target) images. This distribution gap is challenging for many primitive machine learning classification and clustering algorithms such as k-Nearest Neighbor (k-NN) and k-means. In order to minimize this distribution gap, we propose a novel Subspace based Transfer Joint Matching with Laplacian Regularization (STJML) method for visual domain adaptation by jointly matching the features and re-weighting the instances across different domains. Specifically, the proposed STJML-based method includes four key components: (1) considering subspaces of both domains; (2) instance re-weighting; (3) it simultaneously reduces the domain shift in both marginal distribution and conditional distribution between the source domain and the target domain; (4) preserving the original similarity of data points by using Laplacian regularization. Experiments on three popular real-world domain adaptation problem datasets demonstrate a significant performance improvement of our proposed method over published state-of-the-art primitive and domain adaptation methods.

1. Introduction

Indoor–outdoor camera surveillance systems [1,2] are widely used in urban areas, railway stations, airports, smart homes, and supermarkets. These systems play an important role in security management and traffic management [3]. However, cameras with different properties and positions deployed in these systems can create a distribution difference among the capturing images. This difference leads to the system’s poor performance due to considering primitive machine learning algorithms for recognition [4]. For example, if a classifier (or primitive algorithm) is trained from source domain images (taken from a DSLR camera), then the trained classifier will not give as expected results when tested on the images collected from some other target domain (taken from webcam camera). A simple solution to improve the classifier’s performance is that it must only be trained with target domain images. However, there are no labeled images in the target domain in practice, and labeling the target domain images is a time-consuming process. Let us consider an example shown in Figure 1 to discuss in detail how the images collected from different environments can cause differences in distribution across domains. In Figure 1, various possibilities, which can cause distribution differences, are presented, such as (1) the images (keyboards and headphones) as shown in Figure 1a,b are collected from different quality cameras, i.e., low-quality camera (webcam camera) and high-quality camera (DSLR); (2) the images as shown in Figure 1c,d are taken with different weather conditions, i.e., a day or clear weather and night or rainy season.
Recently, the literature [2,4] has seen a growing interest in developing transfer learning (TL) or domain adaptation (DA) algorithms to minimize the distribution gap between domains, so that the structure or information available in the source domain can be effectively transferred to understand the structure available in the target domain. In previous work [5,6,7,8,9,10,11,12], two learning strategies for domain adaptation are considered independently: (1) instance re-weighting [9,10,11,12], which reduces the distribution gap between domains by re-weighting the source domain instances and then training the model with the re-weighted source domain data; (2) feature matching [5,6,8,13,14], which finds a common feature space across both domains by minimizing the distribution gap.
If the distribution difference between both domains is large enough, there will always be a situation where the source domain instances are not relevant to the target domain instance, even after finding a common feature location. In this situation, jointly optimizing instance re-weighting and feature matching is an important and unavoidable task for robust transfer learning. To understand a need for joint-learning instance re-weighting and feature matching more deeply, let us consider an example in which we have source domain data with outlier data samples (or irrelevant instances) as shown in Figure 2a and target domain data as shown in Figure 2b. In this case, if we lean only the common feature space between both domains by existing methods such as Joint Geometrical and Statistical Alignment (JGSA) [8] and Joint Distribution Adaptation (JDA) [6], the new representation of the source and target domain data is shown in Figure 2c, where it can be seen that the domain difference is still large for feature matching due to outlier data samples or irrelevant instances (the symbols with circles). However, if we jointly learn feature matching and instance re-weighting, the data representation is shown in Figure 2d, where it can be seen that all the outlier data samples are down-weighted to reduce domain difference further.
Fortunately, in the literature, there is a method called Transfer Joint Matching (TJM) that performs joint feature matching and instance re-weighting by down-weighting irrelevant features of the source domain [7]. However, only performing feature matching and instance re-weighting is insufficient to successfully transfer knowledge from the source domain to the target domain. Some other DA and TL methods consider other essential properties to minimize distribution differences between both domains. For example, the JDA method considers the conditional distribution in addition to the marginal distribution, and this distribution is needed if the data is conditionally or class-wise distributed. Subspace Alignment (SA) [15] makes use of subspaces (composed of ‘d’ eigenvectors induced by a Principle Component Analysis (PCA)), one for each domain and suggests minimizing the distribution difference between subspaces of both domains rather than the original space data. JGSA preserves source domain discriminant information, among other properties such as SA, marginal, and conditional distributions, to further improve the performance of JDA. However, the feature space obtained by JGSA is not notable because data samples in this space may lose their original similarity so that they can be easily misclassified by the classifier. Kernelized Unified Framework for Domain Adaptation (KUFDA) [16] improves JGSA by adopting the original similarity weight matrix term so that the sample does not lose its original similarity in the learned space. KUFDA follows most of the above discussed important properties but still suffers from outlier data samples, and this is due to not considering instance re-weighting term.
In this paper, to solve all of the above-discussed challenges and to efficiently transfer knowledge from the source domain to the target domain, we propose a novel Subspace based Transfer Joint Matching with Laplacian Regularization (STJML) method for visual domain adaptation by jointly matching the features and re-weighting instances across both the source and the target domains.
The major contributions of this work can be listed as follows:
  • The proposed method STJML is the first framework that crosses the limits of all the comparative cutting edge methods, by considering all inevitable properties such as projecting both domain data into a low dimensional manifold, instance re-weighting, minimizing marginal and conditional distributions, and geometrical structure of both domains in a common framework.
  • With the help of the t-SNE tool, to illustrate the reason for the inclusion of all the components (or inevitable properties), we have graphically visualized the features learned by the proposed method after excluding any component.

2. Related Work

Recently, various DA and TL approaches have been proposed for transferring structure or information from one domain to another domain in terms of features, instances, relational information, and parameters [4,9]. However, the TL approaches, which are closely related to our work, can be divided into three types: feature-based transfer learning [6], instance-based transfer learning [7], and metric-based transfer learning [9].
In the first type, the objective is to minimize the distribution difference between the source and the target domains based on feature learning. For example, Pan et al. [17] proposed a new dimensionality reduction method called maximum mean discrepancy embedding (MMDE) for minimizing the distribution gap between domains. MMDE learns a common feature space on the domains where the distance between distributions can be minimized while preserving data variance. Pan et al. [5] further extended the MMDE algorithm by proposing a new learning method called Transfer Component Analysis (TCA). TCA tries to learn a feature space across domains in a reproducing kernel Hilbert space using Maximum Mean Discrepancy (MMD). Therefore, with the new representation in this feature space, we can apply standard machine learning methods such as k-Nearest Neighbor (k-NN) and Support Vector Machine (SVM) to train classifiers in the source domain for use in the target domain. Long et al. [6] extends TCA by considering not only marginal distribution but also conditional distribution with the help of pseudo-labels in the target domain. Fernando et al. [18] introduced a subspace centric method called Subspace Alignment (SA). SA aims to align the source domain vectors (E) with the target domain one (F) with the help of a transformation matrix (M). Here, E and F can be obtained by Principle Component Analysis (PCA) on the source domain and the target domain, respectively. Shao et al. [19] proposed a low-rank transfer learning method to match both domain samples in the subspace for transferring knowledge. Zhang et al. [8] proposed a unified framework that minimizes the distribution gap between domains both statistically and geometrically, called Joint Geometrical and Statistical Alignment (JGSA). With the help of two coupled projections E (for source domain) and F (for target domain), JGSA projects the source domain and the target domain data into low dimensional feature space, where both domain samples are geometrically and statistically aligned.
In the second type, the objective is to re-weight the domain samples so as to minimize the distribution difference between both domain samples. The TrAdaBoost TL [10] method re-weights the source domain labeled data to filter samples that are most likely not from the target domain. In this way, the re-weighted source domain samples will create the same distribution found on the target domain. Finally, the re-weighted samples can be considered as additional training samples for learning the target domain classifier. As the primitive TrAdaBoost method is applicable to a classification problem, Pardoe et al. [11] extended this method by proposing ExpBoost.R2 and TrAdaBoost.R2 methods to deal with the regression problem.
In the final type, the target domain metric is to be learned by establishing a relationship between the source domain and the target domain tasks. Kulis et al. [20] introduced a method, called ARC-t, to learn a transformation matrix between the source domain and the target domain based on metric learning. Zhang et al. [21] proposed a transfer metric learning (TML) method by establishing the relationship between domains. Ding et al. [22] developed a robust transfer metric learning (RTML) method to effectively assist the unlabeled target learning by transferring the information from source domain labeled data.

3. A Subspace Based Transfer Joint Matching with Laplacian Regularization

This section presents the Subspace based Transfer Joint Matching with Laplacian Regularization (STJML) method in detail.

3.1. Problem Definition

To understand transfer learning or domain adaptation, the domain and the task must be explicitly defined. A domain D consists of two parts: features space X and a marginal probability distribution P ( x ) , i.e., D = { X , P ( x ) } , where x X . If there is a difference in their feature space or marginal distribution, the two domains are said to be different. Given a domain D , a task, that is denoted by T , also consists of two parts: a label space Y and a classifier function f ( x ) , i.e., T = { y , f ( x ) } , where y Y and classifier function f ( x ) predicts label of new instance x. This classifier function f ( x ) can also be interpreted as the conditional probability distribution, i.e., Q ( y | x ) .
Transfer learning, given a labeled source domain D s = { ( x 1 , y 1 ) , , ( x n s , y n s ) } and unlabeled target domain D t = { ( x 1 ) , , ( x n t ) } under the assumptions X s = X t , Y s = Y t , P s ( x s ) P t ( x t ) , Q s ( y s | x s ) Q t ( y t | x t ) , aims to improve the performance of the target domain classifier function f t ( x ) in D t using the knowledge in D s .

3.2. Formulation

To address the limitations of existing TL methods, the STJML method minimizes the distribution gap statistically and geometrically by working on the following components: finding both domain subspaces, matching features, instance re-weighting, and exploiting the similar geometrical property. In our proposed STJML approach, first, we exploit the subspaces of both domains and then with the help of common projection vector-matrix Z for both domains, perform feature matching, instance re-weighting, and the similar geometrical property exploitation in a Reproducing Kernel Hilbert Space (RKHS) to match both first and high-order statistics.

3.3. Subspace Generation

Even though both domain data lie in the same D-dimensional feature space, they are drawn according to different marginal distributions. Consequently, according to [15], instead of working on the original feature space, we need to work on more robust representations of both domain data to allow it to induce stronger classification, which is not subject to local perturbations. For this subspace generation, we use the Principle Component Analysis (PCA) technique, which selects ‘d’ eigenvectors corresponding to the ‘d’ largest eigenvalues. These ‘d’ eigenvectors are used to project original space data on it. For example, if a given input data matrix X = [ x 1 , x 2 , , x n ] R D × n where n = n s + n t and D is the dimension of each data sample in original space, then, the PCA generates the subspace matrix X R d × n by projecting input data matrix ( X ) on selected ‘d’ eigenvectors.

3.4. Feature Transformation

As a dimensionality reduction method, such as PCA, can learn the transformed feature representation by reducing the reconstruction error of given data, it can also be utilized for the data reconstruction. Let us consider subspace data matrix X R d × n , data centering matrix H = I 1 n 1 , and 1 is a n × n matrix of ones. Thus, the covariance matrix of both domain subspace data matrix X can be calculated as X H X T . The objective of PCA is to maximize both domain variances by finding an orthogonal transformation matrix W R n × σ , where σ is the selected number of eigenvectors on which subspace data matrix X to be projected. Thus,
Max W T W = I tr ( W T X H X T W )
where tr(.) is the trace of a matrix and I is an identity matrix. As the problem in Equation (1) is an eigendecomposition problem, it can easily be decomposed by eigendecomposition as X H X T W = W Φ , where Φ = diag ( Φ 1 , , Φ σ ) is the σ largest eigenval matrix. After projecting subspace data matrix X on the selected projection vectors matrix ( W σ ) corresponding to top most σ largest eigenvalues matrix, the optimal σ -dimensional learned projection matrix V = [ v 1 , , v σ ] = W σ T X
To achieve our goal, we need to work in RKHS using some kernel function like linear, polynomial, Gaussian, etc. Let us consider the chosen kernel function θ , which maps the data sample x to θ ( x ) , i.e., θ : x θ ( x ) , or θ ( X ) = [ θ ( x 1 ) , , θ ( x n ) ] , and then the kernel matrix K = θ ( X ) T θ ( X ) R n × n . After the application of the Representer theorem W = θ ( X ) Z , Equation (1) can be written as follows:
Max Z T Z = I tr ( Z T K H K T Z )
where Z R n × σ is the transformation matrix, and the subspace embedding becomes V = Z T K .

3.4.1. Feature Matching with Marginal Distribution

However, even though maximizing the subspace data variance, the distribution difference between both domains will still be quite large. Therefore, the main problem is to minimize the distribution difference between them by applying an appropriate distance metric (or measure). There are many distance measures (such as the Kullback–Leibler (KL) divergence) that can be utilized to compute the appropriate distance between both domain samples. However, many of these methods are parameterized or require estimating the intermediate probability density [5]. Therefore, in this paper, we adopt a non-parametric distance estimate method called Maximum Mean Discrepancy (MMD) [23] to compare distribution difference in a Reproducing Kernel Hilbert Space (RKHS) [5]. MMD estimates the distance between the sample means of both domain data in the σ -dimensional embedding,
M d = 1 n s i = 1 n s Z T k i 1 n t j = n s + 1 n Z T k j 2 = tr ( Z T K M d K T Z )
where M d is the MMD matrix and can be determined as follows
M i j d = 1 n s n s k i , k j D s 1 n t n t k i , k j D t 1 n s n t otherwise

3.4.2. Feature Matching with Conditional Distribution

Minimizing the marginal distribution difference does not guarantee that the conditional distribution between both source and target domains will also be minimized. However, for robust transfer learning, minimizing the conditional distributions, i.e., Q s ( y s | x s ) and Q t ( y t | x t ) between both domains is required [24]. Reducing the conditional distribution is not a trivial process because there is no label data in the target domain. Therefore, we cannot model Q t ( y t | x t ) directly.
Long et al. [6] proposed a Joint Distribution Adaptation (JDA) method for modeling Q t ( y t | x t ) by generating p s e u d o labels of the target data. Initial p s e u d o labels for the target data can be generated by training the classifier with X s and Y s of the source domain, and testing the classifier on target domain subspace X t . Now with Y s , X s , and p s e u d o labels, the conditional distribution between both domains can be minimized by modifying MMD to estimate distance between the class conditional distributions Q s ( x s | y s = c { 1 , , C } ) and Q t ( y t | x t = c { 1 , , C } )
C d = 1 n s c k i D s c Z T k i 1 n t c k j D t c Z T k j 2 = tr ( Z T K M c K T Z )
where D s c = { k i : k i D s y ( k i ) = c } is the set of samples belongs to cth class in the source domain, y ( k i ) is the true label of k i , and n s c = | D s c | . Similarly, for the target domain, D t c = { k j : k j D t y ^ ( k j ) = c } is the set of samples belongs to cth class in the target domain, y ^ ( k j ) is the p s e u d o label of k j , and n t c = | D t c | . Thus, the MMD matrix M c with class labels of both domains can be determined as follows:
M i j c = 1 n s c n s c k i , k j D s c 1 n s c n t c k i , k j D t c 1 n s c n t c k i D s c k j D t c k j D s c k i D t c 0 otherwise
By minimizing Equation (5) such that Equation (2) is maximized, the conditional distributions between both the source and the target domains are drawn close with the new representation V = Z T K . In each iteration, this representation V will be more robust till its convergence. As there is a difference in both the marginal and conditional distributions, the initial p s e u d o labels of the target domain are incorrect. However, we can still take advantage of them and improve the performance of target domain classifiers iteratively.

3.5. Instance Re-Weighting

However, matching features with marginal and conditional distributions is not sufficient for transfer learning, as it can only match first- and higher-order statistics. In particular, when the domain difference is significant enough, even in the feature learning, there will always be some source instances or samples that are not related to the target instance. In this condition, an instance re-weighting method with feature learning should also be included to deal with such a problem.
In this paper, we adopt a L 2 , 1 -norm structured sparsity regularizer as proposed in [7]. This regularizer can introduce row-sparsity to the transformation matrix Z. Because each entry of the matrix Z corresponds to an example, row sparsity can substantially facilitate instance re-weighting. Thus, instance re-weighting regularizer can be defined as follows.
I r = Z s 2 , 1 + Z t F 2
where Z s : = Z 1 : n s is the transformation matrix corresponding to the source samples, and Z t : = Z n s + 1 : n s + n t is the transformation matrix corresponding to the target samples. As the objective is to re-weight the source domain instances, we only impose L 2 , 1 -norm on source domain. Thus, minimizing the Equation (7) such that Equation (2) is maximized, the source domain samples, which are similar (or dissimilar) to the target domain, are re-weighted with less (or greater) importance in the new learned space V = Z T K .

3.6. Exploitation of Geometrical Structure with Laplacian Regularization

However, matching features and instance re-weighting are not enough to convey knowledge transfer by capturing the intrinsic structure of the source domain labeled samples and target domain unlabeled samples. In particular, labeled data samples of the source domain combined with unlabeled data samples of the target domain are used to construct a graph that sets the information of the neighborhood data samples. Here, the graph provides discrete approximations to the local geometry of the manifold data. With the help of the Laplacian regularization term L , the smooth penalty on the graph can be included. Basically, the term regularizer L allows us to incorporate prior knowledge on certain domains, i.e., nearby samples are likely to share same class labels [25].
Given a kernelized data matrix K, we can use a n n -nearest neighbor graph to establish a relationship between nearby data samples. Specifically, we draw an edge between any two samples i and j if k i and k j are “close”, i.e., k i and k j are among n n nearest neighbors of each other. Thus, the similarity weight matrix W can be determined as follows:
W i j = 1 if k i N n n ( k j ) or k j N n n ( k i ) 0 otherwise
where N n n ( k j ) represents the set of n n nearest neighbors of k i .
Here, two data samples are connected with an edge if they are likely to be from the same class. Thus, the regularizer term L can be defined as follows:
L = i j ( Z T k i Z T k j ) 2 W i j = 2 i Z T k i D i i k T Z 2 i j Z T k i W i j k j T Z = 2 Z T K ( D W ) K T Z = 2 Z T K L K T Z
where D is the diagonal matrix, i.e., D i i = j W i j and L is the Laplacian matrix; L = D W ,

3.7. Overall Objective Function

The objective of this work is to minimize the distribution difference between domains by jointly matching the features of both domains and re-weighting the source domain samples, and preserving original similarity of both domain samples. So, by incorporating Equations (3), (5), (7), and (9), the proposed objective function can be obtained as follows:
max Z T K H K T Z 1 tr ( Z T K ( ( δ 1 ) M d + δ C d + η L ) K T Z ) + λ ( Z s 2 , 1 + Z t F 2 )
where δ is a trade-off parameter, which balances the marginal and conditional distributions [13], η is the trade-off parameter that regularizes the Laplacian term, and λ is the regularization parameter to trade-off feature matching and instance re-weighting.

3.8. Optimization

By using the Lagrange multiplier Φ , Equation (10) can be written as follows:
L f = tr ( Z T K ( ( δ 1 ) M d + δ C d + η L ) K T Z ) + λ ( Z s 2 , 1 + Z t F 2 ) + tr ( ( I Z T K H K T Z ) Φ )
In order to find out an optimal value of the projection vector matrix Z, we partial derivative L f with respect to Z and equate it to zero as
( K ( ( δ 1 ) M d + δ C d + η L ) K T + λ G ) Z = K H K T Z Φ
Z s 2 , 1 is a non-smooth function at zero and its partial derivative can be computed as ( Z s 2 , 1 + Z t F 2 ) Z = 2 G Z , where G is a diagonal subgradient matrix and its i th element can be calculated as
G i i = 1 2 a i , k i D s , a i 0 0 , k i D s , a i = 0 1 , k i D t
As the problem in Equation (12) is a generalized eigen decomposition problem, we can solve it and find Φ = d i a g ( ϕ 1 , , ϕ σ ) ( σ leading eigenvalues) and Z = ( z 1 , , z σ ) ( σ leading eigenvectors). The pseudo code of our proposed method is given in Algorithm 1.

4. Experiments

In this section, we present a performance of the proposed STJML method by experimenting on various visual domain classification problems.

4.1. Data Preparation

We have considered three publicly image datasets: Office + Caltech10 with Speeded Up Robust Features (SURF), Office + Caltech10 with VGG-FC6 features, and Pose, Illumination, and Expression (PIE) face Recognition for experimentation. These data sets are well known in the domain adaptation methods and are widely considered in most recent works (such as [25,26]).
Caltech-256 has 30,607 images and 256 classes, while Office-31 is made of three object domains: DSLR (D), Amazon (A), and Webcam (W). It contains a total of 4652 images with 31 classes. As images in Office and Caltech-256 having different distributions, DA methods can help with cross-domain recognition. Since both the datasets contains 10 common classes, we considered Office + Caltech 10 datasets from [8], which has 12 tasks: A D , , C W . For the purpose of experimentation, we considered both the SURF feature and Deep feature (VGG-FC6 features) of this dataset.
Carnegie Mellon University (CMU) PIE (Pose, Illumination, and Expression (PIE)) face dataset [27] contains over 40,000 facial images of 68 people. The images of each person were taken across 13 different poses, under 43 different illumination conditions, and with 4 different expressions. As there are many datasets of different poses, we considered only five poses such as C05, C07, C09, C27, and C29 for experimentation. Here each pose contains images with illumination variation and expression variation. Similar to Office + Caltech 10 datasets, 20 possible combinations of source and target domains or tasks such as C 05 C 07 , , C 29 C 27 can be constructed.
In this paper, we use notation P Q to show knowledge transfer from source domain P to the target domain Q.

4.2. t-SNE Representation of Feature Spaces Learned by the Proposed Method (STJML)

In order to visualize the learned feature space by our proposed method, we considered the t-SNE tool [28], through which high dimensional data is projected to 2-D space (or low dimensional space). To show t-SNE representation of feature spaces for tasks A D (SURF features) and A W (VGG-FC6 features), we randomly selected 150 samples from each domain and then used two different symbols (such as circles and pluses) to represent different domains and 10- different colors (‘black’, ‘red’, ‘lime’, ‘blue’, ‘orange’, ‘cyan’, ‘magenta’, ‘green’, ‘chocolate’, and ‘maroon’) to represent 10 different classes. Furthermore, to clearly understand the distribution differences between both domains, we used different colored ellipses to represent different classes’ variances belonging to different domains. We also used different colored lines to indicate the distribution gap between same class samples belong to different domains and different symbols such as square and star to represent the average point (or mean point) of each class in the source and target domains. For example, Figure 3a shows the initial feature representation of A D task with SURF feature, where it can be seen that different class samples from different domains are too close together or there is no uniform cluster for different classes. Therefore, the classification or clustering algorithms can easily misclassify the samples that are too close or near the edge of their own clusters. However, due to the recent advancement in deep learning approaches, by which we can obtain deep features like VGG-FC6 features for the Office + Caltech10 dataset. The deep features (VGG-FC6) representation of task A W for the Office + Caltech10 dataset is shown in Figure 3b. After comparing the representation of both types of features (as shown in Figure 3a), i.e., VGG-FC6 and SURF, it can be seen that the representation of VGG-FC6 features is much better than the SURF features (Figure 3a). Therefore, the performance of the primitive machine learning algorithm is better for deep learning features. The t-SNE representation of feature spaces learned by the proposed method (STJML) for both the tasks A D (SURF features) and A W (VGG-FC6 features) is shown in Figure 4.
To quantify misclassification samples in the learned feature space by the proposed method STJML for the task ( A D ), we have shown two illustrations such as the first one with the predicted class labels for the target domain (as shown in Figure 4a) and the second one with the given class labels for both domains(as shown in Figure 4b). After carefully analyzing both the graphs, i.e., Figure 4a,b, many samples (highlighted by asterisks (*) and arrows (→))) are being misclassified by the proposed method STJML. However, If we compare the graph (as shown in Figure 3a) with the graph (as shown in Figure 4a), it can be seen that the distribution difference between the source domain samples and the target domain samples is minimized by a small margin. For example, in Figure 3a, it is visible that there is a distribution gap between the red class samples of the circle domain (or source domain) and the red class samples of the plus domain (or target domain). But, in Figure 4a, it can be seen that our proposed method STJML minimizes this distribution gap.
Similar to A D (SURF features) task, if we compare graphs in Figure 3a and Figure 4c for A W (VGG-FC6 features), it is observed that the distribution difference between both domains is satisfactorily reduced by the proposed method STJML. However, after comparing Figure 4c to Figure 4d, it can also be seen that there are only a few samples (marked by asterisks (*) and arrows (→))) which are being misclassified by our proposed method STJML.

4.3. What Happens if One Component Is Omitted from the Proposed Method (STJML)

To reveal the importance of including all of the above discussed components in our proposed method, we have experimented our proposed method on tasks A D (SURF feature) and A W (VGG-FC6 feature) by omitting any of its components at once. Therefore, by excluding any one component from our proposed method STJML, we can divide the proposed method into five new methods: STJML s (Omitting subspaces of both domains), STJML w (Omitting instance-re-weighting term), STJML m (Omitting marginal distribution term), STJML c (Omitting conditional distribution term), and STJML l (Omitting Laplacian regularization term).

4.3.1. Omitting Consideration of Subspaces of Both Domains ( STJML s )

If we execute STJML s on the original VGG-FC6 features of the task A W , the learned feature representation is shown in Figure 5. In Figure 5, the first graph (i.e., Figure 5a) shows feature representation learned by method STJML s with the given source domain labels and predicted target domain labels, while the second graph (i.e., Figure 5b) presents a representation of learned features with given both domain label information.
By comparing the representation of features for the task A W (as shown in Figure 3b) with that of features learned by STJML s (as shown in Figure 5a), the feature space learned by STJML s is much better. From the comparison, it can also be seen that the distribution difference between both domains is minimum, as well as the distance between samples, belong to the same class is minimum, while the distance between different class samples is maximum. Thus, if this learned feature space is given to a classification algorithm, such as 1-NN classifier, the performance of the classifier (in terms of accuracy) will be 86.44%, which is accompanied by a 23% gain over the performance of the trained classifier with original feature space. Although the learned feature space is much better than the original feature space, there are some samples that are still being misclassified. In order to investigate those samples that are being misclassified by STJML s method, we have also visualized these learned features (shown in Figure 5b) with the given both domain labels. If we compare the graphs as shown in Figure 5a,b, it can be observed that some sample class labels (as indicated with asterisks(*) and arrows(→)) predicted by the STJML s are incorrect. Again if we compare the learned feature space (as shown in Figure 5a) by STJML s method with the learned feature space (as shown in Figure 4c) by the proposed method STJML, the clusters for different classes obtained by our proposed method with all components are slightly better than those obtained by STJML s method. For example, a black class cluster obtained by the STJML method is slightly distant from the maroon class samples, but that obtained by the STJML s method collides with the samples with the maroon class. Similarly, the orange class cluster obtained by the STJML s method is the worst compared to that obtained by the STJML method.

4.3.2. Omitting Instance-Re-Weighting Term ( STJML w )

Since, the results obtained by both STJML w and STJML methods for the task A W (VGG-FC6 features) were similar, we considered another task A D (SURF features) to show the effect of instance-re-weighting term. After executing STJML w method on the task A D (SURF features), the learned feature spaces are shown in Figure 6a,b. After comparing both the plots (as shown in Figure 6a,b), we can see that there are many samples (some of them we have highlighted by asterisks (*) and arrows (→)) which are being misclassified by the STJML w method.
As the clusters obtained in Figure 6a,b by STJML w method for the task A D (SURF features) are not as good as obtained by the proposed method for the task A W (VGG-FC6 features), we compare the graph (as shown in Figure 6b) obtained by STJML w method with the graph (as shown in Figure 4b) obtained by the STJML method. After comparing both the plots in Figure 4b and Figure 6b, it can be concluded that some of the samples of source domain for ’lime’ colored class in Figure 6b (look at ‘lime’ colored ellipse) are not efficiently down weighted as compared to the graph in Figure 4b. Therefore, the performance of STJML w method (which is 41.70% accuracy) is not as good as the STJML method (which is 49.10%).

4.3.3. Omitting Marginal Distribution Term ( STJML m )

If we omit consideration of the marginal distribution from our proposed method STJML, the t-SNE views of learned feature spaces by STJML m for A W (VGG-FC6 features) task are shown in Figure 7a,b. After excluding the marginal distribution, the STJML m method achieves 90.85% accuracy, which is similar to the accuracy achieved by the STJML method. Thus, it can be concluded here that even after dropping this distribution, the STJML method does not have much effect on its performance. Moreover, after carefully looking the graphs learned by STJML m (as shown in Figure 7a,b) and STJML (as shown in Figure 4c,d) methods, we find that the graphs learned by both the methods are almost similar.

4.3.4. Omitting Conditional Distribution Term ( STJML c )

If we exclude consideration of the conditional distribution from our proposed method STJML, the t-SNE views of learned feature spaces by STJML c for A W (VGG-FC6 features) task are shown in Figure 8a,b. Without including this conditional distribution term in our proposed method STJML, the STJML c approach achieves 73.90% accuracy, which is much lower than the accuracy (90.85%) achieved by the STJML method. Therefore, we can say that this term greatly impacts the performance of the proposed STJML method if it is not included. If we compare the graph (as shown in Figure 4c) obtained by the STJML method to the graph (as shown in Figure 8a) obtained by STJML c , it can be seen that the distribution difference between both domains has not been effectively reduced by the STJML c method. For example, the distribution difference between green colored class samples of both domains (plus and circle) is not minimized in Figure 8a (i.e., all the green colored class samples are distributed in different green colored circles), but it can be seen in Figure 4c that all green colored class samples are with in a cluster. Because of not minimizing the distribution gap by the STJML c method, we can see that some samples that are being misclassified in Figure 8a (as highlighted by asterisks (*) and arrows (→)) after comparing with the graph in Figure 8b.

4.3.5. Omitting Laplacian Regularization Term ( STJML l )

The samples that are supposed to loss their original similarity in the leaned feature space can preserve their original similarity by adding the Laplacian regularization term. As a result, samples that were supposed to go far away from their respective groups or clusters may come closer together. Thus, in order to see the impact of this term, we omit this term from the proposed method STJML and execute the algorithm. The t-SNE representation of learned feature spaces by STJML l is depicted in Figure 9a,b. If we compare the graphs generated by STJML l (as depicted in Figure 9a,b) with the graphs generated by STJML (as depicted in Figure 4c,d), it can be seen that the samples in each class cluster generated by STJML l are widely spread around their mean point, but they are less spread in the cluster generated by STJML. Therefore, the performance (82.03% accuracy) of STJML l method is slightly lower than the STJML method. After comparing the graphs in Figure 9a,b, we can see that some samples (highlighted by asterisks (*) and arrows (→))) are being misclassified by STJML l method.

4.4. Comparison with State-Of-The-Art Methods

The proposed STJML method was verified and compared with many state-of-the-art primitive and domain adaptation algorithms. A brief description of all the comparative methods is as follows:
  • NN, PCA+1NN, and SVM: These are the traditional machine learning algorithms, which assume that both training and test data should follow a uniform distribution.
  • Transfer Component Analysis (TCA) [5]: TCA is a feature transformation technique that aligns only the marginal distribution of both domains.
  • Joint Distribution Alignment (JDA) [6]: This method aligns both the marginal and conditional distributions of both domains.
  • Geodesic Flow Kernel (GFK) [29]: In order to characterize the changes of geometric and statistical properties from the source domain to the target domain, GFK integrates countless subspaces.
  • Transfer Joint Matching (TJM) [7]: This method demonstrates both feature learning and instance re-weighting for minimizing distribution differences between both domains.
  • Subspace Alignment (SA) [15]: SA first projects both domain samples into lower-dimensional subspaces, then aligns both domains.
  • Scatter Component Analysis (SCA) [30]: This method is based on a simple geometrical measure, i.e., scatter.
  • CORAL [31]: It aligns both domain covariance matrices.
  • Adaptation Regularization (ARTL) [14]: This method learns domain classifier in original space.
  • Cross-Domain Metric Learning (CDML) [32]: It is a novel metric learning algorithm to transfer knowledge in an information-theoretic setting.
  • Close yet Discriminative DA(CDDA), Geometry Aware DA (GA-DA), and Discriminative and Geometry Aware DA (DGA-DA) [33]: CDDA enhances the performance of the JDA method by incorporating a new repulsive force objective into its model to improve the discriminative power of the common feature subspace. Likewise, the GA-DA method includes the original parity of a data point in CDDA to improve its performance. Then, finally, the GA-DA method involves preserving the discriminative information term to improve the DGA-DA method performance further.
  • Invariant Latent Space (ILS) [34]: This TL method makes use of the Riemannian optimization methods to match statistical properties.
  • Balanced Distribution Adaptation (BDA) [35]: It is a novel TL approach to adaptively balance both the marginal and conditional distributions of both domain data.
  • Joint Geometrical and Statistical Alignment (JGSA) [8]: It extends the JDA by adopting two projection vector matrices and considering subspace alignment and source discriminant information.
  • Robust Transfer Metric Learning (RTML) [22]: This method considers two directions, i.e., sample space and feature space, to mitigate the distribution gap.
  • Domain Invariant and Class Discriminative (DICD) [36]: This DA method is to learn a latent feature space while preserving important data properties.
  • Explicit Map-based Feature Selection (EMFS) [37]: It attempts to: (1) reveal high-order invariant features by explicit feature map, (2) integrate feature learning and model learning, and (3) remove non-discriminative features from invariant features.
  • Domain Irrelevant Class clustering (DICE) [38]: This method specifically deals with the intra-domain structure for the target domain in addition to other common properties.
  • Linear Discriminant Analysis-inspired Domain Adaptation (LDADA) [39]: The key insight of this approach is to leverage the discriminative information from the target task, even when the target domain labels are not given.
  • Kernelized Unified Framework for Domain Adaptation (KUFDA) [16]: This TL method improves the JGSA method by adding the Laplacian regularization term.

4.5. Parameter Sensitivity

Our proposed STJML method contains various parameters such as n n , k , σ , η , λ , and δ , along with other state-of-the-art domain adaptation methods [5,8,13,16]. Similar to previous methods [14,40], we also need to analyze the parameter sensitivity of the STJML method on all possible tasks of both datasets to validate that an appropriate value of each parameter can be chosen to obtain satisfactory performance. Analyzing the parameter sensitivity of STJML, we vary one parameter value and keep the other parameter values constant. For example, we vary parameter value k from 1 to 10 with an interval of 1 and keep other parameter values n n = 1 , σ = 100 , η = 10 1 , λ = 10 3 , and δ = 0.5 constant. Here, we have provided a description of each parameter and performed a parameter sensitivity test for all the considered datasets. But, we have shown parameter sensitivity analysis graphs for Office + Caltech10 with VGG-FC6 features and PIE face datasets. The description and possible values of each parameter are as follows:

4.5.1. k Parameter

In our proposed method, we considered the k-NN classifier to predict the label of the target domain, and the performance of this classifier depends on an appropriate value of parameter k for each task. Therefore, we need to find out its proper value for each task of the datasets. For each task of the considered datasets, we varied k value from 1 to 10 with an interval of 1 and keep other parameter values constant as shown in Figure 10a,b. The resultant graphs for Office + Caltech10 with VGG-FC6 features and PIE face datasets are shown in Figure 10a,b. From Figure 10a,b, it can be seen that the STJML method outperforms at k = 1 for most of the tasks of both the datasets. Therefore, we keep k = 1 for most of the tasks of all the datasets except some tasks such as 9 7 ,   9 27 ,   9 29 ,   27 7 ,   27 09 ,   C D ( VGG−FC6 features task ) ,   C A ( SURF features task ) ,   C W ( SURF features task ) ,   D A ( SURF features task ) ,   D C ( SURF features task ) ,   W A (SURF features task), but for these tasks k-parameter values are kept 9, 9, 5, 10, 10, 9, 3, 3, 10, 3, and 2, respectively.

4.5.2. n n Parameter

Similar to the parameter k, we require an appropriate value of parameter n n for construing the Laplacian graph as discussed in Section 3.6. So, we vary n n value from 1 to 10 and keep other parameter values constant. Here also, in Figure 10c,d, there is no unique value of n n for which STJML is outperforming for all tasks of Office + Caltech with the SURF features dataset. However, STJML is outperforming for the values (1 and 2) of n n for the tasks of PIE dataset. Therefore, we keep n n = 1 for all tasks (except 5 27 ( n n = 2 ) , 7 5 ( n n = 2 ) , and 7 09 ( n n = 2 ) ) of PIE face dataset. Similarly, we keep n n = 3 for all tasks (except A C ( n n = 1 ) and A W ( n n = 7 ) ) of Office + Caltech with VGG-FC6 features dataset, but n n = 10 for all tasks (except A D ( n n = 1 ) , A W ( n n = 1 ) , C A ( n n = 8 ) , C D ( n n = 2 ) , and W C ( n n = 1 ) ) of Office + Caltech with SURF features dataset.

4.5.3. δ Parameter

This parameter quantitatively evaluates the importance of aligning marginal and conditional distributions in domain adaptation. In this evaluation, the existing work [6,8] in DA fails by the assumption that both distributions are equally important. However, this assumption may not be true for real-world problems. Wang et al. [13] introduced the adaptive factor parameter to measure the importance of these two distributions dynamically. However, in this paper, we manually performed the parameter sensitivity tests to ascertain a reasonable value of this factor for each task. Thus, we varied its value from 0 to 1 with an interval of 0.1, and the resulted graphs are shown in Figure 11. It is clearly shown in Figure 11, that STJML performs well for different values of δ for different tasks. Thus, to achieve best performance of the STJML method, we keep δ = 0.5 for all tasks (except 5 7 ( δ = 0.9 ) and 7 29 ( δ = 0.6 ) ) of PIE face dataset. Similarly, we keep δ = 0.5 for all tasks (except A W ( δ = 0.9 ) ) of Office + Caltech with VGG-FC6 dataset, but δ = 0.9 for all tasks (except C W ( δ = 0.9 ) ) of Office + Caltech with SURF features dataset.

4.5.4. Parameter: λ and η

As parameter λ is the trade-off parameter which regularizes the instance re-weighting term, we vary its values from 10 9 to 10 2 and keep other parameter values constant as shown in Figure 12a,b, where the x-axis represents numbers from ‘−n’ to ‘n’, which are actually numbers from 10 n to 10 n . In Figure 12a,b, it is shown that STJML performs well for 10 4 and 10 3 for most of the tasks of both datasets. For better accuracy with respect to this parameter, we keep λ = 10 3 for all tasks (except 5 7 ( λ = 10 7 ) , 5 9 ( λ = 10 6 ) , 7 29 ( λ = 10 7 ) , and 27 29 ( λ = 10 6 ) ) of PIE face dataset. But, we keep λ = 10 4 for all tasks (except A W ( λ = 10 1 ) ) of Office + Caltech with VGG-FC6 features dataset and λ = 10 1 for all tasks (except A D ( λ = 10 3 ) , A W ( λ = 10 2 ) , D A ( λ = 10 4 ) , D C ( λ = 10 3 ) , W C ( λ = 10 3 ) , and W D ( λ = 10 3 ) ) of Office + Caltech with SURF features dataset.
Similar to the parameter λ , we also vary η parameter values from 10 4 to 10 4 and keep other parameter values constant as shown in Figure 12c,d. From Figure 12c,d, it can be determined that STJML outperforms at 10 1 for all tasks of PIE face dataset, while at 10 1 for some tasks of Office + Caltech dataset with VGG-FC6 features. For better performance, we keep η = 10 1 for all tasks (except C A ( η = 10 0 ) and W A ( η = 10 2 ) tasks of VGG-FC6 features, and D W ( η = 10 0 ) and W C ( η = 10 0 ) tasks of SURF features) of considered datasets.

4.5.5. Parameter: σ

As the performance of the proposed STJML method depends on choosing the eigenvectors ( σ ) corresponding to the leading eigenvalues, we ran STJML with varying values of σ (70 to 130 with an interval of 5 for PIE dataset and 9 to 30 with an interval of 3 for Office + Caltech dataset ) and report the results in Figure 13a,b. We plot classification accuracy graph with respect to different values of σ in Figure 13a,b. In Figure 13a,b, it can be seen that proposed STJML the gives best accuracy for different values of this parameter for different tasks of both the datasets. For better accuracy with respect to this parameter, we keep σ = 100 for all tasks (except 5 29 ( σ = 115 ) ,   7 27 ( σ = 75 ) ,   9 05 ( σ = 105 ) ,   27 05 ( σ = 85 ) , and 27 09 ( σ = 80 ) ) of PIE face dataset. We keep σ = 30 for all tasks of Office + Caltech with SURF features dataset, and σ = 14 for all tasks (except A C ( σ = 27 ) ,   A D ( σ = 20 ) ,   A W ( σ = 20 ) ,   D W ( σ = 30 ) ,   D C ( σ = 13 ) , and W A ( σ = 13 ) of Office + Caltech with VGG-FC6 features dataset.

4.5.6. Parameter: d

In order to find a low dimensional subspace for both the domains, we need to project original data from D-dimensional space to a d-dimensional subspace. However, projecting data from D-dimensional space to a d-dimensional subspace, it can lose some information. Therefore, we need to find the appropriate value of d, so that the original information of both domains can remain in the low dimensional space as well.
Like other parameters, we also vary its value and keep other parameter values constant, and find that the proposed STJML method outperforms at d = 140 for all tasks of the PIE face dataset, while d = 100 for SURF features and d = 200 for VGG-FC6 features of the Office + Caltech datasets.

4.6. Experimental Setup

To show the strength of the STJML method over previous state-of-the-art methods, we considered 12 tasks of Office + Caltech10 with SURF feature, 12 tasks of Office + Caltech10 with VGG-FC6 features, and 20 tasks of PIE face datasets. With the help of the parameter sensitivity test, we explored an appropriate value of each parameter of the STJML method and then used those values to ascertain the proposed method’s accuracy for each task of the considered dataset. Thus, after experimenting on the proposed STJML method, the accuracy of each task of all datasets is stated in Table 1, Table 2 and Table 3. However, for the accuracy of other comparative methods in Table 1, Table 2 and Table 3, we have derived directly from their respective papers or previous papers [5,7,8,13,14,16,37,38].

4.7. Experimental Results and Analysis

The recognition performance of the proposed method and the other compared state-of-the-art methods on three widely used domain adaptation datasets is reported in Table 1, Table 2 and Table 3. From the results reported in Table 1, Table 2 and Table 3, we can conclude the following observations:
  • Primitive machine learning approaches such as NN, PCA, and SVM are not performing well due to the distribution gap between training (source data) and testing (target data) datasets.
  • Among domain adaptation methods, the GFK method’s performance is worse for an average accuracy of all tasks in the Office + Caltech dataset for both SURF and VGG-FC6 features.
  • The JDA method’s performance for all the three datasets is higher than that of the TCA method because it adopts the conditional distribution in addition to the marginal distribution.
  • The ILS method works well compared to other subalignment methods (such as SA, GFK, and CORAL) because of considering the more robust discriminative loss function for the Office + Caltech dataset with deeper features.
  • As TJM adopts the term instance re-weighting, its performance is better than other DA methods such as TCA, GFK, JDA, SA, CORAL, ILS, and BDA for the deep features of Office + Caltech dataset. However, for the SURF features, TJM gives better average accuracy than SCA, ARTL, GFK, and TCA, but performs poorly compared to JGSA, CORAL, LDADA, DICE, RTML, ILS, and JDA.
  • The average accuracy (65.09%) of the DGA-DA method for all tasks in the PIE face dataset is higher than that of other methods (i.e., TCA, JDA, CDML, TJM, TDA-AL, CDDA, BDA, RTML, EMFS, and LDADA) because it involves novel repulsive force term and the Laplacian regularization term.
  • Since JGSA improves JDA by considering two projection vector matrices and preserving source domain discriminant information, the mean accuracy (82.60%) of the JGSA method for all tasks in the Office + Caltech dataset with deep features is higher than other methods (i.e., TCA, JDA, GFK, SA, TJM, ILS, BDA, DICE, and CDDA). Similarly, for the Office + Caltech dataset with SURF features, it performs well compared to all other comparative methods except KUFDA and DICE.
  • As KUFDA improves JGSA by considering the term Laplacian regularization, its average accuracy is much higher than other methods for the deep features of the Office + Caltech dataset, but less than the DICE method for most of the tasks in the PIE face dataset.
  • For the PIE face and the Office + Caltech10 with SURF features datasets, DICE is performing better than all other methods (except STJML) because it is taking care of the intra-domain structure, especially for the target domain. However, its performance is abysmal for deep features of the Office + Caltech 10 dataset.
  • Since our proposed method covers all the important objectives, as well as works on the projected subspaces of both the domains, the average accuracy of the proposed STJML method for all the tasks in all the considered datasets, is higher than all the other comparative methods. However, KUFDA beats our proposed algorithm for some tasks in the Office + Caltech dataset with deep features such as A C ,   D W ,   W A ,   W C , and W C . Similarly, DICE beats the proposed method for eight tasks of the PIE face dataset.

4.8. Computational Complexity

Here, we analyze the computational complexity of Algorithm 1 by the O . The computational cost is detailed as follows: O ( D 3 + n D d ) for finding subspaces of both the source and the target domains, where D < n , i.e., Line 1; O ( n 2 d ) for constructing Laplacian matrix, i.e., Line 2; O ( n 2 d ) for computing kernel matrix, i.e., Line 3; O ( n 2 ) for generating initial pseudo labels, i.e., Line 5; O ( t ( d n 2 + C d n 2 ) ) for constructing marginal and conditional distribution matrices, i.e., Line 7; O ( t n 2 d ) for solving the generalized eigendecomposition problem with dense matrices, i.e., Line 9; O ( t ( σ n n s + σ n n t ) ) for computing X s and X t matrices, i.e., Line 11; O ( t n 2 ) for generating pseudo labels, i.e., Line 12; O ( t n 2 ) for computing the subgradient matrix, i.e., Line 13. In total, the computational complexity of Algorithm 1 is O ( D 3 + n D d + n 2 d + t ( d n 2 + C d n 2 + σ n n s + σ n n t + n 2 ) ) . The complexity of this model can be greatly reduced by low-rank approximation.
Algorithm 1: Subspace based Transfer Joint Matching with Laplacian Regularization (STJML)
Sensors 20 04367 i001

4.9. Running Time Analysis

To analyze the actual running time taken by each algorithm, we considered three tasks for the experiment such as C A , C D , and C W , and reported the total time taken by the proposed and some other comparative methods to execute all the three tasks in Table 4. If we use the 1-NN classifier for the experiment, then the running time taken by it is minimal due to not performing any feature transformation. As TCA first presents feature transformation and then uses the 1-NN classifier for classification, it costs slightly more than the 1-NN classifier. Since GFK adds infinite subspaces for feature changes, it takes a longer time than the TCA method. JDA method improves the performance of TCA by considering the conditional distribution. To calculate the conditional distribution, we require a pseudo label for the target domain till the tth iteration. Therefore, its computational cost is higher than the TCA and GFK methods. The TJM method enhances JDA performance by considering additional instance-weighting terms. Thus, the computation cost of TJM also increases compared to JDA. Since the CORAL method first finds the co-variance matrices for both domains and then aligns the source domain to the target domain, its computation cost is higher than that of TJM, JDA, GFK, and TCA. JGSA includes two additional terms such as discriminative terms and subalignments and also uses two projection vector matrices for both domains. Therefore, JGSA consumes much time compared to other TJM, JDA, GFK, CORAL, and TCA. Since our proposed method improves the TJM by considering the subspace of both domains as well as adding the Laplacian term, it has a higher cost than TJM, but lower than JGSA.

5. Conclusions and Future Work

In this paper, we proposed a novel Subspace based Transfer Joint Matching with Laplacian Regularization (STJML) method for efficiently transferring knowledge from the source domain to the target domain. Because of jointly optimizing all the inevitable components, the proposed STJML method is robust for reducing the distribution differences between both domains. Extensive experiments on several cross-domain image datasets suggest that the STJML method performs much better than state-of-the-art primitive and transfer learning methods.
In the future, there are several ways through which we can extend our proposed method STJML, and some of them are:
Firstly, we will extend the STJML method to multi-task learning environments [42], where multiple tasks may contain some label samples. Thus, by using the label information of all tasks, all of them’ generalization performance can be enhanced.
Secondly, since the STJML method has many parameters and conducting manual parameter sensitive tests to find appropriate values is a hectic and time-consuming process. Furthermore, the STJML method uses the original features to find a common feature space. Still, the original features itself are distorted, then the STJML method will not become a robust classifier. Therefore, in the future, we will use the particle swarm optimization [43] method to select the appropriate value of each parameter and the proper subset of excellent features across both domains. So, the STJML method for selecting parameters will be strengthened, and its performance will also improve due to the elimination of distorted features.
Lastly, nowadays, there is increasing interest in neural-network-based learning models [44] due to their outstanding performance; we will also extend the STJML method to deep learning framework. In our deep learning STJML method, we will extract deep features concerning our proposed method overall objective function.

Author Contributions

Conceptualization, R.K.S. and L.Y.; methodology, R.K.S. and L.Y.; software, R.K.S.; validation, R.K.S.; formal analysis, R.K.S. and L.Y.; investigation, R.K.S. and L.Y.; resources, R.K.S. and L.Y.; data curation, R.K.S.; writing—original draft preparation, R.K.S.; writing—review and editing, R.K.S. and L.Y.; visualization, R.K.S.; supervision, L.Y.; project administration, L.Y.; funding acquisition, L.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Ministry of Science and Technology, Taiwan, under Grant MOST107-2221-E-027-086-MY3.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, X.; Liu, F. Triplet loss guided adversarial domain adaptation for bearing fault diagnosis. Sensors 2020, 20, 320. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Chen, Z.; Li, X.; Zheng, H.; Gao, H.; Wang, H. Domain adaptation and adaptive information fusion for object detection on foggy days. Sensors 2018, 18, 3286. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Joshi, K.A.; Thakore, D.G. A survey on moving object detection and tracking in video surveillance system. Int. J. Soft Comput. Eng. 2012, 2, 44–48. [Google Scholar]
  4. Pan, S.J.; Yang, Q. A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 2009, 22, 1345–1359. [Google Scholar] [CrossRef]
  5. Pan, S.J.; Tsang, I.W.; Kwok, J.T.; Yang, Q. Domain adaptation via transfer component analysis. IEEE Trans. Neural Netw. 2011, 22, 199–210. [Google Scholar] [CrossRef] [Green Version]
  6. Long, M.; Wang, J.; Ding, G.; Sun, J.; Yu, P.S. Transfer feature learning with joint distribution adaptation. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2200–2207. [Google Scholar]
  7. Long, M.; Wang, J.; Ding, G.; Sun, J.; Yu, P.S. Transfer joint matching for unsupervised domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1410–1417. [Google Scholar]
  8. Zhang, J.; Li, W.; Ogunbona, P. Joint geometrical and statistical alignment for visual domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1859–1867. [Google Scholar]
  9. Xu, Y.; Pan, S.J.; Xiong, H.; Wu, Q.; Luo, R.; Min, H.; Song, H. A unified framework for metric transfer learning. IEEE Trans. Knowl. Data Eng. 2017, 29, 1158–1171. [Google Scholar] [CrossRef]
  10. Dai, W.; Yang, Q.; Xue, G.R.; Yu, Y. Boosting for transfer learning. In Proceedings of the 24th International Conference on Machine Learning, Corvallis, OR, USA, 20–24 June 2007; pp. 193–200. [Google Scholar]
  11. Pardoe, D.; Stone, P. Boosting for regression transfer. In Proceedings of the 27th International Conference on International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010; pp. 863–870. [Google Scholar]
  12. Wan, C.; Pan, R.; Li, J. Bi-weighting domain adaptation for cross-language text classification. In Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence, Barcelona, Spain, 16–22 July 2011. [Google Scholar]
  13. Wang, J.; Feng, W.; Chen, Y.; Yu, H.; Huang, M.; Yu, P.S. Visual domain adaptation with manifold embedded distribution alignment. In Proceedings of the 2018 ACM Multimedia Conference on Multimedia Conference, Seoul, Korea, 22–26 October 2018; ACM: New York, NY, USA, 2018; pp. 402–410. [Google Scholar]
  14. Long, M.; Wang, J.; Ding, G.; Pan, S.J.; Philip, S.Y. Adaptation regularization: A general framework for transfer learning. IEEE Trans. Knowl. Data Eng. 2014, 26, 1076–1089. [Google Scholar] [CrossRef]
  15. Fernando, B.; Habrard, A.; Sebban, M.; Tuytelaars, T. Unsupervised visual domain adaptation using subspace alignment. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2960–2967. [Google Scholar]
  16. Sanodiya, R.K.; Mathew, J.; Paul, B.; Jose, B.A. A kernelized unified framework for domain adaptation. IEEE Access 2019, 7, 181381–181395. [Google Scholar] [CrossRef]
  17. Pan, S.J.; Kwok, J.T.; Yang, Q. Transfer Learning via Dimensionality Reduction; AAAI: Chicago, IL, USA, 2008; Volume 8, pp. 677–682. [Google Scholar]
  18. Hofmann, T.; Schölkopf, B.; Smola, A.J. Kernel methods in machine learning. Ann. Stat. 2008, 1171–1220. [Google Scholar] [CrossRef] [Green Version]
  19. Shao, M.; Kit, D.; Fu, Y. Generalized transfer subspace learning through low-rank constraint. Int. J. Comput. Vis. 2014, 109, 74–93. [Google Scholar] [CrossRef]
  20. Kulis, B.; Saenko, K.; Darrell, T. What you saw is not what you get: Domain adaptation using asymmetric kernel transforms. In Proceedings of the IEEE CVPR 2011, Providence, RI, USA, 20–25 June 2011; pp. 1785–1792. [Google Scholar]
  21. Zhang, Y.; Yeung, D.Y. Transfer metric learning by learning task relationships. In Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 25–28 July 2010; pp. 1199–1208. [Google Scholar]
  22. Ding, Z.; Fu, Y. Robust transfer metric learning for image classification. IEEE Trans. Image Process. 2017, 26, 660–670. [Google Scholar] [CrossRef] [PubMed]
  23. Gretton, A.; Borgwardt, K.; Rasch, M.; Schölkopf, B.; Smola, A.J. A kernel method for the two-sample-problem. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2007; pp. 513–520. [Google Scholar]
  24. Sun, Q.; Chattopadhyay, R.; Panchanathan, S.; Ye, J. A two-stage weighting framework for multi-source domain adaptation. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2011; pp. 505–513. [Google Scholar]
  25. Zhou, D.; Bousquet, O.; Lal, T.N.; Weston, J.; Schölkopf, B. Learning with local and global consistency. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2004; pp. 321–328. [Google Scholar]
  26. Sanodiya, R.K.; Mathew, J.; Saha, S.; Thalakottur, M.D. A new transfer learning algorithm in semi-supervised setting. IEEE Access 2019, 7, 42956–42967. [Google Scholar] [CrossRef]
  27. Sim, T.; Baker, S.; Bsat, M. The CMU Pose Illumination and Expression Database of Human Faces. Available online: https://www.ri.cmu.edu/pub_files/pub2/sim_terence_2001_1/sim_terence_2001_1.pdf (accessed on 10 July 2020).
  28. Maaten, L.v.d.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  29. Gong, B.; Shi, Y.; Sha, F.; Grauman, K. Geodesic flow kernel for unsupervised domain adaptation. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2066–2073. [Google Scholar]
  30. Ghifary, M.; Balduzzi, D.; Kleijn, W.B.; Zhang, M. Scatter component analysis: A unified framework for domain adaptation and domain generalization. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1414–1430. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Sun, B.; Saenko, K. Subspace Distribution Alignment for Unsupervised Domain Adaptation. In Proceedings of the British Machine Vision Conference, Swansea, UK, 7–10 September 2015; Volume 4, pp. 24–31. [Google Scholar]
  32. Wang, H.; Wang, W.; Zhang, C.; Xu, F. Cross-domain metric learning based on information theory. In Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence, Quebec City, QC, Canada, 27–31 July 2014. [Google Scholar]
  33. Luo, L.; Chen, L.; Hu, S.; Lu, Y.; Wang, X. Discriminative and geometry aware unsupervised domain adaptation. arXiv 2017, arXiv:1712.10042. [Google Scholar] [CrossRef] [Green Version]
  34. Herath, S.; Harandi, M.; Porikli, F. Learning an invariant hilbert space for domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3845–3854. [Google Scholar]
  35. Wang, J.; Chen, Y.; Hao, S.; Feng, W.; Shen, Z. Balanced distribution adaptation for transfer learning. In Proceedings of the 2017 IEEE International Conference on Data Mining (ICDM), New Orleans, LA, USA, 18–21 November 2017; pp. 1129–1134. [Google Scholar]
  36. Li, S.; Song, S.; Huang, G.; Ding, Z.; Wu, C. Domain invariant and class discriminative feature learning for visual domain adaptation. IEEE Trans. Image Process. 2018, 27, 4260–4273. [Google Scholar] [CrossRef]
  37. Deng, W.Y.; Lendasse, A.; Ong, Y.S.; Tsang, I.W.H.; Chen, L.; Zheng, Q.H. Domain adaption via feature selection on explicit feature map. IEEE Trans. Neural Netw. Learn. Syst. 2018, 30, 1180–1190. [Google Scholar] [CrossRef]
  38. Liang, J.; He, R.; Sun, Z.; Tan, T. Aggregating randomized clustering-promoting invariant projections for domain adaptation. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 1027–1042. [Google Scholar] [CrossRef]
  39. Lu, H.; Shen, C.; Cao, Z.; Xiao, Y.; van den Hengel, A. An embarrassingly simple approach to visual domain adaptation. IEEE Trans. Image Process. 2018, 27, 3403–3417. [Google Scholar] [CrossRef]
  40. Zhang, J.; Li, W.; Ogunbona, P. Transfer learning for cross-dataset recognition: A survey. arXiv 2017, arXiv:1705.04396. [Google Scholar]
  41. Shu, L.; Latecki, L.J. Transductive domain adaptation with affinity learning. In Proceedings of the 24th ACM International on Conference on Information and Knowledge Management, Melbourne, Australia, 19–23 October 2015; pp. 1903–1906. [Google Scholar]
  42. Ruder, S. An overview of multi-task learning in deep neural networks. arXiv 2017, arXiv:1706.05098. [Google Scholar]
  43. Nguyen, B.H.; Xue, B.; Andreae, P. A particle swarm optimization based feature selection approach to transfer learning in classification. In Proceedings of the Genetic and Evolutionary Computation Conference, Kyoto, Japan, 15–19 July 2018; pp. 37–44. [Google Scholar]
  44. Liu, W.; Wang, Z.; Liu, X.; Zeng, N.; Liu, Y.; Alsaadi, F.E. A survey of deep neural network architectures and their applications. Neurocomputing 2017, 234, 11–26. [Google Scholar] [CrossRef]
Figure 1. Shows different possibilities th at can cause distribution differences in the source domain (training) or target domain (test) images: (a) Images captured from low-resolution camera such as Webcam, (b) Images captured from high-resolution camera such as Digital Single-lens Reflex(DSLR), (c) Images captured in day time, and (d) Images captured in night or raining time.
Figure 1. Shows different possibilities th at can cause distribution differences in the source domain (training) or target domain (test) images: (a) Images captured from low-resolution camera such as Webcam, (b) Images captured from high-resolution camera such as Digital Single-lens Reflex(DSLR), (c) Images captured in day time, and (d) Images captured in night or raining time.
Sensors 20 04367 g001
Figure 2. (a) Initial representation of source domain data with two classes (i.e., red plus (+) symbol and blue minus (−) symbol) in 2-d space, where the symbols with circles are outlier data samples. (b) Initial representation of target domain data with two classes (i.e., yellow plus (+) symbol and green minus (−) symbol) in 2-d space. (c) Data representation after application of only feature learning. (d) Data representation after application of jointly feature learning and instance re-weighting.
Figure 2. (a) Initial representation of source domain data with two classes (i.e., red plus (+) symbol and blue minus (−) symbol) in 2-d space, where the symbols with circles are outlier data samples. (b) Initial representation of target domain data with two classes (i.e., yellow plus (+) symbol and green minus (−) symbol) in 2-d space. (c) Data representation after application of only feature learning. (d) Data representation after application of jointly feature learning and instance re-weighting.
Sensors 20 04367 g002
Figure 3. The t-SNE view of initial feature spaces on the tasks of Office + Caltech data set.
Figure 3. The t-SNE view of initial feature spaces on the tasks of Office + Caltech data set.
Sensors 20 04367 g003
Figure 4. The t-SNE view of learned features from the proposed method Subspace based Transfer Joint Matching with Laplacian Regularization (STJML) after considering all the terms. (a) A D (SURF features) with the predicted class labels for the target domain. (b) A D (SURF features) with the given class labels for both domains. (c) A W (VGG-FC6 features) with the predicted class labels for the target domain. (d) A W (VGG-FC6 features) with the given class labels for both domains.
Figure 4. The t-SNE view of learned features from the proposed method Subspace based Transfer Joint Matching with Laplacian Regularization (STJML) after considering all the terms. (a) A D (SURF features) with the predicted class labels for the target domain. (b) A D (SURF features) with the given class labels for both domains. (c) A W (VGG-FC6 features) with the predicted class labels for the target domain. (d) A W (VGG-FC6 features) with the given class labels for both domains.
Sensors 20 04367 g004
Figure 5. The t-SNE view of learned features from the proposed method after omitting the subspace of both domains. (a) A W (VGG-FC6 features) with the predicted class labels for the target domain. (b) A W (VGG-FC6 features) with the given class labels for both domains.
Figure 5. The t-SNE view of learned features from the proposed method after omitting the subspace of both domains. (a) A W (VGG-FC6 features) with the predicted class labels for the target domain. (b) A W (VGG-FC6 features) with the given class labels for both domains.
Sensors 20 04367 g005
Figure 6. The t-SNE view of learned features from the proposed method after omitting the instance-re-weighting term. (a) A D (SURF features) with the predicted class labels for the target domain. (b) A D (SURF features) with the given class labels for both domains.
Figure 6. The t-SNE view of learned features from the proposed method after omitting the instance-re-weighting term. (a) A D (SURF features) with the predicted class labels for the target domain. (b) A D (SURF features) with the given class labels for both domains.
Sensors 20 04367 g006
Figure 7. The t-SNE view of learned features from the proposed method after omitting the marginal distribution term. (a) A W (VGG-FC6 features) with the predicted class labels for the target domain. (b) A W (VGG-FC6 features) with the given class labels for both domains.
Figure 7. The t-SNE view of learned features from the proposed method after omitting the marginal distribution term. (a) A W (VGG-FC6 features) with the predicted class labels for the target domain. (b) A W (VGG-FC6 features) with the given class labels for both domains.
Sensors 20 04367 g007
Figure 8. The t-SNE view of learned features from the proposed method after omitting the conditional distribution term. (a) A W (VGG-FC6 features) with the predicted class labels for the target domain. (b) A W (VGG-FC6 features) with the given class labels for both domains.
Figure 8. The t-SNE view of learned features from the proposed method after omitting the conditional distribution term. (a) A W (VGG-FC6 features) with the predicted class labels for the target domain. (b) A W (VGG-FC6 features) with the given class labels for both domains.
Sensors 20 04367 g008
Figure 9. The t-SNE view of learned features from the proposed method after omitting the Laplacian distribution term. (a) A W (VGG-FC6 features) with the predicted class labels for the target domain. (b) A W (VGG-FC6 features) with the given class labels for both domain.
Figure 9. The t-SNE view of learned features from the proposed method after omitting the Laplacian distribution term. (a) A W (VGG-FC6 features) with the predicted class labels for the target domain. (b) A W (VGG-FC6 features) with the given class labels for both domain.
Sensors 20 04367 g009
Figure 10. Performance of the proposed STJML method, varying the values of the parameters ‘k’ and ‘ n n ’ on all tasks in both the datasets. (a) k-Pose, Illumination, and Expression (PIE) face dataset. (b) k-Office + Caltech10 with VGG-FC6 features dataset. (c) n n -PIE face dataset. (d) n n -Office + Caltech10 with VGG-FC6 features dataset.
Figure 10. Performance of the proposed STJML method, varying the values of the parameters ‘k’ and ‘ n n ’ on all tasks in both the datasets. (a) k-Pose, Illumination, and Expression (PIE) face dataset. (b) k-Office + Caltech10 with VGG-FC6 features dataset. (c) n n -PIE face dataset. (d) n n -Office + Caltech10 with VGG-FC6 features dataset.
Sensors 20 04367 g010
Figure 11. Performance of the proposed STJML method, varying the values of the parameter δ on all tasks in both the datasets. (a) δ -PIE face dataset. (b) δ -Office + Caltech10 with VGG-FC6 features dataset.
Figure 11. Performance of the proposed STJML method, varying the values of the parameter δ on all tasks in both the datasets. (a) δ -PIE face dataset. (b) δ -Office + Caltech10 with VGG-FC6 features dataset.
Sensors 20 04367 g011
Figure 12. Performance of the proposed STJML method, varying the values of the parameters λ and η on all tasks in both the datasets. (a) λ -PIE face dataset. (b) λ -Office + Caltech10 with VGG-FC6 features dataset. (c) η -PIE face dataset. (d) η -Office + Caltech10 with VGG-FC6 features dataset.
Figure 12. Performance of the proposed STJML method, varying the values of the parameters λ and η on all tasks in both the datasets. (a) λ -PIE face dataset. (b) λ -Office + Caltech10 with VGG-FC6 features dataset. (c) η -PIE face dataset. (d) η -Office + Caltech10 with VGG-FC6 features dataset.
Sensors 20 04367 g012
Figure 13. Performance of the proposed STJML method, varying the values of the parameter σ on all tasks in both the datasets. (a) σ -PIE face dataset. (b) σ -Office + Caltech10 with VGG-FC6 features dataset.
Figure 13. Performance of the proposed STJML method, varying the values of the parameter σ on all tasks in both the datasets. (a) σ -PIE face dataset. (b) σ -Office + Caltech10 with VGG-FC6 features dataset.
Sensors 20 04367 g013
Table 1. Accuracy (%) on the PIE face recognition dataset.
Table 1. Accuracy (%) on the PIE face recognition dataset.
TasksPrimitive
Algorithms
Transfer Learning Algorithms
NNPCATCA
[5]
GFK
[29]
JDA
[6]
CDML
[32]
TJM
[7]
TDA-AL
[41]
CDDA
[33]
BDA
[35]
GA-DA
[33]
DGA-DA
[33]
JGSA
[8]
RTML
[22]
EMFS
[37]
DICD
[36]
LDADA
[39]
KUFDA
[16]
DICE
[38]
STJML
Proposed
PIE 1
5→7
26.0924.8040.7626.1558.8153.2229.5235.9760.2223.9857.4065.3268.0760.1261.873.034.567.6784.182.32
PIE 2
5→9
26.5925.1841.7927.2754.2353.1233.7632.9758.7024.0060.5462.8167.5255.2158.872.044.970.3477.973.22
PIE 3
5→27
30.6729.2659.6331.1584.5080.1259.2035.2483.4848.9384.0583.5482.8785.1986.892.261.586.0695.994.65
PIE 4
5→29
16.6716.3029.3517.5949.7548.2326.9628.4354.1724.0052.2156.0746.5052.9852.666.935.449.0266.562.99
PIE 5
7→5
24.4924.2241.8125.2457.6252.3939.4038.9062.3349.0057.8963.6925.2158.1359.269.931.472.6281.281.27
PIE 6
7→9
46.6345.5351.4747.3762.9354.2337.7449.3964.6424.0061.5861.2754.7763.9264.565.934.974.3474.082.1
PIE 7
7→27
54.0753.3564.7354.2575.8268.3649.8053.2679.9048.9782.3482.3758.9676.1677.985.353.587.8688.691.91
PIE 8
7→29
26.5325.4333.7027.0839.8937.3417.0936.9544.0024.0041.4246.6335.4140.3844.348.726.461.7068.869.54
PIE 9
9→5
21.3720.9534.6921.8250.9643.5437.3934.0358.4649.0054.1456.7222.8153.1253.869.438.273.9178.877.16
PIE 10
9→7
41.0140.4547.7043.1657.9554.8735.2949.5459.7323.9560.7761.2644.1958.6759.865.430.572.5676.780.9
PIE 11
9→27
46.5346.1456.2346.4168.4562.7644.0348.9977.2048.9777.2377.8356.8669.8170.683.460.686.9685.290.68
PIE 12
9→29
26.2325.3133.1526.7839.9538.2117.0339.3447.2424.0043.5044.2441.3642.1341.961.440.769.8570.871.01
PIE 13
27→5
32.9531.9655.6434.2480.5875.1259.5142.2083.1049.0079.8381.8472.1481.1282.793.161.390.0093.395.55
PIE 14
27→7
62.6860.9667.8362.9282.6380.5360.5863.9082.2623.9684.7185.2788.278.9285.690.156.788.4095.0093.01
PIE 15
27→9
73.2272.1875.8673.3587.2583.7264.8861.6486.6424.0089.1790.9586.0989.5188.289.067.884.6292.390.37
PIE 16
27→29
37.1935.1140.2637.3854.6652.7825.0646.3258.3324.0053.6253.8074.3256.2657.275.650.475.2481.182.65
PIE 17
29→5
18.4918.8526.9820.3546.4627.3432.8632.9248.0249.0052.7357.4417.5229.1149.462.931.354.0573.863.17
PIE 18
29→7
24.1923.3929.9024.6242.0530.8222.8937.2645.6123.8947.6453.8441.0633.2845.157.024.167.4671.275.5
PIE 19
29→9
28.3127.2129.9028.4953.3136.3422.2436.6452.0224.0051.6655.2749.2039.8555.965.935.470.7774.176.83
PIE 20
29→27
31.2430.3433.6431.3357.0140.6130.7238.9655.9948.9458.8261.8234.7547.1359.674.848.276.7881.882.9
Average34.7633.8544.7535.3560.2453.6937.2942.1463.1033.9862.5665.0953.3958.8062.873.143.474.4280.580.88
Table 2. Accuracy (%) on Office + Caltech dataset with VGG-FC6 features.
Table 2. Accuracy (%) on Office + Caltech dataset with VGG-FC6 features.
TasksPrimitive
Algorithms
Transfer Learning Algorithms
NNPCASVMTCA
[5]
GFK
[29]
JDA
[6]
SA
[15]
TJM
[7]
CORAL
[31]
CDDA
[33]
ILS
[34]
BDA
[35]
JGSA
[8]
KUFDA
[16]
DICE
[38]
STJML
Proposed
A→C70.176.4974.280.1477.7382.0177.182.4579.082.178.980.2381.1285.1283.684.14
A→D52.359.8751.765.6059.2370.0664.972.6167.168.272.564.9768.7878.3466.078.34
A→W69.969.1563.176.9473.8983.7276.082.7174.878.182.476.6178.3080.1676.691.11
C→A81.986.4386.786.6386.0188.1083.985.8089.486.587.686.0186.2289.8389.590.51
C→D55.661.1461.569.4262.4272.6166.275.7967.666.173.066.8877.0780.1369.987.89
C→W65.974.2374.874.9174.9180.6776.077.9677.677.184.475.9376.6187.8379.889.15
D→A57.067.4358.775.1568.5877.1369.080.7975.682.679.274.3286.9585.2183.286.84
D→C48.058.5055.569.1859.5770.5262.374.4464.776.166.569.7278.0980.3878.782.63
D→W86.795.5991.896.6195.9397.6290.596.9494.693.794.297.6397.6298.8795.898.3
W→A62.475.1569.880.2779.0184.276.682.2581.286.585.980.7990.8191.5688.890.51
W→C57.569.0164.775.2470.1674.7970.778.4575.280.177.076.2276.6684.1282.083.17
W→D83.994.9089.493.6394.9096.8190.494.9092.692.887.492.3692.9910088.1100
Average65.9373.970.1578.6475.1981.5275.382.0978.280.880.778.4782.6086.8381.888.59
Table 3. Accuracy (%) on Office + Caltech dataset with SURF features.
Table 3. Accuracy (%) on Office + Caltech dataset with SURF features.
TasksPrimitive
Algorithm
Transfer Learning Algorithms
1NNPCASVMGFK
[29]
TCA
[5]
JDA
[6]
CORAL
[31]
TJM
[7]
SCA
[30]
JGSA
[8]
ARTL
[14]
ILS
[34]
RTML
[22]
DICD
[36]
LDADA
[39]
DICE
[38]
STJML
Proposed
C→A23.739.553.146.045.643.152.146.845.651.544.148.549.347.354.850.249.69
C→W25.834.641.737.039.339.346.439.040.045.431.541.444.746.460.248.146.1
C→D25.544.647.840.845.949.045.944.647.145.939.545.947.649.741.551.050.32
A→C26.039.041.740.742.040.945.139.539.741.536.140.043.742.438.442.741.85
A→W29.835.931.937.040.038.044.442.034.945.833.639.044.345.149.352.253.22
A→D25.533.844.640.135.742.039.545.239.547.136.940.143.938.939.149.749.04
W→C19.928.228.824.831.533.033.730.231.133.229.731.234.833.631.737.833.21
W→A23.029.127.627.630.529.836.030.030.039.938.337.635.334.135.137.543.01
W→D59.289.278.385.491.192.486.689.287.390.587.986.091.089.874.687.393.9
D→C26.329.726.429.333.031.233.831.430.729.930.534.634.634.629.933.733.3
D→A28.533.226.228.732.833.437.732.831.638.034.941.233.334.540.641.139.35
D→W63.486.152.580.387.589.284.785.484.491.988.585.889.091.274.784.193.22
Average31.443.641.143.146.246.848.846.345.250.044.347.649.349.047.551.352.18
Table 4. Running time complexity of proposed method (STJML) and some other methods.
Table 4. Running time complexity of proposed method (STJML) and some other methods.
MethodsRunning Time (s)MethodsRunning Time (s)
1-NN2.22CORAL47.05
TCA3.42JGSA482.665
GFK19.64TJM22.15
JDA21.96Proposed Method72.74

Share and Cite

MDPI and ACS Style

Sanodiya, R.K.; Yao, L. A Subspace Based Transfer Joint Matching with Laplacian Regularization for Visual Domain Adaptation. Sensors 2020, 20, 4367. https://doi.org/10.3390/s20164367

AMA Style

Sanodiya RK, Yao L. A Subspace Based Transfer Joint Matching with Laplacian Regularization for Visual Domain Adaptation. Sensors. 2020; 20(16):4367. https://doi.org/10.3390/s20164367

Chicago/Turabian Style

Sanodiya, Rakesh Kumar, and Leehter Yao. 2020. "A Subspace Based Transfer Joint Matching with Laplacian Regularization for Visual Domain Adaptation" Sensors 20, no. 16: 4367. https://doi.org/10.3390/s20164367

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop