Next Article in Journal
Analysis and Performance Evaluation of Transfer Learning Algorithms for 6G Wireless Networks
Previous Article in Journal
Bionic Design of a Novel Portable Hand-Elbow Coordinate Exoskeleton for Activities of Daily Living
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Pseudo Labels for Unsupervised Domain Adaptation: A Review

School of Information Science and Technology, North China University of Technology, Beijing 100144, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(15), 3325; https://doi.org/10.3390/electronics12153325
Submission received: 10 July 2023 / Revised: 27 July 2023 / Accepted: 27 July 2023 / Published: 3 August 2023
(This article belongs to the Topic Computer Vision and Image Processing)

Abstract

:
Conventional machine learning relies on two presumptions: (1) the training and testing datasets follow the same independent distribution, and (2) an adequate quantity of samples is essential for achieving optimal model performance during training. Nevertheless, meeting these two assumptions can be challenging in real-world scenarios. Domain adaptation (DA) is a subfield of transfer learning that focuses on reducing the distribution difference between the source domain ( D s ) and target domain ( D t ) and subsequently applying the knowledge gained from the D s task to the D t task. The majority of current DA methods aim to achieve domain invariance by aligning the marginal probability distributions of the D s . and D t . Recent studies have pointed out that aligning marginal probability distributions alone is not sufficient and that alignment of conditional probability distributions is equally important for knowledge migration. Nonetheless, unsupervised DA presents a more significant difficulty in aligning the conditional probability distributions because of the unavailability of labels for the D t . In response to this issue, there have been several proposed methods by researchers, including pseudo-labeling, which offer novel solutions to tackle the problem. In this paper, we systematically analyze various pseudo-labeling algorithms and their applications in unsupervised DA. First , we summarize the pseudo-label generation methods based on the single and multiple classifiers and actions taken to deal with the problem of imbalanced samples. Second, we investigate the application of pseudo-labeling in category feature alignment and improving feature discrimination. Finally, we point out the challenges and trends of pseudo-labeling algorithms. As far as we know, this article is the initial review of pseudo-labeling techniques for unsupervised DA.

1. Introduction

Deep learning has achieved remarkable success in diverse fields, including object detection, speech recognition, health care, and computer vision in bygone years. Its effectiveness is heavily dependent on a substantial quantity of training data, yet collecting vast labeled data is challenging, costly, and time intensive. Meanwhile, the model’s performance can be compromised when dealing with new domains owing to domain shifts. Hence, it is a significant and arduous task to maximize the utilization of existing labeled data to boost the model’s generalization capability and compensate for the sample scarcity.
To address the above issues, the research field of domain adaptation (DA) was established. DA endeavors to migrate the knowledge acquired from labeled data in the source domain ( D s ) to the target domain ( D t ), with the purpose of enhancing the model’s performance in the D t [1,2,3,4,5]. Tan et al. [1] provided a definition of deep transfer learning and divided it into four groups: adversarial, instance-based, network, and mapping. Mei et al. [2] present a comprehensive survey of deep DA methods and categorize them according to their loss functions. Wilson et al. [3] discuss various DA methods in single-source DA. Kouw et al. [4] divide DA into three categories in terms of how classifiers learn from the D s and generalize to the D t : methods based on a single observation (sample-based), methods based on the set of observations representation (feature-based) methods, and parameter estimator-based (inference-based) methods. Fan et al. [5] classified DA into different types based on the label sets in the D t and D t , including open-set, close-set, partial, generalized, and zero-shot DA.
The target of DA is to boost a model that can generalize well to the D t that are related but not identical to the D s by leveraging the knowledge learned from the D s [6]. The main challenge in DA is addressing the distribution shift stemming from the disparities in feature distributions between the D s and D t . DA can be classified as supervised DA, semi-supervised DA, and unsupervised DA, derived from whether the D t is labeled or not [2], among which unsupervised DA is the most challenging and is currently receiving a lot of attention from researchers. According to the theoretical analysis of DA by Ben-david et al. [7], the generalization error on the D t is defined by (1) the empirical error of the D s classifier, (2) the empirical error between the D s and D t , and (3) the ideal joint error. At first, researchers mainly focused on reducing the empirical error between the D s and D t , assuming that the ideal joint error is small. From a probability statistics perspective, this involves aligning the marginal probability distribution between the D s and D t . Such methods include maximum mean difference (MMD) [8], deep adversarial neural networks (DANN) [9], and adversarial discriminative domain adaptation (ADDA) [10]. As the research progresses, more and more researchers have started to focus on the effect of the ideal joint error. Zhang et al. [11] argue that merely aligning the edge distribution between two domains is not enough, and alignment without considering the class-conditional distribution will lead to an increase in the ideal joint error, constraining the upper bound of the model’s theoretical error. The majority of existing DA methods only consider aligning the global features of the D s and D t but fail to ensure the alignment of class-specific features, which may constrain the model’s performance on particular tasks. Worse still, inter-domain category-level alignment often requires labels of both domains to achieve, which is difficult for unsupervised DA. Recently, motivated by the pseudo-labeling techniques in semi-supervised learning, an increasing number of researchers have made significant improvements in model performance by assigning pseudo-labels to D t to aid in achieving inter-domain category-level alignment. Yet, almost no one has undertaken a systematic organization and analysis of these works, which is the motivation for our review.
We investigated the papers in some top sessions from 2017 to 2022 and counted and analyzed the number of DA papers on pseudo-labeling methods, as shown in Figure 1. Overall, the number of DA-related papers grows year by year, and it is expected that more and more research on DA will be conducted in the coming years. We use darker colors to represent papers on DA that utilize pseudo-labeling methods, and it is evident that the quantity of these papers has been growing steadily over the years. It is worth noting that its share in the overall DA papers is also increasing year by year; in particular, nearly 50% of the DA papers in the ICCV and CVPR conferences in 2022 used pseudo-labeling methods, which indicates the successful application of pseudo-labels in DA.
The primary emphasis of this review is to analyze the methods for generating pseudo-labels and their applications in unsupervised DA. First, the generation methods are categorized as either single-classifier-based or multi-classifier-based, and the measures for dealing with sample imbalance are explored. Second, the paper examines the applications of pseudo-labels in unsupervised DA, which are divided into two parts: using pseudo-labels for category feature alignment and for enhancing the feature discrimination of classifiers.
The structural framework of this paper is shown in Figure 2.
As far as we know, there are no existing review papers on DA methods based on pseudo-labeling. Specifically, this paper’s primary contributions are:
  • We review in detail the background knowledge related to DA and pseudo-labeling methods and sort out the connections and differences between them.
  • We have organized and analyzed the paper in detail in terms of both the pseudo-labeling generation method and the application of pseudo-labeling in unsupervised DA. To the best of our knowledge, it is the first attempt to summarize pseudo labels used in the community of domain adaptation.
  • We conducted a comprehensive review of various pseudo-labeling methods within each category through experimental evaluations. This analysis enables readers to grasp the nuances of each technique and make informed decisions.
  • We point out possible challenges and future directions for pseudo-labeling methods in DA applications.

2. Background

2.1. Unsupervised Domain Adaptation

In this section, we give a formal definition of unsupervised DA. We denote the D s input data and labels as x s = x i i n s and y s = y i i n s , and the D t input data as x t = x i i n t , where n s and n t represent the number of samples in the D s and D t , respectively, so that the sample spaces of the D s and D t can be denoted as D s = x i s , y i s i n s   and D t = x i t i n t , respectively. The feature space and label space of both the D s and D t are assumed to be identical, X s = X t ,     Y s = Y t , while the joint probability distributions are different, P s x , y P t x , y . The objective of unsupervised DA is to learn a mapping function using the aforementioned data, f : x t y t , to make predictions on the labels in y t Y t for the D t . Interventionary studies involving animals or humans, and other studies that require ethical approval, must list the authority that provided approval and the corresponding ethical approval code.

2.2. Pseudo-Labeling

Lee et al. [12] first proposed the method of pseudo-labeling, which uses the token with the highest prediction probability as a pseudo-label, y ^ = argmax x f θ x , for unlabeled data, and then assigns a weight, w , to the unlabeled data and slowly increases it during the training process to perform the training. In contrast to the consistent regularization approach, the pseudo-labeling approach does not rely on region-specific data enhancement and is easier to implement [13]. We categorized the current pseudo-labeling methods into two types: divergence-based methods and self-training methods. Divergence-based methods utilize multiple networks to perform a task and leverage the divergence of different networks, f θ 1 · and f θ 2 · , to boost the quality of pseudo-labels, thereby improving the overall model’s performance [14]. Self-training methods, on the other hand, use the model’s own confident predictions to predict the pseudo-label for D t unlabeled data, thereby augmenting the training data [13].

3. Pseudo-Labeling Generation Methods

Pseudo-labeling generation methods are methods dedicated to improving the accuracy of model-generated D t pseudo-labels and to further facilitating domain alignment. We classify the pseudo-labeling generation methods into three categories: single-classifier-based generation methods, multi-classifier-based generation methods, and category-balancing methods for difficult samples. Single-classifier-based generation methods refer to obtaining pseudo-labels of the D t by one classifier and completing the DA task by self-training. The multi-classifier-based generation method refers to obtaining more accurate pseudo-labels by the difference of more than two classifiers and then completing the DA task by self-training. The category-balancing method for difficult samples refers to further considering the pseudo-labels category-balancing problem based on the quality of generated pseudo-labels, as a way to obtain higher-quality pseudo-labels.

3.1. Single-Classifier-Based Generation Method

The basic assumption of the single-classifier-based generation approach is that the model’s own highly confident predictions are correct [15]. The single-classifier-based approach generates pseudo-labels by using the model’s own confident predictions for unlabeled data. In semi-supervised classification tasks, it can predict unlabeled data by using a limited quantity of available labeled data, filtering according to some criterion, and finally training the model together with true labels and pseudo-labels [12,16]. In contrast, in unsupervised DA problems, using a model trained from labeled D s data and pseudo-labeling the unlabeled D t data can be of great help in promoting domain alignment, especially category-level alignment.
In Wang et al. [17], a structured prediction-based selective pseudo-labeling approach was proposed. This method utilizes the structural information of the D t data through clustering and labels the D t samples collectively based on the clusters they belong to. The distance from the D t sample to the cluster center is used as the criterion for pseudo-labeling, with samples closer to the center being more prone to be chosen for pseudo-labeling and for participating in the next round of iterative training. Deng et al. [18] employed a teacher–student-model structure with pseudo-labeling provided by the teacher model. The discriminative learning and category-level alignment goals are achieved through discriminative clustering loss and clustering-based alignment loss.
To address the issue of sparse pseudo-labels generated by single-classifier models, Shin et al. [19] proposed a two-phase pseudo-label densification framework that uses a bootstrapping mechanism in the self-training loss function to boost the model’s generalization ability. Wang et al. [20] introduced a binary soft-constrained information entropy to improve the credibility of the mined class prototypes and class anchors, particularly for samples at the decision boundary. This method increased the accuracy of the model in estimating pseudo-labels for the D t . Zhang et al. [21] proposed AuxSelf-Training for the auxiliary model from the perspective of training samples, in which the sample selection is founded on reducing the proportion of D s data and increasing the proportion of the D t data proportion to construct the intermediate domain and gradually overcome the distance bias across the domain.

3.2. Multi-Classifier-Based Generation Methods

Multi-classifier-based generation methods are extensively used in semi-supervised learning, including the classical approaches proposed by Qin et al. [22] and Zhou et al. [23]. Unlike the single-classifier-based approach, the multi-classifier-based approach usually trains two or three different networks and uses the divergence between different networks to allocate high-quality pseudo-labels to unlabeled samples. These pseudo-labeled samples are then used in training together with the labeled samples, leading to effective DA results.
Inspired by [22,23], Saito et al. [24] proposed asymmetric tri-training for unsupervised domain adaptation (ATDA) for unsupervised DA, as shown in Figure 3.
The approach involves utilizing a shared feature extractor and three classifiers. To enable and to classify from different perspectives, ATDA adds as a regularization term to the cross-entropy loss, as shown in the following Equation (1).
E θ F , θ F 1 , θ F 2 = 1 n i = 1 n L y F 1 F x i , y i + L y F 2 F x i , y i + λ W 1 T W 2
where L y is the cross-entropy loss, and θ F 1 and θ F 2 are hyper-parameters. ATDA screens out the D t samples with the consistent sum output of F 1 and F 2 confidence greater than a certain value to be pseudo-labeled for training. Inspired by ATDA to allocate pseudo-labels to unlabeled D t samples and Mixup [25], Li et al. [26] put forward a three-branch CNN model based on an electrocardiogram (ECG), which demonstrated superior performance on the task of classifying heartbeats in the presence of domain shift. Similar to ATDA, Venkat et al. [27] proposed a multi-source DA method that employs pseudo-labels generated by multiple classifiers ground on the consistency of their predictions. This approach achieved promising results in their experiments.
To further mitigate the undesirable consequences of incorrect pseudo-labeling on training, Zheng et al. [28] used uncertainty to mitigate the undesirable consequences of incorrect pseudo-labeling on DA. Unlike the fixed threshold used by Saito et al. [24] and Zou et al. [29], Zheng et al. [28] used a dynamic thresholding approach, as shown in Figure 4.
The method models the uncertainty of pseudo-labeling by accessing networks with different depths of primary and secondary classifiers to obtain different perspectives and prediction variance. The prediction variance and the classification loss on pseudo-labeling are defined as follows in Equations (2) and (3).
D k l = Ε F x t j | θ t l o g F x t j | θ t F a u x x t j | θ t
L c e = E [ p ^ t j l o g F ( x t j | θ t ) ]
where F is the main classifier and F a u x is the secondary classifier. The modified pseudo-label loss function is expressed by Equation (4).
L r e c t = E e x p D k l L c e + D k l
When the prediction results of the main classifier and the subclassifier are very different, the value of D k l will be larger, indicating that the pseudo-labels may be inaccurate. Du et al. [30] enhanced the performance of the dual-classifier adversarial training network construction put forward by Saito et al. [31] by introducing additional losses. Specifically, they added a self-supervised loss on the D t and a gradient difference loss on both domains on top of the classification loss on the D s . The self-supervised loss on the D t improved the discriminability of the D t distribution, which is beneficial for subsequent category-level alignment. In terms of pseudo-label generation, Du et al. [30] used the softmax outputs of two classifiers to weight the samples to obtain the k class prime c k and finally pseudo-labeled the D t data with the nearest prime strategy. Li et al. [32] put forward a method for obtaining accurate pseudo-labels for D t data using the prediction consistency of multiple classifiers. The method explicitly adapts the multi-order classifier from the D s to the D t , ensuring that the pseudo-labels are both accurate and diverse. Ge et al. [33] proposed a simultaneous training symmetric network to achieve mutual supervision under collaborative training, thus avoiding the formation of overfitting to the network’s own output error, which leads to the amplification of pseudo-labeling noise. Qin et al. [22] added a pseudo-labels training set to the training model of MCD [31] to participate in the training, enhancing the efficacy of this network.

3.3. Category-Balancing Methods for Difficult Samples

In the DA problem, the complexity of various DA tasks varies because of domain differences. Similarly, within the same DA task, the alignment difficulty may vary across different classes. Samples in the D s that the classifier finds difficult to label can be referred to as difficult samples. For difficult samples, several of its categories may mostly be misclassified into other classes or sieved out because of low confidence in the classifier output, leading to severe category-balance bias in the selected samples and further negatively impacting DA. Methods for alleviating the class-balance problem for difficult samples are relatively novel and have achieved promising results.
Chen et al. [34] used the easy-to-hard transfer strategy (ETHS) to select reliable pseudo-label samples and then used adaptive prototype alignment (APA) to achieve cross-domain category alignment, as shown in Figure 5.
The network has the same construction as DANN [9] and consists of a feature extractor G, a label predictor F, and a domain discriminator D. In EHTS, by first computing the mean value of the features of each class in the D s as the class prototype c k S = 1 N s k x i s , y i s D s k G x i s , in which c k S represents the class prototype of the k class of the D s , and then using the cosine similarity to estimate the distance between the D t and each class of the D s and finally setting a threshold value, the samples in the D t that have phase degrees exceeding the threshold value are assigned pseudo-labels.
In the APA phase, the distance between each class in the D s   and D t is defined as d c k S , c k T = c k S c k T 2 , and cross-domain class-level alignment is achieved by minimizing the APA loss. The APA loss and the total loss are defined as Equations (5) and (6), respectively:
L a p a θ g = k = 1 C d c k I S , c k I T
min θ g , θ f   max θ d i = 1 n s L c F G x i S ; θ g ; θ f , y i S + λ L d θ g , θ d + γ L a p a θ g  
where L c is the standard cross-entropy loss, and λ and γ are the weights controlling the interaction between source categorization loss, domain confusion loss L d , and APA loss. Zou et al. [29] mainly address the DA problem in semantic segmentation. To alleviate the issue of imbalanced classes caused by fixed thresholds, the paper sets a threshold K c for each class and gradually performs DA through self-step learning [35]. Zhang et al. [11] delved into the negative impact of inter-class imbalance of D t samples on DA and proposed adaptive prediction calibration (APC) to mitigate the problem of hard classes by boosting hard classes, keeping common classes, and eliminating easy classes and introduced TE and SE (temporal fusion and self-fusion, respectively) to improve the reliability of prediction, as shown in Figure 6.
Recently, Liu et al. [36] suggested utilizing cyclic self-training as a replacement for standard self-training to tackle the issue of distribution bias in DA, as shown in Figure 7.
The network structure is the same as MCD [31] and contains a feature extractor and two classifiers. The difference is that the training alternates between two steps, the inner loop and the outer loop. In the inner loop, the D t pseudo-labels are used to train the target classifier; in the outer loop, the shared representation is updated to boost the capability of the target classifier on the D s . To address the issue of noise amplification caused by high pseudo-label confidence, this study introduces an uncertainty metric derived from the information-theoretic Tsallis entropy. This metric can automatically minimize the pseudo-label uncertainty without requiring any manual adjustment or setting of the confidence threshold.

4. Application of Pseudo-Labeling in Domain Adaptation

Unlike pseudo-labeling generation methods, the application of pseudo-labeling in DA refers to the application of pseudo-labeling in traditional DA methods (e.g., adversarial-based, difference-based, and reconstruction-based methods, etc.). We classify them into two major categories: the application of pseudo-labeling in improving classifier discrimination and the application of pseudo-labeling in category feature alignment. The first category refers to methods that obtain classifiers with high generalization ability through supervised learning in the D s and weakly supervised learning in the D t that is labeled with high-quality pseudo-labels. The second category refers to methods that use pseudo-labeling to facilitate category feature alignment in the D s and D t .

4.1. Application of Pseudo-Labeling in Improving Classifier Discrimination

Zhao et al. [37] integrated a DANN with a teacher–student network model [38] to learn feature representations with target differentiation using a consistency-forcing approach. It used prediction averaging and label sharpening to generate pseudo-labels for unlabeled D t and introduced interpolation consistency into the unsupervised DA task to enhance the clarity of the decision boundaries. Zhang et al. [39] divided a CNN feature extractor into several blocks, each block being a set of CNN layers, as shown in Figure 8.
The figure illustrates that each block for feature extraction includes a series of CNN layers. The domain classifier comprises several FC layers that serve to differentiate the domain to which each sample belongs. As the samples are propagated forward from the lower to the higher layers, the learned feature distribution changes smoothly from domain-relevant information to domain-independent information. Notably, the authors proposed an extension of the CAN method called incremental CAN (iCAN), i.e., incorporating the idea of self-training by leveraging the image classifier and the domain classifier from the previous training period, and iteratively choose a set of D t samples with pseudo-labels. A dynamic thresholding method is employed to add these samples to the training set, achieving better results. The dynamic-thresholding-related settings are as follows in Equations (7)–(9).
T C = 1 1 + e ρ A
A = 1 N s i = 1 N s I y i s , a r g m a x c   p c x i s
I a , b = 1 ,   i f   a = b 0 ,   o t h e r w i s e
where T C denotes the threshold value, in which ρ was set to a fixed value of 3, p c x i t denotes the possibility of the i th sample, and x i t pertains to class c. Xie et al. [40] put forward a method to evaluate the contribution of edge distributions (global) and conditional distributions (local) to the target task in the DA problem. Specifically, they learned the semantic representation of unlabeled D t samples by aligning the labeled D s examples and pseudo-labeled D t examples. To alleviate the adverse effect of incorrect pseudo-labels, instead of aligning these newly acquired primes directly in each iteration, MSTN aligns exponentially moving average primes. Wang et al. [41] suggested a method called confidence-aware pseudo-labeling selection (CAPLS), which employs an iterative learning approach to gradually achieve domain alignment. Based on MMD, Kang et al. [42] introduced a difference measure called contrast domain difference (CDD) to explicitly model intra-class domain differences and inter-class domain differences. Recently, Chen et al. [43] further improved migration performance by using higher-order statistics for domain matching and using pseudo-labeled samples from the D t to learn domain-invariant representations. Dong et al. [44] designed a confidence-anchor-induced pseudo-labels generator to mine the confidence pseudo-labels of the D t by building confidence anchor groups and capturing consistent cross-domains by class-relationship-aware consistency loss inter-class relationships. Li et al. [45] defined an attention-aware transmission distance to measure domain differences using predictive feedback from an iterative learning classifier.
Yang et al. [46] proposed a bidirectional generation cross-domain generation framework by adding MMD loss and consistency loss to the loss function and pseudo-labeling the D t data using the D s classifier obtained from pretraining to implement a bidirectional cross-domain generation method. Hu et al. [47] preserved the category structure of the D t by a duplex discriminator that also included classification tasks while aligning the overall features of the domain, as shown in Figure 9.
It consists of four parts: encoder E , generator D , duplex discriminators D s and D t , and classifier C . The role of E is to compress the image pixel-level features into z . Under the D s and D t constraint, the domain alignment is achieved by transforming the two-domain sample styles, while the D s and D t simultaneously classify the real image to preserve the class information of z . The D t pseudo-labels missing during training are provided by Russo et al. [48], using GANS to introduce a symmetric mapping between the two domains and adding a class-consistency loss to enhance the structural stability and image quality of the reconstructed samples.

4.2. Application of Pseudo-Labeling in Category Feature Alignment

In recent years, the generative adversarial network (GAN) proposed by Goodfellow et al. [49] has been extensively employed in unsupervised learning. The GAN network primarily includes a generator and a discriminator. The generator in the GAN network generates synthetic samples using random noise, while the discriminator is responsible for distinguishing between real and synthetic samples. The GAN is trained by the strategy of maximum–minimum alternating optimization, and the ability of the discriminator to discriminate the authenticity is used as the “yardstick”, thus generating samples that bear closer resemblance to the true samples. The GAN has been described in detail in [49,50,51], and interested researchers can refer to the above literature. The true and fake samples in the GAN can correspond to the D s and D t in DA, respectively, and the generator corresponds to the feature extractor, while the discriminator is an implicit alignment scale. Due to the clear logic of GAN and its natural structural adaptation to the DA task, it has become a popular method in DA for learning transferable features that are domain invariant between the D s and D t .
The domain adversarial neural network (DANN) [9] is comprised of a feature extractor, a classifier, and a domain discriminator. It maximizes the domain confusion loss by using a gradient reversal layer (GRL) while minimizing the label prediction loss on the D s data to achieve feature alignment between the D s and D t . Unlike DANN, adversarial discriminative domain adaptation (ADDA) [10] uses separate feature extractors for each domain to capture more domain-specific information, aligns the D t features toward the D s through a pretraining, fine-tuning training model, and finally tests the D t samples using the D t exclusive feature extractor and the D s classifier. The above two simple and effective adversarial DA methods have become the basic architectures of many current DA methods [52,53,54]. Nonetheless, these two techniques only take into account the alignment of the marginal distribution between the D s and D t and do not consider the alignment of the conditional distribution (class-level alignment), so even if domain confusion is achieved, the classifier may perform poorly on the target task. As an analogy, by adversarial training, even with perfectly aligned marginal distributions, the feature space can still blend the characteristics of apples in the D s with those of oranges in the D t [55].
To alleviate the above problems, more and more adversarial-based DA methods have started to consider category-level alignment, where the combination of adversarial training and pseudo-labeling methods is notable. Based on DANN, Zhang et al. [52] introduced center loss to achieve conditional distribution alignment. They proposed a method to deal with unlabeled samples in the D t . They used the predictions of the D s classifier to allocate pseudo-labels to each sample and defined the loss function as shown in Equation (10).
min   θ E L c t = x i Φ X t E x i c y ^ i 2 2
where y ^ i is the label of x i predicted by the classifier, and c y ^ i denotes the center of the i class. To alleviate the negative impact of incorrect pseudo-labeling during training, Zhang et al. [52] filtered a subset of D t samples for training by means of a card-fixed threshold, and the filtering function is as follows in Equation (11).
Φ X t = { x i | x i X t   and   m a x p x i T }
where p x i is a K-dimensional vector, dimension i corresponds to the predicted probability of class i , m a x ( p x i ) is the possibility that sample x i pertains to the predicted class, and T is a fixed threshold. In [56], the selection of training samples is implicitly guided by pseudo-labels from the perspective of class-conditional domain alignment, focusing on the problem of intra-domain class imbalance and inter-domain class distribution shift.
Yu et al. [57] proposed transfer learning with a dynamic adversarial adaptation network (DAAN), which consists of three main components: labeled classifier, global domain discriminator, and local subdomain discriminator. The overall loss function is as follows in Equation (12).
L θ f , θ y , θ d , θ d c | c = 1 C = L y λ 1 ω L g + ω L l
where λ is a constant value, while ω is a dynamic factor measuring the importance of L g and L l . L y , L g , and L l denote classification loss, global loss, and local loss, respectively, where the pseudo-label of the D t is also used in the calculation of L l . Wang et al. [58] proposed an entropy-based adaptive reweighting adversarial DA method from the perspective of the conditional distribution, as shown in Figure 10.
To promote positive migration and curb negative migration, the method uses an entropy criterion to reveal the degree of sample transferability, which is then reweighted and fed back into the discriminative network to force the underlying distribution closer. In this paper, the loss function of domain adversarial training incorporates a conditional entropy term, and the weights assigned to different samples in the adversarial training are determined by the following Equations (13) and (14).
L a d v θ f , θ d = 1 n s + n t x i D s D t 1 + H p L d G d f x i , d i
where       H p = 1 C c = 1 C p c l o g p c
In addition to this, the authors use triplet loss to facilitate category-level alignment. Samples are randomly selected on the basis of the sampling approach. The pseudo-labels for the D t are obtained by maximizing the posterior probability of the D s cross-entropy, which is gradually optimized as the model is trained. In addition, based on the intuitive consideration that images with high prediction scores are more likely to be correctly classified, only D t samples with prediction scores above a certain threshold, T, are chosen for training in this paper, and the threshold is set as a constant in this paper. To avoid mislabeled target instances from propagating errors to the next iteration to disrupt the subspace learning process, Tanwani et al. [55] employ a network trained on D s data in the initial stages of training to predict pseudo-labels for unlabeled D t and, then, retain only the most confident pseudo-labels for each category, resulting in a balanced mini-batch consisting of equal numbers of D s and D t data for replacement sampling during training.
The construction of the graph convolutional adversarial network (GCAN) is proposed in Xinhong et al. [59]. The GCAN approach includes three alignment mechanisms: structure-aware, class-mass, and domain alignment. In class-mass alignment, class-mass alignment loss is computed using pseudo-labeled D t features and labeled D s features to ensure that samples belonging to the same class from different domains are embedded closely. In order to develop the module for aligning the class centers of mass, the method uses a target classifier to assign pseudo-tags and obtains pseudo-tagged D t . Both labeled and pseudo-labeled samples are utilized to calculate the center of mass for each class. The DART (domain-against-residual transfer) network proposed by Fang et al. [60] comprises a deep feature extractor, a deep label classifier, and a domain classifier in its architecture. The entropy minimization method is used in computing the D t label prediction loss by setting its loss function to Equation (15):
L H = 1 N t i = 1 N t j = 1 c p ( y i t = j | x i t ) l o g p ( y i t = j | x i t )
where c represents the total number of classes, and p ( y i t = j | x i t ) can be obtained by p ( y i t | x i t ) = G t G f x i t . Through minimizing the entropy penalty, the target classifier G t will self-adjust to expand the likelihood difference between predictions and predict more indicative labels accordingly. The alignment of the conditional distributions between the D s and D t is accomplished in Cicek et al. [61] by incorporating an extra joint predictor. This predictor learns the distributions on the domain and class labels. The encoder is trained to deceive this predictor in the same class of samples for each domain. In [53], the confusion matrix is computed using a domain discriminator based on DANN as a way to correct the noise in the pseudo-labels. In [6], discriminator D discriminates the domain distribution along with the class distribution. Given that there exist some transferable regions between the D s and D t images, we propose an attention module embedded in the GAN. In this way, we can remove as much background information as possible and further minimize the domain shift between the D s and D t . The corresponding experimental results can support our conclusion. To fully utilize the label information in the D t , we present a straightforward yet effective approach to pseudo-label the unlabeled D t samples. This idea can enhance the performance of classifier C while mitigating negative migration. Gu et al. [62] introduced an adversarial DA approach based on a spherical feature space and employed a Gaussian mixture model in the spherical space to obtain more robust pseudo-labels.
One of the most popular approaches in deep DA is to minimize the distributional discrepancy of domain features to achieve domain alignment, which employs deep neural networks to extract informative feature representations for the D s and D t samples. Among them, two broad categories of domain distribution disparity metrics are commonly used: explicit and implicit. The explicit metrics are generally MMD distance, form-center distance, class-prototype distance, etc. In addition, implicit metrics are adversarial-based methods, popular learning, optimal transmission methods, etc. Fortunately, the pseudo-labeling approach can still be applied in a flexible manner in the aforementioned methods and lead to improved performance.
To achieve alignment of the conditional distributions of the D s and D t , Long et al. [63] modified the MMD to estimate the distance between the class-conditional distributions Q s ( x s | y s = c ) and Q t ( x t | y t = c ) . The inter-class MMD distance is defined as follows in Equation (16).
1 n s c x i D s c A T x i 1 n t c x j D t c A T x j 2
where the norm represents the L 2 norm, which is defined as the square root of the sum of the squared elements of a vector, A is the orthogonal transformation matrix, D s c is the set of samples of class c in the D s , and n s c = D s c . The same is true for the D t . Since the D t c has no label, the authors incorporated the prediction of the D t classifier directly as its pseudo-label in the computation.
Chadha et al. [54] enhanced the performance of ADDA by referring to the framework of semi-supervised GAN and exploiting the MMD loss. To fully leverage the discriminative information present in the distribution of labels, Luo et al. [64] put forward a method in which the features from the D s and D t are mapped into a regenerated Hilbert kernel space, and the conditional distribution of the domains is represented by the conditional covariance operator in the kernel space. Then, the conditional kernel Bures (CKB) metric put forward in the paper is estimated and optimized based on the variance feedback. For each class, the semantic difference of that class between two domains is modeled using a multivariate Gaussian distribution that utilizes the inter-domain feature mean difference and the intra-class feature covariance on the D t , and then the D s features are augmented by randomly sampling semantic enhancement directions from the constructed distribution [65]. As a result of the absence of labels for the data in the D t , its pseudo-label is defined as y t j = a r g m a x c P t j c , where P t j c is the softmax output of the D t sample x t j . Tanwisuth et al. [66] provides a framework for extracting class prototypes and aligning D t features with them. Liang et al. [67], based on a nearest form-centered classifier, project the form centers of the D s and D t features into an invariant subspace, where the pseudo-labels are computed using the feature transformation matrix and the maximum likelihood estimate of the D s expectation. Zhao et al. [68] defined a symmetric mirror loss based on Kullback–Leibler scatter to enhance the degree of domain alignment and followed an unsupervised discriminative clustering approach [69] to introduce auxiliary distributions as soft pseudo-labels. Li et al. [70] introduced a two-layer optimization strategy using pseudo-labels generated by the optimal classifier. With the purpose of boosting the accuracy of the pseudo-labels, Liang et al. [71] reduced the classifier bias by introducing auxiliary classifiers only for the D t and incorporated the maximum prediction probability as a weight to the standard cross-entropy loss in Equations (17) and (18):
y ^ i = a r g m a x k   p i , k ,   i = 1 , 2 , , N t
L p l o u r s = λ N t u i = 1 N t u p i , y ^ i l o g p i , y ^ i
Sharma et al. [72] added a link to feature processing based on Gani et al. [9], as shown in Figure 11.
As in [9], the network includes an encoder, a classifier, and a discriminator. The encoder G is shared between the D s and D t and used to reduce the dimensionality of the data. After dimensionality reduction of the D s data by encoder G, the classifier C is used to generate a softmax predictive distribution of the categories, and then, supervised training is performed on the D s labeled data using standard cross-entropy loss. To achieve global feature alignment between the two domains, the authors train the domain discriminator D using L D to classify the D s and D t features and train G using L a d v to generate features for the confusion discriminator. In this way, domain-invariant features are extracted through a min-max training process between L D and L a d v .
The classification loss and the adversarial loss are defined as follows in Equations (19)–(21).
L s u p = E x , y ~ D s l o g C G x y
L a d v = E x ~ D t l o g D G x
L D = E x ~ D s l o g D G x E x ~ D t l o g 1 D G x
Notably, The authors adopt the K-nearest neighbor (KNN) method to allocate pseudo-labels to the D t samples ground on their similarity to nearby labeled D s samples. To avoid some D t samples being assigned incorrect pseudo-labels because some D t samples may not have corresponding true D s samples, the authors use a class-balanced small-batch sampling method to alleviate this problem. Finally, the correlation matrix is constructed based on similarity, and multiple-sample contrast loss is used to achieve class-level alignment of the D s and D t features.
Xu et al. [73] proposed a weighted optimal transfer strategy that uses spatial prototype information and intra-domain structure to reduce the negative transfer from samples near the decision boundary in the D t . Luo et al. [74] proposed a Riemannian manifold embedding and alignment framework that projects D s and D t features into manifold space and uses a manifold metric to measure domain differences while taking into account both category-level alignment and global alignment.
In addition to the aforementioned mainstream methods, there are some DA methods using pseudo-labeling that have also achieved better results.
Hou et al. [75] proposed a source-free domain image translation (SFIT) method in which the model is split into two branches, one branch inputting D t images and one branch using cycle-GAN as a generator to generate D s -style images guided by the D s and D t models.
The reconstruction-based approach data reconstruction refers to the addition of a data reconstruction task, typically using an autoencoder or generative adversarial network to ensure feature invariance during migration. Zhu et al. [76] proposed a new low-dimensional visual attribute (LDVA) coding method based on an autoencoder that can train end-to-end models for tasks such as DA, few-sample learning, and zero-sample learning.
In a data-enhancement-based approach, adversarial domain adaptation with domain mixup (DM-ADA), Xu et al. [77] proposes a mixup alignment. The features are gradually aligned by constructing some synthetic data, which serves as a bridge between the D s and D t . To realize domain alignment at the category level and ensure that features pertaining to the same category in both domains are mapped closely to the same latent space, the authors introduce classification loss to ensure category consistency between the decoded image and the input and mitigate the detrimental effects of mislabeling by filtering out samples with classification confidence below a certain threshold. Zhong et al. [78] introduced a general approach named E-MixNet, which improves the model performance by applying an enhanced mixup technique on labeled D s samples and pseudo-labeled D t samples to curb the combinatorial risk in the target risk.
In a heterogeneous-based approach, Paolo et al. [79] propose a novel heterogeneous-distributed unsupervised DA method that focuses on the challenging setting of positive-unlabeled (PU) learning, where only positive and unlabeled examples are available. The method aims to enhance predictive models for a target domain by leveraging knowledge from a related source domain, even when the two domains are described with different feature spaces. The proposed method not only handles heterogeneous feature spaces but also efficiently distributes the workload to manage large volumes of data.
Existing unsupervised DA methods for time-series data have mainly centered on aligning the marginal distribution between the source and target domains. However, they tend to overlook the conditional distribution discrepancy, which can lead to misclassification in the target domain. He et al. [80] propose a novel method called ARADA-TK (attentive recurrent adversarial domain adaptation with top-k time-series pseudo-labeling) for unsupervised domain adaptation in time-series data. It focuses on learning domain-invariant representations by capturing temporal dependencies and reducing conditional distribution discrepancies.

5. Experience Evaluation

Given that image classification is a crucial task in various computer vision applications, the majority of the aforementioned algorithms were initially developed to address this issue. Therefore, in this section, we compare the current leading pseudo-labeling methods in unsupervised DA on the Office-31 classification dataset, showing how much benefit this method can bring to image classification.
The Office-31 dataset is widely used as a benchmark in visual DA, and it consists of 4652 images belonging to 31 object categories commonly found in office environments, such as laptops, filing cabinets, keyboards, etc. [81]. These images were primarily sourced from three different domains: Amazon (product images from online e-commerce websites), webcam (low-resolution images captured by webcams), and DSLR (high-resolution images captured by digital SLR cameras). There are 2817 images in the Amazon dataset, with an average of 90 images and one image background per category, 795 images in the webcam dataset, where the images show obvious noise, color, and white balance artifacts, and 498 images in the DSLR dataset. There are five objects in each category, and each object is pictured, on average, three times from different viewpoints.
Images are collected from online retailers (e.g., Amazon) and webcams (e.g., webcam) under various office-related categories. For the DSLR domain, images are taken using a high-quality DSLR camera. The collected images are resized and standardized to a fixed size, such as 224 × 224 pixels, to ensure uniformity and facilitate data processing. Then, they are divided into three domains: Amazon (A), webcam (W), and DSLR (D). Each image is associated with a domain label to indicate its source. To perform domain adaptation experiments, the dataset is split into training and test sets. The training set contains images from the source domain (e.g., Amazon), along with their corresponding labels. The test sets are composed of images from the target domains (webcam and DSLR) without any labeled data.
Due to the variations in the parameters, experimental protocols, and tuning strategies used in different studies apart from pseudo-labeling, it is challenging to conduct a direct and fair comparison of all the methods. Therefore, we present comparison between the proposed method using pseudo-labels and an unsupervised DA method using only deep networks. In particular, we modularize these methods according to Section 3 and Section 4 from the pseudo-label generation method and the application of pseudo-label adaptation in unsupervised domains to reflect the effectiveness of different modules. It is important to mention that all approaches are uniformly set to unsupervised DA of isomorphic closed sets with Resnet-50 as the framework, and we report the highest performance reported in the respective papers. The tables present the accuracy results of the unsupervised DA model (JAN) using only deep networks as a baseline, as shown in Table 1 and Table 2.

6. Challenges and Future Directions

Although pseudo-labeling methods have a large number of applications in deep DA with good results, there are still some problems, and we present them and indicate potential areas for future research.
(1)
There is a lack of a common, universal indicator to evaluate the quality of pseudo-labels.
In DA with pseudo-labeling methods, the quality of pseudo-labels can directly affect the effectiveness of DA. Through literature research, we found that only a few researchers have analyzed the quality of pseudo-labels: Zhang et al. [11] portrayed the variation of pseudo-label accuracy with training cycles in the ablation experiment section, while Liu et al. [36] plotted ROC curves and used the AUC metric to quantify pseudo-label effectiveness. We believe that a generalized metric for evaluating the quality of pseudo-labels would be useful for the development of DA methods that use pseudo-labels.
(2)
Cross-domain issues affect the quality of pseudo-labels.
Unlike semi-supervised learning where the training and test sets obey the same distribution, DA faces the challenge of cross-domain distribution shift. Liu et al. [36] experimentally confirmed that in the cross-domain case (where the pseudo labels of the D t are generated by the D s model), the quality of the obtained pseudo-labels is lower than when the D s and D t obey the same distribution. Moreover, the difficulty (inter-domain distance) of different migration tasks is different, and finding a pseudo-labeling method that is applicable to different difficulty DA problems is a worthwhile research direction.
(3)
The dataset is more homogeneous, while the real scenario is more complex.
At present, the public datasets used for DA are generally Digits, Office-31, VisDA-2017, etc. It is fair to compare with public datasets for theoretical studies, which is beneficial to the theoretical development in this direction. However, it is possible that the proposed method performs better only on the mentioned datasets. Further expansion of more complex public datasets in the future will facilitate the application of DA methods in real scenarios.
(4)
Research has mainly focused on classification problems, and there is a lack of research on other DA problems.
The current application of the pseudo-labeling method for DA mainly solves the classification problem, and its application in semantic segmentation DA, weakly supervised DA, and domain generalization can be further tried in the future.

7. Conclusions

Deep DA is a research area with important real-world applications. The successful application of pseudo-labeling methods in deep DA has further contributed to its rapid development. This review classifies DA methods using pseudo-labeling into self-training-based methods, divergence-based methods, adversarial-based methods, difference-based methods, and other methods. Finally, we discuss the challenges faced by pseudo-labeling in DA applications and some directions that deserve further research in the future.

Author Contributions

Conceptualization, Y.L. and L.G.; methodology, Y.L.; software, Y.G.; validation, Y.L. and L.G.; formal analysis, Y.L.; investigation, Y.G.; resources, Y.L.; data curation, L.G.; writing—original draft preparation, Y.L.; writing—review and editing, L.G.; visualization, Y.G.; supervision, Y.G.; project administration, Y.L.; funding acquisition, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by [the National Natural Science Foundation of China] grant number [62071006].

Data Availability Statement

Data available in a publicly accessible repository that does not issue DOIs. Publicly available datasets were analyzed in this study. This data can be found here: [https://paperswithcode.com/dataset/office-31].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tan, C.; Sun, F.; Kong, T.; Zhang, W.; Yang, C.; Liu, C. A Survey on Deep Transfer Learning. arXiv 2018. [Google Scholar] [CrossRef]
  2. Mei, W.; Deng, W. Deep Visual Domain Adaptation: A Survey. Neurocomputing 2018, 312, 135–153. [Google Scholar] [CrossRef] [Green Version]
  3. Wilson, G.; Cook, D.J. A Survey of Unsupervised Deep Domain Adaptation. arXiv 2020. [Google Scholar] [CrossRef] [PubMed]
  4. Kouw, W.M.; Loog, M. A Review of Domain Adaptation without Target Labels. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 766–785. [Google Scholar] [CrossRef] [Green Version]
  5. Fan, M.; Cai, Z.; Zhang, T.; Wang, B. A survey of deep domain adaptation based on label set classification. Multimedia Tools Appl. 2022, 81, 39545–39576. [Google Scholar] [CrossRef]
  6. Chen, W.; Hu, H. Generative attention adversarial classification network for unsupervised domain adaptation. Pattern Recognit. 2020, 107, 107440. [Google Scholar] [CrossRef]
  7. Ben-David, S.; Blitzer, J.; Crammer, K.; Kulesza, A.; Pereira, F.C.; Vaughan, J.W. A theory of learning from different domains. Mach. Learn. 2010, 79, 151–175. [Google Scholar] [CrossRef] [Green Version]
  8. Gretton, A.; Borgwardt, K.; Rasch, M.; Schölkopf, B.; Smola, A.J. A Kernel Two-Sample Test. J. Mach. Learn. Res. 2012, 13, 723–773. [Google Scholar]
  9. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M.; Lempitsky, V. Domain-Adversarial Training of Neural Networks. J. Mach. Learn. Res. 2015, 17, 2096-2030. [Google Scholar]
  10. Tzeng, E.; Hoffman, J.; Saenko, K.; Darrell, T. Adversarial Discriminative Domain Adaptation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2962–2971. [Google Scholar]
  11. Zhang, Y.; Jing, C.; Lin, H.; Chen, C.; Huang, Y.; Ding, X.; Zou, Y. Hard Class Rectification for Domain Adaptation. Knowl. Based Syst. 2020, 222, 107011. [Google Scholar] [CrossRef]
  12. Lee, D.-H. Pseudo-Label: The Simple and Efficient Semi-Supervised Learning Method for Deep Neural Networks. In Proceedings of the ICML 2013 Workshop: Challenges in Representation Learning (WREPL), Atlanta, GA, USA, 16–21 June 2013. [Google Scholar]
  13. Kong, L.; Hu, B.; Liu, X.; Lu, J.; You, J.; Liu, X. Constraining pseudo-label in self-training unsupervised domain adaptation with energy-based model. Int. J. Intell. Syst. 2022, 37, 8092–8112. [Google Scholar] [CrossRef]
  14. Zhou, Z.-H.; Li, M. Semi-supervised learning by disagreement. Knowl. Inf. Syst. 2010, 24, 415–439. [Google Scholar] [CrossRef]
  15. Zhu, X. Semi-Supervised Learning Literature Survey; Comput Sci, University of Wisconsin-Madison: Madison, WI, USA, 2008; p. 2. [Google Scholar]
  16. Grandvalet, Y.; Bengio, Y. Semi-supervised Learning by Entropy Minimization. In Proceedings of the Conférence Francophone sur L’apprentissage Automatique, Montpellier, LIF, France, 16–19 June 2004. [Google Scholar]
  17. Wang, Q.; Breckon, T. Unsupervised Domain Adaptation via Structured Prediction Based Selective Pseudo-Labeling. Proc. AAAI Conf. Artif. Intell. 2020, 34, 6243–6250. [Google Scholar] [CrossRef]
  18. Deng, Z.; Luo, Y.; Zhu, J. Cluster Alignment with a Teacher for Unsupervised Domain Adaptation. arXiv 2019. [Google Scholar] [CrossRef]
  19. Shin, I.; Woo, S.; Pan, F.; Kweon, I. Two-Phase Pseudo Label Densification for Self-Training Based Domain Adaptation; Springer International Publishing: Cham, Switzerland, 2020. [Google Scholar]
  20. Wang, F.; Han, Z.; Yin, Y. Source Free Robust Domain Adaptation Based on Pseudo Label Uncertainty Estimation. J. Softw. 2022, 33, 1183–1199. [Google Scholar] [CrossRef]
  21. Zhang, Y.; Deng, B.; Jia, K.; Zhang, L. Gradual Domain Adaptation via Self-Training of Auxiliary Models. arXiv 2021. [Google Scholar] [CrossRef]
  22. Qin, C.; Wang, L.; Zhang, Y.; Fu, Y. Generatively Inferential Co-Training for Unsupervised Domain Adaptation. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019; pp. 1055–1064. [Google Scholar] [CrossRef]
  23. Zhou, Z.-H.; Li, M. Tri-training: Exploiting unlabeled data using three classifiers. IEEE Trans. Knowl. Data Eng. 2005, 17, 1529–1541. [Google Scholar] [CrossRef] [Green Version]
  24. Saito, K.; Ushiku, Y.; Harada, T. Asymmetric Tri-training for Unsupervised Domain Adaptation. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 11–15 August 2017. [Google Scholar]
  25. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond Empirical Risk Minimization. arXiv 2018. [Google Scholar] [CrossRef]
  26. Li, J.; Wang, G.; Chen, M.; Ding, Z.; Yang, H. Mixup Asymmetric Tri-Training for Heartbeat Classification under Domain Shift. IEEE Signal Process. Lett. 2021, 28, 718–722. [Google Scholar] [CrossRef]
  27. Venkat, N.; Kundu, J.; Singh, D.K.; Revanur, A.; VenkateshBabu, R. Your Classifier can Secretly Suffice Multi-Source Domain Adaptation. arXiv 2021. [Google Scholar] [CrossRef]
  28. Zheng, Z.; Yang, Y. Rectifying Pseudo Label Learning via Uncertainty Estimation for Domain Adaptive Semantic Segmentation. Int. J. Comput. Vis. 2020, 129, 1106–1120. [Google Scholar] [CrossRef]
  29. Zou, Y.; Yu, Z.; Kumar BV, K.V.; Wang, J. Domain Adaptation for Semantic Segmentation via Class-Balanced Self-Training. arXiv 2018. [Google Scholar] [CrossRef]
  30. Du, Z.; Li, J.; Su, H.; Zhu, L.; Lu, K. Cross-Domain Gradient Discrepancy Minimization for Unsupervised Domain Adaptation. arXiv 2021. [Google Scholar] [CrossRef]
  31. Saito, K.; Watanabe, K.; Ushiku, Y.; Harada, T. Maximum Classifier Discrepancy for Unsupervised Domain Adaptation. arXiv 2018. [Google Scholar] [CrossRef]
  32. Li, S.; Zhang, J.; Ma, W.; Liu, C.H.; Li, W. Dynamic Domain Adaptation for Efficient Inference. arXiv 2021. [Google Scholar] [CrossRef]
  33. Ge, Y.; Chen, D.; Li, H. Mutual Mean-Teaching: Pseudo Label Refinery for Unsupervised Domain Adaptation on Person Re-identification. arXiv 2020. [Google Scholar] [CrossRef]
  34. Chen, C.; Xie, W.; Xu, T.; Huang, W.; Rong, Y.; Ding, X.; Huang, Y.; Huang, J. Progressive Feature Alignment for Unsupervised Domain Adaptation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 627–636. [Google Scholar]
  35. Kumar, M.P.; Packer, B.; Koller, D. Self-Paced Learning for Latent Variable Models. NIPS 2010, 1, 1189–1197. [Google Scholar]
  36. Liu, H.; Wang, J.; Long, M. Cycle Self-Training for Domain Adaptation. arXiv 2021. [Google Scholar] [CrossRef]
  37. Zhao, X.; Wang, S. (11 2019). Adversarial Learning and Interpolation Consistency for Unsupervised Domain Adaptation. IEEE Access 2019, 7, 170448–170456. [Google Scholar] [CrossRef]
  38. Hinton, G.E.; Vinyals, O.; Dean, J. Distilling the Knowledge in a Neural Network. arXiv 2015. [Google Scholar] [CrossRef]
  39. Zhang, W.; Ouyang, W.; Li, W.; Xu, D. Collaborative and Adversarial Network for Unsupervised Domain Adaptation. In Proceedings of the2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3801–3809. [Google Scholar]
  40. Xie, S.; Zheng, Z.; Chen, L.; Chen, C. Learning Semantic Representations for Unsupervised Domain Adaptation. In Proceedings of the International Conference on Machine Learning, Stockholm Sweden, 10–15 July 2018. [Google Scholar]
  41. Wang, Q.; Bu, P.; Breckon, T. Unifying Unsupervised Domain Adaptation and Zero-Shot Visual Recognition. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019. [Google Scholar] [CrossRef] [Green Version]
  42. Kang, G.; Jiang, L.; Yang, Y.; Hauptmann, A. Contrastive Adaptation Network for Unsupervised Domain Adaptation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4888–4897. [Google Scholar]
  43. Chen, C.; Fu, Z.; Chen, Z.; Jin, S.; Cheng, Z.; Jin, X.; Hua, X. HoMM: Higher-order Moment Matching for Unsupervised Domain Adaptation. arXiv 2019. [Google Scholar] [CrossRef]
  44. Dong, J.; Fang, Z.; Liu, A.; Sun, G.; Liu, T. Confident Anchor-Induced Multi-Source Free Domain Adaptation. In Proceedings of the Neural Information Processing Systems, Virtual, 6–12 December 2021. [Google Scholar]
  45. Li, M.; Zhai, Y.; Luo, Y.; Ge, P.; Ren, C. Enhanced Transport Distance for Unsupervised Domain Adaptation. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 13933–13941. [Google Scholar]
  46. Yang, G.; Xia, H.; Ding, M.; Ding, Z. Bi-Directional Generation for Unsupervised Domain Adaptation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020. [Google Scholar]
  47. Hu, L.; Kan, M.; Shan, S.; Chen, X. Duplex Generative Adversarial Network for Unsupervised Domain Adaptation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1498–1507. [Google Scholar]
  48. Russo, P.; Carlucci, F.M.; Tommasi, T.; Caputo, B. From Source to Target and Back: Symmetric Bi-Directional Adaptive GAN. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8099–8108. [Google Scholar]
  49. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.C.; Bengio, Y. Generative Adversarial Networks. arXiv 2014. [Google Scholar] [CrossRef]
  50. Gui, J.; Sun, Z.; Wen, Y.; Tao, D.; Ye, J. A Review on Generative Adversarial Networks: Algorithms, Theory, and Applications. IEEE Trans. Knowl. Data Eng. 2020, 35, 3313–3332. [Google Scholar] [CrossRef]
  51. Radford, A.; Metz, L.; Chintala, S. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. arXiv 2015. [Google Scholar] [CrossRef]
  52. Zhang, Y.; Zhang, Y.; Wang, Y.; Tian, Q. Domain-Invariant Adversarial Learning for Unsupervised Domain Adaption. arXiv 2018. [Google Scholar] [CrossRef]
  53. Chen, M.; Zhao, S.; Liu, H.; Cai, D. Adversarial-Learned Loss for Domain Adaptation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020. [Google Scholar]
  54. Chadha, A.; Andreopoulos, Y. Improved Techniques for Adversarial Discriminative Domain Adaptation. IEEE Trans. Image Process. 2019, 29, 2622–2637. [Google Scholar] [CrossRef] [Green Version]
  55. Tanwani, A.K. DIRL: Domain-Invariant Representation Learning for Sim-to-Real Transfer. arXiv 2021. [Google Scholar] [CrossRef]
  56. Jiang, X.; Lao, Q.; Matwin, S.; Havaei, M. Implicit Class-Conditioned Domain Alignment for Unsupervised Domain Adaptation. arXiv 2020. [Google Scholar] [CrossRef]
  57. Yu, C.; Wang, J.; Chen, Y.; Huang, M. Transfer Learning with Dynamic Adversarial Adaptation Network. arXiv 2019. [Google Scholar] [CrossRef]
  58. Wang, S.; Zhang, L. Self-adaptive Re-weighted Adversarial Domain Adaptation. arXiv 2020. [Google Scholar] [CrossRef]
  59. Xinhong, M.; Zhang, T.; Xu, C. GCAN: Graph Convolutional Adversarial Network for Unsupervised Domain Adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8258–8268. [Google Scholar] [CrossRef]
  60. Fang, X.; Bai, H.; Guo, Z.; Shen, B.; Hoi, S.; Xu, Z. DART: Domain-Adversarial Residual-Transfer networks for unsupervised cross-domain image classification. Neural Netw. 2020, 127, 182–192. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  61. Cicek, S.; Soatto, S. Unsupervised Domain Adaptation via Regularized Conditional Alignment. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Long Beach, CA, USA, 15–20 June 2019; pp. 1416–1425. [Google Scholar]
  62. Gu, X.; Sun, J.; Xu, Z. Spherical Space Domain Adaptation with Robust Pseudo-Label Loss. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 9098–9107. [Google Scholar]
  63. Long, M.; Wang, J.; Ding, G.; Sun, J.; Yu, P.S. Transfer Feature Learning with Joint Distribution Adaptation. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2200–2207. [Google Scholar]
  64. Luo, Y.; Ren, C. Conditional Bures Metric for Domain Adaptation. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 13984–13993. [Google Scholar]
  65. Li, S.; Xie, M.; Gong, K.; Liu, C.H.; Wang, Y.; Li, W. Transferable Semantic Augmentation for Domain Adaptation. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 11511–11520. [Google Scholar]
  66. Tanwisuth, K.; Fan, X.; Zheng, H.; Zhang, S.; Zhang, H.; Chen, B.; Zhou, M. A Prototype-Oriented Framework for Unsupervised Domain Adaptation. Adv. Neural Inf. Process. Syst. 2021, 34, 17194–17208. [Google Scholar]
  67. Liang, J.; He, R.; Sun, Z.; Tan, T. Distant Supervised Centroid Shift: A Simple and Efficient Approach to Visual Domain Adaptation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 2970–2979. [Google Scholar]
  68. Zhao, Y.; Wang, M.; Cai, L. Reducing the Covariate Shift by Mirror Samples in Cross Domain Alignment. In Proceedings of the 35th Conference on Neural Information Processing Systems, Online, 6–14 December 2021. [Google Scholar]
  69. Jabi, M.; Pedersoli, M.; Mitiche, A.; Ayed, I.B. Deep Clustering: On the Link Between Discriminative Models and K-Means. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 43, 1887–1896. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  70. Li, M.; Jiang, K.; Zhang, X. Implicit Task-Driven Probability Discrepancy Measure for Unsupervised Domain Adaptation. Adv. Neural Inf. Process. Syst. 2021, 34, 25824–25838. [Google Scholar]
  71. Liang, J.; Hu, D.; Feng, J. Domain Adaptation with Auxiliary Target Domain-Oriented Classifier. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 16627–16637. [Google Scholar]
  72. Sharma, A.; Kalluri, T.; Chandraker, M. Instance Level Affinity-Based Transfer for Unsupervised Domain Adaptation. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 5357–5367. [Google Scholar]
  73. Xu, R.; Liu, P.; Wang, L.; Chen, C.; Wang, J.; He, K.; Zhang, X.; Ren, S.; Long, M.; Cao, Z.; et al. Reliable Weighted Optimal Transport for Unsupervised Domain Adaptation. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 4393–4402. [Google Scholar]
  74. Luo, Y.; Ren, C.; Ge, P.; Huang, K.; Yu, Y. Unsupervised Domain Adaptation via Discriminative Manifold Embedding and Alignment. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020. [Google Scholar]
  75. Hou, Y.; Zheng, L. Visualizing Adapted Knowledge in Domain Transfer. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 13819–13828. [Google Scholar]
  76. Zhu, P.; Wang, H.; Saligrama, V. Learning Classifiers for Target Domain with Limited or No Labels. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019. [Google Scholar]
  77. Xu, M.; Zhang, J.; Ni, B.; Li, T.; Wang, C.; Tian, Q.; Zhang, W. Adversarial Domain Adaptation with Domain Mixup. In Proceedings of the AAAI Conference on Artificial Intelligence, Long Beach, CA, USA, 10–15 June 2019. [Google Scholar]
  78. Zhong, L.; Fang, Z.; Liu, F.; Lu, J.; Yuan, B.; Zhang, G. How does the Combined Risk Affect the Performance of Unsupervised Domain Adaptation Approaches? arXiv 2020. [Google Scholar] [CrossRef]
  79. Mignone, P.; Pio, G.; Ceci, M. Distributed Heterogeneous Transfer Learning for Link Prediction in the Positive Unlabeled Setting. In Proceedings of the 2022 IEEE International Conference on Big Data (Big Data), Osaka, Japan, 17–20 December 2022; pp. 5536–5541. [Google Scholar]
  80. He, Q.; Siu, S.W.; Si, Y. Attentive recurrent adversarial domain adaptation with Top-k pseudo-labeling for time series classification. Appl. Intell. 2022, 53, 13110–13129. [Google Scholar] [CrossRef]
  81. Yu, Z.; Li, J.; Du, Z.; Zhu, L.; Shen, H.T. A Comprehensive Survey on Source-free Domain Adaptation. arXiv 2023. [Google Scholar] [CrossRef]
  82. Long, M.; Zhu, H.; Wang, J.; Jordan, M.I. Deep Transfer Learning with Joint Adaptation Networks. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016. [Google Scholar]
Figure 1. Relevant literature statistics of top conferences.
Figure 1. Relevant literature statistics of top conferences.
Electronics 12 03325 g001
Figure 2. The taxonomy of unsupervised DA based on pseudo-labeling.
Figure 2. The taxonomy of unsupervised DA based on pseudo-labeling.
Electronics 12 03325 g002
Figure 3. The ATDA architecture. (Image: courtesy of Saito et al. [24]).
Figure 3. The ATDA architecture. (Image: courtesy of Saito et al. [24]).
Electronics 12 03325 g003
Figure 4. The rectifying pseudo-label learning via uncertainty estimation for DA semantic segmentation architecture. (Image: courtesy of Zheng et al. [28]).
Figure 4. The rectifying pseudo-label learning via uncertainty estimation for DA semantic segmentation architecture. (Image: courtesy of Zheng et al. [28]).
Electronics 12 03325 g004
Figure 5. The PFAN architecture. (Image: courtesy of Chen et al. [34]).
Figure 5. The PFAN architecture. (Image: courtesy of Chen et al. [34]).
Electronics 12 03325 g005
Figure 6. The hard class rectification for domain adaptation architecture. (Image: courtesy of Zhang et al. [11]).
Figure 6. The hard class rectification for domain adaptation architecture. (Image: courtesy of Zhang et al. [11]).
Electronics 12 03325 g006
Figure 7. The cycle self-training for domain adaptation architecture. (Image: courtesy of Liu et al. [36]).
Figure 7. The cycle self-training for domain adaptation architecture. (Image: courtesy of Liu et al. [36]).
Electronics 12 03325 g007
Figure 8. The CAN architecture. (Image: courtesy of Zhang et al. [39]).
Figure 8. The CAN architecture. (Image: courtesy of Zhang et al. [39]).
Electronics 12 03325 g008
Figure 9. The DupGAN architecture. (Image: courtesy of Hu et al. [47]).
Figure 9. The DupGAN architecture. (Image: courtesy of Hu et al. [47]).
Electronics 12 03325 g009
Figure 10. The self-adaptive reweighted adversarial DA architecture. (Image: courtesy of Wang et al. [58]).
Figure 10. The self-adaptive reweighted adversarial DA architecture. (Image: courtesy of Wang et al. [58]).
Electronics 12 03325 g010
Figure 11. The ILA-DA architecture. (Image: courtesy of Sharma et al. [72]).
Figure 11. The ILA-DA architecture. (Image: courtesy of Sharma et al. [72]).
Electronics 12 03325 g011
Table 1. Classification Accuracy (%) Comparison for Different Pseudo-label Generation Methods on the Office-31 Dataset (ResNet-50).
Table 1. Classification Accuracy (%) Comparison for Different Pseudo-label Generation Methods on the Office-31 Dataset (ResNet-50).
Generation MethodsMethod
( D s     D t )
A → WD → WW → DA → DD → AW → AAvg
BaselinesJAN [82]85.4 ± 0.496.7 ± 0.399.7 ± 0.185.1 ± 0.469.2 ± 0.470.7 ± 0.584.6
SPL [17]92.798.799.893.076.476.889.6
Single-classifierCAT [18]94.4 ± 0.198.0 ± 0.2100.0 ± 0.090.8 ± 1.872.2 ± 0.670.2 ± 0.187.6
PLUE-SFRDA [20]92.598.3100.096.474.572.289.0
SImpAI [27]97.9 ± 0.297.9 ± 0.299.4 ± 0.299.4 ± 0.271.2 ± 0.471.2 ± 0.489.5 ± 0.3
Multi-classifierMCS [67]97.297.299.499.461.361.386.0
CAiDA [44]98.998.999.899.875.875.891.6
Difficult samplesHCRPL [11]95.9 ± 0.298.7 ± 0.1100.0 ± 0.094.3 ± 0.275.0 ± 0.475.4 ± 0.489.9
Table 2. Classification Accuracy (%) Comparison for Different Pseudo-label Application Scenario on the Office-31 Dataset (ResNet-50).
Table 2. Classification Accuracy (%) Comparison for Different Pseudo-label Application Scenario on the Office-31 Dataset (ResNet-50).
Application Scenario Method   ( D s     D t ) A → WD → WW → DA → DD → AW → AAvg
BaselinesJAN [80]85.4 ± 0.496.7 ± 0.399.7 ± 0.185.1 ± 0.469.2 ± 0.470.7 ± 0.584.6
DIAL [52]91.7 ± 0.497.1 ± 0.399.8 ± 0.089.3 ± 0.471.7 ± 0.771.4 ± 0.286.8
MDD + Alignment [56]90.3 ± 0.298.7 ± 0.199.8 ± 0.092.1 ± 0.575.3 ± 0.274.9 ± 0.388.8
SRADA [58]95.298.6100.091.774.573.789.0
DART [60]87.3 ± 0.198.4 ± 0.199.9 ± 0.191.6 ± 0.170.3 ± 0.169.7 ± 0.186.2
ALDA [53]95.6 ± 0.597.7 ± 0.1100.094.0. ± 0.472.2 ± 0.472.5 ± 0.288.7
GAACN [6]90.298.4100.090.467.467.785.6
RSDA-MSTN [62]96.1 ± 0.299.3 ± 0.2100.0 ± 095.8 ± 0.377.4 ± 0.878.9 ± 0.391.1
TSA [65]94.899.1100.092.674.974.489.3
Classifier discriminationPCT [66]94.6 ± 0.598.7 ± 0.499.9 ± 0.193.8 ± 1.877.2 ± 0.576.0 ± 0.990.0
MCS [67]97.297.299.499.461.361.386.0
Mirror [68]98.5 ± 0.399.3 ± 0.1100.0 ± 0.096.2 ± 0.177.0 ± 0.178.9 ± 0.191.7
i-CDD [70]95.4 ± 0.498.5 ± 0.2100.0 ± 0.096.3 ± 0.377.2 ± 0.378.3 ± 0.290.9
ATDOC [71]94.698.199.795.477.577.086.1
ILA-DA [72]95.799.2100.093.372.175.489.3
RWOT [73]95.1 ± 0.294.5 ± 0.299.5 ± 0.2100.0 ± 0.077.5 ± 0.177.9 ± 0.390.8
Fine-tuning [75]91.898.799.989.973.972.087.7
E-MixNet [78]93.0 ± 0.399.0 ± 0.1100.0 ± 0.095.6 ± 0.278.9 ± 0.574.7 ± 0.790.2
iCAN [39]92.598.8100.090.172.169.987.2
CAPLS [41]90.698.699.688.675.476.388.2
CAN [42]94.5 ± 0.399.1 ± 0.299.8 ± 0.295.0 ± 0.378.0 ± 0.377.0 ± 0.390.6
Category feature alignmentHoMM [43]91.7 ± 0.398.8 ± 0.0100.0 ± 0.089.1 ± 0.371.2 ± 0.270.6 ± 0.386.9
CAiDA [44]98.998.999.899.875.875.891.6
ETD [45]92.1100.0100.088.071.067.886.2
BDG [46]93.6 ± 0.499.0 ± 0.1100.0 ± 0.93.6 ± 0.373.2 ± 0.272.0 ± 0.188.5
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Guo, L.; Ge, Y. Pseudo Labels for Unsupervised Domain Adaptation: A Review. Electronics 2023, 12, 3325. https://doi.org/10.3390/electronics12153325

AMA Style

Li Y, Guo L, Ge Y. Pseudo Labels for Unsupervised Domain Adaptation: A Review. Electronics. 2023; 12(15):3325. https://doi.org/10.3390/electronics12153325

Chicago/Turabian Style

Li, Yundong, Longxia Guo, and Yizheng Ge. 2023. "Pseudo Labels for Unsupervised Domain Adaptation: A Review" Electronics 12, no. 15: 3325. https://doi.org/10.3390/electronics12153325

APA Style

Li, Y., Guo, L., & Ge, Y. (2023). Pseudo Labels for Unsupervised Domain Adaptation: A Review. Electronics, 12(15), 3325. https://doi.org/10.3390/electronics12153325

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop