Next Article in Journal
Accuracy of the Fluorescence-Aided Identification Technique (FIT) for Detecting Residual Composite Remnants after Trauma Splint Removal—A Laboratory Study
Next Article in Special Issue
On the Relative Impact of Optimizers on Convolutional Neural Networks with Varying Depth and Width for Image Classification
Previous Article in Journal
Comprehensive Numerical Analysis of a Porous Piezoelectric Ceramic for Axial Load Energy Harvesting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Extending Partial Domain Adaptation Algorithms to the Open-Set Setting

by
George Pikramenos
1,
Evaggelos Spyrou
2,3,* and
Stavros J. Perantonis
2
1
Department of Informatics and Telecommunications, University of Athens, 157 72 Athens, Greece
2
Institute of Informatics and Telecommunications, National Center for Scientific Research—“Demokritos”, 153 41 Athens, Greece
3
Department of Informatics and Telecommunications, University of Thessaly, 351 00 Lamia, Greece
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(19), 10052; https://doi.org/10.3390/app121910052
Submission received: 16 September 2022 / Revised: 28 September 2022 / Accepted: 29 September 2022 / Published: 6 October 2022
(This article belongs to the Special Issue Deep Learning Architectures for Computer Vision)

Abstract

:
Partial domain adaptation (PDA) is a framework for mitigating the covariate shift problem when target labels are contained in source labels. For this task, adversarial neural network (ANN) methods proposed in the literature have been proven to be flexible and effective. In this work, we adapt such methods to tackle the more general problem of open-set domain adaptation (OSDA), which further allows the existence of target instances with labels outside the source labels. The aim in OSDA is to mitigate the covariate shift problem and to identify target instances with labels outside the source label space. We show that the effectiveness of ANN methods utilized in the PDA setting is hindered by outlier target instances, and we propose an adaptation for effective OSDA.

1. Introduction

Supervised learning techniques work under the assumption that training and testing datasets are drawn from the same distribution. Thus, traditional techniques require at least some labeled data for the problem at hand so as to obtain a useful model upon training, which, in many cases, is not possible. Domain adaptation (DA) [1] is a technique for generalizing classifiers to different domains. More specifically, given two datasets with the same label space, with one of them labeled (the source), a DA algorithm may retrieve labels for the other (the target). Note that this does not make any assumption on the underlying distributions of data, yet it is still quite restrictive, since it requires the two datasets to have the same labels.
In practice, we are typically interested in extracting labels from datasets with much larger label spaces than any specific task dataset [2]. To this end, the partial domain adaptation (PDA) [2,3] framework was proposed, in which the assumption is that labels for target data are contained within the larger label space of source data. Of course, the creation of such general-purpose source datasets, which represent the entire label space of any given specific task well, is unrealistic. As an example, some of the largest image datasets currently available [4,5] are comprised of millions of images, but many classes are under-represented.
To address this issue, open-set domain adaptation (OSDA) was proposed [6], essentially combining open-set recognition (OSR) [7] and DA. The aim in OSDA is to automatically extract the information from the source relevant to the target and to identify target instances relevant to the source. That is, a non-empty intersection between the target and source label spaces is assumed, and the goal is to identify source and target instances with labels in the common label subspace. The advantage of identifying such instances is two-fold: It allows us to minimize negative transfer (i.e., the inclusion of source-specific features in the adapted representation) and to identify the subset of target instances that can be reliably labeled through the source.
Note that we cannot hope to predict the label of a target instance belonging to a class not represented in the source label space; however, the advantage of identifying which instances belong to such classes is two-fold. Firstly, it allows us to minimize negative transfer (i.e., the inclusion of source-specific features in the adapted representation) during the adaptation procedure. Secondly, it allows us to identify the subset of target instances that can be reliably labeled using this source.
In this paper, we propose an algorithm for tackling OSDA, which we term the =Doubly Importance Weighted Adversarial Network (DIWAN), which was inspired by the distribution reweighing techniques introduced in [8]. Our algorithm bridges the gap between the adversarial neural network (ANN) techniques used in PDA and OSDA by taking into account outlier target instances during adaptation. We prove that our algorithm constructs a representation in the which source and target distributions of transfer-relevant instances (TRI) are aligned. Moreover, we empirically demonstrate that in the open-set setting, DIWAN outperforms non-adapted versions of the DA and PDA algorithms.
The rest of this paper is organized as follows: In Section 2, we present related work and the contribution of DIWAN. Then, in Section 3, we provide the theoretical background and the methodology used by DIWAN. Moreover, the theoretical support for this work is presented in Section 4. An experimental evaluation of DIWAN is presented and discussed in Section 5, and conclusions are drawn in Section 7.

2. Related Work and Contributions

Our algorithm relies on ANNs for DA, as they were successful in classical DA [9,10] and PDA [2,3,8,11,12]. Such adversarial schemes for DA typically include a source model, which is comprised of a classifier and a representer network, a target representer network, and a domain discriminator. The source and target representer networks are embeddings of the source and target data, respectively, into some latent space. During the adaptation, the target representer and the domain discriminator antagonize each other, while the source representer and classifier are either fixed with pre-trained weights or trained in a supervised fashion.
At each iteration, the discriminator is presented with data from the source and target domains embedded in latent space and trained to discriminate between the two. The target representer is then trained using reversed discriminator gradients. This process can be shown to minimize the Jensen–Shannon Divergence [13] between the distributions of source and target data in latent space, mitigating the covariate shift problem. This standard scheme is typically augmented by other networks to deal with PDA problems. For example, in [2,3], a collection of domain discriminators were used to reweigh source instances so as to ignore outlier classes in the source domain. In [8], a second domain discriminator was used to reweigh source instances in a similar fashion. Moreover, in [11], both a domain discriminator and classifier information were used to separate outlier source classes. Finally, in [12], a shared-label classifier was used to re-weight instances, and it was trained using information from the source classifier.
Contrary to previous works, we reweigh both source and target instances to correct for target outliers. As in [8,11], we re-weight using a second domain discriminator, but our scheme may be modified to use any heuristic for TRIs. Moreover, we use the obtained weights to identify target outlier instances by setting a threshold on the target instance weights.
In [6], OSDA was introduced and an algorithm was proposed for tackling OSDA problems based on a constrained integer programming model. A rejection parameter was used to tune open-space risk tolerance. In [14], the authors introduced a method that relied on classifying outlier target instances as “unknown”, removing the need for a rejection parameter but making classification harder. The model training utilized techniques introduced in [15] for OSR that generated “unknown” target instances.
The main contribution of this work is the adaptation of existing algorithms for PDA to achieve improved performance in the OSDA setting. Our method mitigates the covariate-shift problem only for source and target instances that have labels in the common labels’ subspace. We further show that our algorithm gives rise to a natural heuristic for identifying target instances that are probably transfer relevant. We perform experiments and empirically validate our approach. In Figure 1, we illustrate the effect of domain adaptation in the presence of source and target outliers.

3. Theory and Methodology

The upshot of OSDA is the development of methods for partially labeling an unlabeled dataset by using a model trained on a related dataset. Unlike typical DA, we assume that there exist target instances with labels that are not contained in the source label space, i.e., outlier instances. We are interested in identifying those instances so as to ignore them during the adaptation procedure. The rest of the instances, i.e., transfer-relevant ones, can potentially be reliably labeled through the adaptation procedure.
As in typical DA works, we also assume a machine learning problem described by a domain D = { X , X } and a task T = { Y , Y | X } , where X is a random variable (the covariates) taking values in X and Y is a random variable (the labels) taking values in Y . We assume the existence of two problems ( D S , T S ) and ( D T , T T ) , where (a) X S = X T and (b) P r ( X S ) P r ( X T ) . In the OSDA setting, we assume that Y S Y T and
P ( Y S = y | X S = x , y Y S Y T ) = P r ( Y T = y | X T = x , y Y S Y T ) .
Our aim is to align the distributions of instances in the source and target domains with labels in Y S Y T . Moreover, we want to identify instances that are probably transfer relevant with respect to some confidence measure or heuristic. In the following, we describe DIWAN. Specifically, five neural networks are used: the source and target representer networks M S , M T , a re-weighting network W , a constrained domain discriminator D, and a source classifier C. M S and C are pre-trained in a standard supervised way on the source dataset. The target representer is typically initialized as M S . During training, weights are assigned to each instance x. These are:
w S ( x ) : = 1 W ( M S ( x ) ) 1 E [ W ( M S ( x ) ) ] , w T ( x ) : = W ( M T ( x ) ) E [ W ( M T ( x ) ) ] ,
for the source and target, respectively. These choices are explained in the next section. W S , W T denote the collection of weights for a fixed M S , M T and W . The expectation above is estimated by averages over mini-batches. The objective function used is
min W L W = E x S X S [ log ( W ( M S ( x S ) ) ) ] E x T X T [ log ( 1 W ( M T ( x T ) ) ] .
For W and for M T , D , we have, given W S , W T ,
max M T min D L = E x S X S [ W S log ( D ( M S ( x S ) ) ) ] E x T X T [ W T log ( 1 D ( M T ( x T ) ) ] .
Here, W S ( x ) , W T ( x ) is used to emphasize that the weights are computed for all instances before M T is updated. The pseudocode for DIWAN is provided in Algorithm 1. A visual overview of the five neural networks and the training procedure is illustrated in Figure 2.
Algorithm 1: Doubly Importance Weighted Adversarial Network (DIWAN).
Applsci 12 10052 i001

4. Analysis

We now provide theoretical support regarding the design of the proposed algorithm.
Proposition 1.
Let p 1 and p 2 be distributions over some bounded space X R n and let F : X R be some real-valued function. We define
L ˜ [ F , p 1 , p 2 ] = E x p 1 [ log ( F ( x ) ) ] E x p 2 [ log ( 1 F ( x ) ) ] .
Then,
F p 1 , p 2 * ( x ) = arg min F L ˜ [ F , p 1 , p 2 ] = p 1 ( x ) p 1 ( x ) + p 2 ( x ) .
Proof. 
We will use a variational calculus argument. We denote F = x F = [ F / x 1 , , F / x n ] , define the functional
F ^ p 1 , p 2 [ F ( x ) , F ( x ) , x ] = p 1 ( x ) log ( F ( x ) ) p 2 ( x ) log ( 1 F ( x ) ) ,
and note that L ˜ [ F , p 1 , p 2 ] = X F ^ [ F ( x ) , F ( x ) , x ] d x . We solve the Euler–Lagrange equations, yielding
F ^ F = 0 p 1 ( x ) F * ( x ) = p 2 ( x ) 1 F * ( x ) F * ( x ) = p 1 ( x ) p 2 ( x ) + p 1 ( x )
Corollary 1.
Let p S , p T have bounded support. Then,
W * ( x ) = arg min W L W = p S ( x ) p S ( x ) + p T ( x )
Lemma 1.
Let p 1 , p 2 have bounded support. Let w 1 , w 2 be non-negative weight functions that satisfy E [ w 1 ] = E [ w 2 ] = 1 . Then, L ˜ [ F , w 1 p 1 , w 2 p 2 ] is well defined.
Proof. 
Let P 1 ( x ) = w 1 ( x ) p 1 ( x ) and P 2 ( x ) = w 2 ( x ) p 2 ( x ) . Since X d P 1 ( x ) = E [ w 1 ] = X d P 2 ( x ) = E [ w 2 ] = 1 and w 1 , w 2 0 , these are valid probability densities. Furthermore, for i = 1 , 2 , the support of P i is contained in the support of p i . □
Corollary 2.
Let p S , p T have bounded support. Let w S , w T be as in (2) and let D * be the minimizer of L for a fixed M T . Then,
D * ( x ) = w S ( x ) p S ( x ) w S ( x ) p S ( x ) + w T ( x ) p T ( x )
Proposition 2.
min D L L [ D * ] = log ( 4 ) + J S D ( P w S | | P w T )
Proof. 
We have
min D L = E p S [ w S ( x ) log ( D * ) ] + E p T [ w T ( x ) log ( 1 D * ) ] = ( C o r r o l a r y 1 ) E P w S [ log ( P w S P w T + P w S ) ] + E P w T [ log ( P w T P w T + P w S ) ] = log ( 4 ) + J D S ( P w S | | P w T )
Essentially, we see that OSDA can be cast as an alternating optimization problem. Moreover, we see that a heuristic for identifying probable TRIs can define an algorithm for OSDA by simply normalizing its heuristic scores by using them as weights in L . Intuitively, we aim to down-weight the outliers in the source and target domains. We use an adapted heuristic similar to the one used in [8,11] for re-weighting. The idea behind it is that TRIs will lie near the decision boundary of a “good” domain discriminator, and outlier classes will be far from it; a good domain discriminator will easily distinguish that an outlier source (target) instance belongs to the source (target) domain. In particular, source instances x S for which W * ( x S ) 1 are likely to be outliers, since W * can easily tell that they originate from the source domain. Thus, they should be down-weighted. Similarly, for target instances x T , W * ( x T ) 0 .

5. Experiments

For the empirical verification of our approach, we performed three different experiments. In the first experiment, our aim was to show that the adversarial neural network techniques commonly used in DA are prone to negative transfer when applied in settings with target outlier classes. In particular, we demonstrated that if adaptation is performed using the ADDA algorithm [9], the accuracy on transfer-relevant instances will be lower when target outliers are present.
In the second experiment, our goal was to demonstrate that DIWAN mitigated this negative transfer and to compare the results obtained with ADDA and IWPDA [8]. To this end, we used only transfer-relevant instances to evaluate the methods. Clearly, target outlier instances were misclassified by all three methods and were thus ignored. However, DIWAN offers a heuristic for identifying transfer-relevant instances, and this was evaluated in the final experiment.
Lastly, we demonstrated that we could select a threshold for the assigned weights of target instances after DIWAN was run, such that target instances with weights above this threshold were very likely to be transfer relevant. In particular, we calculated the accuracy of the target model obtained through DIWAN over all data instances with weights above certain thresholds, and we showed that there were threshold choices that could be selected through cross-validation, yielding high-quality classification.

5.1. Dataset and Task Description

Our experiments were conducted for four different DA tasks for images. These were image obstruction, rotation, displacement, and re-scaling. The samples used within these tasks are depicted in Figure 3. Our dataset was comprised of the MNIST [16] and USPS [17] handwritten digit datasets, as in [8,9]. The former contained 28 × 28 pixel gray-scale images, while the latter contained 16 × 16 pixel gray-scale images. Both datasets contained 10 classes corresponding to each of the 10 digits. The USPS images were padded to obtain a homogeneous DA problem. That is, we appended zeros on all four edges of each USPS image so that it became a 28 × 28 pixel image, where the original image occupied the central pixels. In all of our experiments, MNIST was used as a source dataset and USPS was used to generate a synthetic dataset for each of the tasks.
For the obstruction task, we set a 14 pixel area starting from the top-left corner in each image in the USPS dataset to have an intensity of 1. Similarly, each image in USPS was rotated counter-clockwise by 40 for the rotation task. For the re-scaling task, the image was shrunk by a factor of 0.6 on both the vertical and horizontal axes, and then it was re-centered. Finally, for the displacement task, each image was displaced by five pixels to the left and three pixels upwards. The covariate shift introduced in all four tasks was systematic; the same transformation was applied on all USPS images.

5.2. Experiments and Methodology

We started by running experiments to test whether outlier target labels were a source of negative transfer in DA. ADDA was run for each task for 50 trials on a full DA scenario with five source labels. The labels were chosen randomly for each task and trial. We then repeated for a scenario where there were 10 target labels and five source labels, and the accuracy was only measured on transfer-relevant instances. The source network was kept the same for each scenario, and the hyperparameters were tuned to optimize the performance for each task. A convolutional architecture was used for the source representer network. We performed this experiment for all four tasks and plotted average accuracies ± 2 σ against the number of iterations.
For the second experiment, for each task, we varied the number of outlier instances in both the source and target domains, and for each combination, we reported the mean results of four algorithms on ten trials. The algorithms that we used were ADDA, IWPDA, DIWAN, and the source model (without any adaptations). For each trial, the target and source labels were selected randomly. Note also that the number of source labels was always equal to 5.
For our final experiment, recall that once training was finished, we could use W * to calculate (2), where the expectation was replaced by the empirical average over the entire target dataset. We present histograms of weights for each task, where representative setups of labels are chosen. Our aim was to illustrate that the weights tended to be larger for transfer-relevant instances. We then went on to compute the adapted model accuracies after a threshold was imposed. We kept the models from experiment 2 for the case in which there were 10 target labels. The average over the ten trials is presented for each threshold.
For each task, we used four different thresholds, and for each, we computed the accuracy of our model on the subset of the target dataset that we obtained after the cut-off procedure. We further gave the percentage of total transfer-relevant target instances that remained in the dataset after applying the threshold and the percentage of transfer-relevant instances in the remaining dataset.

6. Results and Discussion

6.1. Evidence for Negative Transfer

There was a clear indication that the target outlier classes had a negative impact on the adaptation procedure and could have led to higher instability during training, as indicated by the higher variance obtained for the displacement and re-scaling tasks. The results are presented in Figure 4.

6.2. Effectiveness of DIWAN

The results for our second experiment indicated that full DA would often fail in the OSDA setting. This was observed across all tasks. The PDA model performed better than ADDA, but still performed worse than DIWAN. The results are summarized in Table 1 and Table 2. We further noted that in the partial DA setting (when the target labels were all common labels), DIWAN performed comparably well with respect to IWPDA, while ADDA could completely fail if there were too many outlier source labels.

6.3. Testing Cut-Off Thresholds

Finally, we present our results for using thresholds on the obtained weights to identify probable transfer-relevant instances. Histograms of the weights are presented in Figure 5, Figure 6, Figure 7 and Figure 8. The histograms suggest that the heuristic that we utilized was useful; the transfer-relevant instances (depicted in red) tended to have above-average weights, while the outlier instances (in blue) tended to have below-average weights. The separation was much more apparent in certain tasks (e.g., obstruction). We then demonstrated that a threshold on target weights could be used to select only instances that could be reliably classified from the adapted model. In practice, the threshold should be selected by using cross-validation to fit the needs of the problem at hand; in some cases, we may not mind a few mistakes if we can label more instances correctly, and in other cases, we may only want to produce labels with very high certainty.
In addition, note that, especially when increasing the threshold above 1.0 , many transfer-relevant instances were removed from the dataset, as it may be seen in Table 3. These, however, were transfer-relevant instances that did not benefit much from the adaptation. This was seen because the accuracies on subsets resulting from higher thresholds were better than the overall accuracy on transfer-relevant instances.
For example, for the obstruction task, the adapted model had an accuracy of 76.22 % on transfer-relevant instances, but after a threshold of 1.5 was imposed, the accuracy went up to 99.22 % . For the rescaling task, the accuracy on the dataset with a threshold of 1.5 reaches 100 % . To sum up, using our heuristic not only yielded transfer-relevant instances, but also yielded the “best” transfer-relevant instances in the sense that they were more likely to have benefited from the adaptation. This was not observed for the rotation task, which raises an interesting question that may be investigated in future work.

7. Conclusions

We have extended the popular ANN algorithms used in the DA and PDA settings to work on the OSDA setting. We showed how to utilize heuristics for transfer relevance in order to obtain algorithms for the constrained latent distribution alignment problem and cut-off thresholds for identifying subsets of the target dataset that can be reliably labeled. In particular, we introduced the DIWAN algorithm, which uses an adapted popular heuristic from the literature. Extensive experimentation illustrated the benefits of our method in the OSDA setting. This methodology can be very effective for improving the applicability of machine learning models to real-world data. It could be used for image processing that is generated in an online manner in uncontrolled environments to detect objects or events of interest. Examples include social media content, assisted living, surveillance data, and traffic monitoring. The proposed methodology can also be extended in the future to utilize other weighting methods for target domain classes.

Author Contributions

Conceptualization, G.P.; methodology, G.P., E.S. and S.J.P.; validation, G.P.; writing—original draft preparation, G.P. and E.S.; writing—review and editing, G.P., E.S. and S.J.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was co-financed by the European Union and Greek national funds through the Operational Program Competitiveness, Entrepreneurship, and Innovation, under the call RESEARCH–CREATE–INNOVATE (project code: T1EDK-02070).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ADDAAdversarial Discriminative Domain Adaptation
ANNAdversarial Neural Networks
DADomain Adaptation
DIWANDoubly Importance Weighted Adversarial Network
IWPDAImportance Weighted Partial Domain Adaptation
MNIST datasetModified National Institute of Standards and Technology dataset
OSDAOpen-Set Domain Adaptation
OSROpen-Set Recognition
PDAPartial Domain Adaptation
TRITransfer-Relevant Instance
USPS datasetUnited States Postal Service dataset

References

  1. Ben-David, S.; Blitzer, J.; Crammer, K.; Kulesza, A.; Pereira, F.; Vaughan, J.W. A theory of learning from different domains. Mach. Learn. 2010, 79, 151–175. [Google Scholar] [CrossRef] [Green Version]
  2. Cao, Z.; Long, M.; Wang, J.; Jordan, M.I. Partial transfer learning with selective adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  3. Cao, Z.; Ma, L.; Long, M.; Wang, J. Partial adversarial domain adaptation. In Proceedings of the ECCV, Munich, Germany, 8–14 September 2018. [Google Scholar]
  4. Wu, B.; Chen, W.; Fan, Y.; Zhang, Y.; Hou, J.; Liu, J.; Zhang, T. Tencent ml-images: A large-scale multi-label image database for visual representation learning. IEEE Access 2019, 7, 172683–172693. [Google Scholar] [CrossRef]
  5. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  6. Panareda, B.P.; Gall, J. Open set domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 22–25 July 2017. [Google Scholar]
  7. Geng, C.; Huang, S.J.; Chen, S. Recent advances in open set recognition: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 3614–3631. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Zhang, J.; Ding, Z.; Li, W.; Ogunbona, P. Importance weighted adversarial nets for partial domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  9. Tzeng, E.; Hoffman, J.; Saenko, K.; Darrell, T. Adversarial discriminative domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 22–25 July 2017. [Google Scholar]
  10. Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; March, M. Domain-adversarial neural networks. arXiv 2014, arXiv:1412.4446. [Google Scholar]
  11. Cao, Z.; You, K.; Long, M.; Wang, J.; Yang, Q. Learning to transfer examples for partial domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  12. Hu, J.; Tuo, H.; Wang, C.; Qiao, L.; Zhong, H.; Jing, Z. Multi-Weight Partial Domain Adaptation. In Proceedings of the BMVC 2019, Cardiff, UK, 9–12 September 2019; p. 5. [Google Scholar]
  13. Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; Wiley Series in Telecommunications and Signal Processing; John Wiley Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
  14. Saito, K.; Yamamoto, S.; Ushiku, Y.; Harada, T. Open set domain adaptation by backpropagation. In Proceedings of the ECCV, Munich, Germany, 8–14 September 2018. [Google Scholar]
  15. Ge, Z.; Demyanov, S.; Chen, Z.; Garnavi, R. Generative openmax for multi-class open set classification. arXiv 2017, arXiv:1707.07418. [Google Scholar]
  16. Deng, L. The mnist database of handwritten digit images for machine learning research. IEEE Signal Process. Mag. 2012, 29, 141–142. [Google Scholar] [CrossRef]
  17. Friedman, J.; Hastie, T.; Tibshirani, R. The Elements of Statistical Learning (Volume 1, No. 10); Springer: New York, NY, USA, 2001. [Google Scholar]
Figure 1. Alignment with source and target outliers. (Left): Source and target domain distributions. The blue dashed line denotes a decision boundary before adaptation. (Right): The effect of adaptation; common labels align, while source/target outlier instances are mainly ignored. This is best viewed in color.
Figure 1. Alignment with source and target outliers. (Left): Source and target domain distributions. The blue dashed line denotes a decision boundary before adaptation. (Right): The effect of adaptation; common labels align, while source/target outlier instances are mainly ignored. This is best viewed in color.
Applsci 12 10052 g001
Figure 2. Illustration of the five neural networks involved in the DIWAN algorithm. On the top, the re-weighting network is trained from the source and target representer network representations of the source and target data, respectively. On the bottom, the domain discriminator is trained using the weights obtained from the re-weighting network.
Figure 2. Illustration of the five neural networks involved in the DIWAN algorithm. On the top, the re-weighting network is trained from the source and target representer network representations of the source and target data, respectively. On the bottom, the domain discriminator is trained using the weights obtained from the re-weighting network.
Applsci 12 10052 g002
Figure 3. Sample images from the obstruction, rotation, rescaling, and displacement tasks with different parameters.
Figure 3. Sample images from the obstruction, rotation, rescaling, and displacement tasks with different parameters.
Applsci 12 10052 g003
Figure 4. From top to bottom: Plots of accuracy vs. number of iterations for the obstruction, displacement, rotation, and re-scaling tasks. The mean over 50 trials is drawn with the shaded region, representing ± 2 σ calculated empirically. The source domain included five random classes from the source domain, which were different for each trial. The red-shaded region illustrates the mean accuracy for the experiment where no target outliers were present. The blue-shaded region illustrates the mean accuracy when five outlier labels were added to the target domain. The figure is best viewed in color.
Figure 4. From top to bottom: Plots of accuracy vs. number of iterations for the obstruction, displacement, rotation, and re-scaling tasks. The mean over 50 trials is drawn with the shaded region, representing ± 2 σ calculated empirically. The source domain included five random classes from the source domain, which were different for each trial. The red-shaded region illustrates the mean accuracy for the experiment where no target outliers were present. The blue-shaded region illustrates the mean accuracy when five outlier labels were added to the target domain. The figure is best viewed in color.
Applsci 12 10052 g004
Figure 5. Obstruction task: Transfer-relevant (outlier) target instance histograms are depicted in red (blue). The figure is best viewed in color.
Figure 5. Obstruction task: Transfer-relevant (outlier) target instance histograms are depicted in red (blue). The figure is best viewed in color.
Applsci 12 10052 g005
Figure 6. Rotation task: Transfer-relevant (outlier) target instance histograms are depicted in red (blue). The figure is best viewed in color.
Figure 6. Rotation task: Transfer-relevant (outlier) target instance histograms are depicted in red (blue). The figure is best viewed in color.
Applsci 12 10052 g006
Figure 7. Rescaling task: Transfer-relevant (outlier) target instance histograms are depicted in red (blue). The figure is best viewed in color.
Figure 7. Rescaling task: Transfer-relevant (outlier) target instance histograms are depicted in red (blue). The figure is best viewed in color.
Applsci 12 10052 g007
Figure 8. Displacement task: Transfer-relevant (outlier) target instance histograms are depicted in red (blue). The figure is best viewed in color.
Figure 8. Displacement task: Transfer-relevant (outlier) target instance histograms are depicted in red (blue). The figure is best viewed in color.
Applsci 12 10052 g008
Table 1. The results of Experiment 2 for the obstruction and displacement cases. “T” is the number of target labels and “C” is the number of target labels common to the source. “A” is the performance of ADDA [9], “I” is the performance of IWPDA [8], “D” is the performance of DIWAN, and “S” is the performance of the source model without adaptations. The results are the average accuracy percentages over 10 trials. Best accuracy indicated in bold.
Table 1. The results of Experiment 2 for the obstruction and displacement cases. “T” is the number of target labels and “C” is the number of target labels common to the source. “A” is the performance of ADDA [9], “I” is the performance of IWPDA [8], “D” is the performance of DIWAN, and “S” is the performance of the source model without adaptations. The results are the average accuracy percentages over 10 trials. Best accuracy indicated in bold.
ObstructionDisplacement
TCAIDSAIDS
3390.193.291.434.859.885.784.680.6
4480.183.078.653.274.085.385.166.1
4257.385.687.857.460.260.065.256.4
523.570.788.47.855.360.178.972.2
6260.978.580.149.279.979.290.559.2
4333.663.964.529.383.688.392.769.9
5338.974.974.841.534.635.255.835.3
6367.769.570.961.639.041.459.839.0
7370.274.690.068.634.036.142.635.3
5475.277.385.653.279.281.382.074.4
6479.977.581.869.739.741.666.838.7
10563.167.376.257.066.673.182.370.7
Table 2. Results of Experiment 2 for the rotation and rescaling cases. “T” is the number of target labels and “C” is the number of target labels common to the source. “A” is the performance of ADDA [9], “I” is the performance of IWPDA [8], “D” is the performance of DIWAN, and “S” is the performance of the source model without adaptation. The results are the average accuracy percentages over 10 trials. Best accuracy indicated in bold.
Table 2. Results of Experiment 2 for the rotation and rescaling cases. “T” is the number of target labels and “C” is the number of target labels common to the source. “A” is the performance of ADDA [9], “I” is the performance of IWPDA [8], “D” is the performance of DIWAN, and “S” is the performance of the source model without adaptation. The results are the average accuracy percentages over 10 trials. Best accuracy indicated in bold.
RotationRescaling
TCAIDSAIDS
3366.966.764.862.177.381.381.060.6
4473.374.973.968.877.379.477.962.5
4270.065.168.965.055.867.582.769.1
5274.073.874.572.881.074.585.365.1
6266.160.167.249.372.383.794.683.4
4376.377.077.873.666.085.290.079.2
5371.268.471.762.259.087.993.879.2
6384.883.284.973.344.747.053.346.5
7375.477.281.562.849.156.458.154.1
5470.270.470.869.153.055.056.855.2
6475.677.078.968.870.176.283.074.6
10561.365.867.065.874.277.884.568.6
Table 3. TTR denotes the percentage of total transfer-relevant instances captured by the threshold. TR denotes the percentage of TRIs in the dataset after the threshold is applied.
Table 3. TTR denotes the percentage of total transfer-relevant instances captured by the threshold. TR denotes the percentage of TRIs in the dataset after the threshold is applied.
CaseThresholdTTRTR#Target InstancesAccuracy
Obstruction0100.0047.10414236.12
0.5100.0062.53312047.95
197.5480.13237562.95
1.532.3999.2263798.43
Rescaling0100.0058.42445537.15
153.5178.39177767.36
1.2532.8598.1687192.66
1.520.86100.00543100.00
Rotation0100.0042.96324742.22
155.8459.19131638.98
1.2521.5786.2434950.43
1.55.9698.969735.16
Displacement0100.0043.37600249.92
161.9365.90244660.95
1.2530.4392.6385584.21
1.57.9996.7421691.16
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Pikramenos, G.; Spyrou, E.; Perantonis, S.J. Extending Partial Domain Adaptation Algorithms to the Open-Set Setting. Appl. Sci. 2022, 12, 10052. https://doi.org/10.3390/app121910052

AMA Style

Pikramenos G, Spyrou E, Perantonis SJ. Extending Partial Domain Adaptation Algorithms to the Open-Set Setting. Applied Sciences. 2022; 12(19):10052. https://doi.org/10.3390/app121910052

Chicago/Turabian Style

Pikramenos, George, Evaggelos Spyrou, and Stavros J. Perantonis. 2022. "Extending Partial Domain Adaptation Algorithms to the Open-Set Setting" Applied Sciences 12, no. 19: 10052. https://doi.org/10.3390/app121910052

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop