1. Introduction
With the development of artificial intelligence, especially the emergence of neural networks, artificial intelligence technology driven by big data has played its role in many complex fields, such as automatic driving, medical care, and finance. People want this technology to be realized in all walks of life. However, the actual situation is very disappointing. Except for a few limited fields, many industries have limited data and poor data quality, which is insufficient to support the implementation of artificial intelligence technology [
1]. Even if there are many data in more application fields, these data are distributed among different data sources and marked unevenly. For example, there are 537,100 schools of all levels and types in China, with 289 million students. These students’ data are stored in different schools and on various online education platforms. These data sources are not connected to each other, and the data exist in the form of isolated islands. It is almost impossible to integrate these data, or the cost required is enormous.
On the other hand, with the further development of big data, data privacy and security have become a worldwide trend. Because of IoT devices’ increasing storage and computing capabilities, it is common to save data and models on local devices [
2]. Federated learning is a new frontier field [
1,
3,
4]. Google has launched a federated learning system [
3,
4,
5], in which a federated of distributed participants updates the global machine learning model and their data are stored locally. Their framework requires all contributors to share the same feature space. On the other hand, secure machine learning for partitioning data in feature space is also studied [
6,
7,
8], which finds that, in machine learning, each participant can make joint modeling with the help of other parties’ data. However, all parties do not need to share data resources. They can conduct federated data training and establish a shared machine learning model when the data do not come out locally. Federated learning enables multiple participants to continue machine learning and deep learning in protecting data privacy and meeting legal compliance requirements to solve the data island problem. Federated learning technology can be used among mobile terminals, wearable devices, and different platforms to realize a training machine learning model that pays more attention to privacy and efficiency. However, when one party collects a lot of unlabeled data, even when the features are statistically different from those collected by the other party, domain shift will arise [
9]. This makes the model obtained by federated learning and training unable to be extended to new equipment. For example, in the field of education, two different online education platforms save local student data. One platform may collect more student behavior logs, while the other platform saves students’ personal information.
In the field of educational data mining, the classification of students’ grades is usually limited by the size of datasets and the number of data features. Moreover, because of data privacy protection, each data platform cannot carry out practical federated training. Therefore, it is an urgent problem to solve the imbalance of data and the inconsistency of characteristics among data owners to train student models.
This paper proposes a framework based on federated transfer learning for students’ grades classification with privacy protection. In this paper, heterogeneous domain adaptive technology is introduced into federated transfer learning, and knowledge transfer between different nodes is solved by exchanging encryption parameters between data. In this framework, each data owner uses the network of label classifier and domain classifier for training and aligns the features of all parties in the new feature space for transfer learning. Furthermore, according to the actual situation, we also simulate the following four scenarios by constraining different characteristics and data scales:
Scenario 1: Both Party A and Party B have the same feature distribution and data scales;
Scenario 2: Party A and Party B have the same characteristics but different data scales;
Scenario 3: Party A and Party B have different characteristics but the same data scales;
Scenario 4: Both Party A and Party B have different feature distribution and data scales.
We conducted experiments on datasets from two middle schools in Portugal and obtained the training results of student models in these four situations. Experimental results show that both parties with different data feature spaces and inconsistent data scales can effectively transfer knowledge under this framework. It can help the target data party with missing data labels or fewer feature data to train and improve the generalization ability of the model of main data participants.
The rest of this paper is organized as follows: In
Section 2, we outline related work. In
Section 3, we present the details of the model proposed in this paper. Then, in
Section 4, we describe the experimental details. The analysis of the experimental results is in
Section 5. Finally, in
Section 6, we summarize our work and present some ideas for future work.
2. Related Work
2.1. Domain Adaptation
Domain adaptation is an active research topic in many fields of artificial intelligence, including machine learning, natural language processing, and computer vision. The methods of early domain adaptation mainly focus on constructing domain-invariant features. This can be achieved by the weight and selection mechanism of features [
10,
11] or by learning the explicit feature transformation that aligns the distribution of source domain and target domain [
12,
13,
14]. In recent work, deep neural networks have been used to learn powerful representation and perform unsupervised domain adaptation [
15]. Revgrad [
16] uses a domain classification network, which aims to distinguish between source and target embedding. The goal of the feature extraction network is to produce embedding that maximizes the loss of domain classifiers and minimizes the loss of tag prediction. This is achieved by negating the gradient from the domain classification network.
Driven by GAN [
17], many transfer learning methods are based on the assumption that a good feature representation contains almost no specific information about the original domain of an instance. For example, Ganin et al. proposed a depth structure called Domain-Adversarial Neural Network (DANN) for domain adaptation [
16,
18]. Its architecture consists of a feature extractor, tag predictor, and domain classifier. DANN can be trained by inserting a particular gradient reversal layer (GRL). After the training of the whole system, the feature extractor learns the deep features of instances, and the output classifies the prediction tags of unmarked target domain instances.
In the field of domain adaptation, some researchers have improved and applied domain adaptation technology. Cao et al. put forward an adaptive model with semantic consistency (ADASC), which gradually and effectively aligns the distinguishing features across domains using the class hierarchy relationship between domains [
19]. Hu et al. developed a framework DATSING based on transfer learning, which effectively utilized the potential representation of cross-domain time series to enhance target domain prediction [
20]. Su et al. introduced the task of continuous domain adaptation in machine reading comprehension [
21].
There are some other related impressive works. Zeng et al.’ s work puts forward a unified framework for adaptation in adversarial fields [
15]. Shen et al. used Wasserstein distance for domain adaptation [
22]. Hoffman et al. adopted cycle-consistency loss to ensure the consistency of structure and semantics [
23]. Long et al. put forward the condition domain adversarial network, which uses the condition domain discriminator to assist the adversarial adaptation [
24]. Zhang et al. adopted the symmetrical design of source classifier and target classifier [
25]. Zhao et al. used a domain confrontation network to solve the problem of multi-source transfer learning [
26]. Yu et al. proposed a dynamic adversarial adaptation network [
27]. Chen et al. proposed a reinforcement learning (DARL) framework combined with adversarial learning for partial domain adaptation [
28]. The DARL framework adopts the deep Q-learning method to learn the strategy by approximating the action value function, so that the agent can make a choice decision. In order to minimize the domain differences in cross domain knowledge transfer, Chen et al. used self similarity consistency (SSC) constraints to match the deep feature structures in the source and target domains, rather than aligning the global distribution statistics across domains [
29]. Tran et al. proposed a generative adversary domain adaptive (GADA) algorithm, which significantly improved the process of distinguishing feature extraction by injecting additional classes and using the generated samples for training [
30]. Guo et al. proposed a new distribution calibration method by learning an adaptive weight matrix between new samples and base classes, which built on the Hierarchical Optimal Transport (H-OT) framework by minimizing the high-level OT distance between the new sample and the base class to convey the adaptive weight information of the base class [
31]. In the multi-source domain adaptation task, Deng et al. proposed a two-layer optimization-based robust object training (BORT2) [
32]. BORT2 first learns a labeling function using source and target data, and then trains a noise-robust model only on the pseudo-labeled target domain. Noise-robust models exploit feature uncertainty to detect label noise and mitigate its negative effects.
To sum up, although there are many kinds of research in the field of domain adaptation in recent years, this research is all based on typical scene training. Their data are stored on the same server, which leads to incompatibility under the condition of federated learning. Therefore, our research hopes to achieve domain adaptation under the framework of federated learning.
2.2. Federated Learning
Google recently proposed the concept of federated learning [
3,
4,
5]. Their main idea is to build a machine learning model based on datasets distributed on multiple devices while preventing data leakage. Recent improvements focus on overcoming statistical challenges [
2,
33] and improving the security of federated learning [
34,
35], and there are some research efforts to make federated learning more personalized [
36]. All of the above work focuses on device federated learning, involving distributed mobile user interaction and communication cost, unbalanced data distribution, and device reliability in large-scale distribution are some of the main factors of optimization. To extend the concept of federated learning to cover collaborative learning scenarios among organizations, Yang et al. expanded the original federated learning into the general idea of all decentralized collaborative machine learning technologies for privacy protection [
1]. They made a comprehensive overview of federated learning and federated transfer learning technologies. At the same time, they further investigated the related security foundation and discussed the relationship with other associated fields, such as multi-agent theory and privacy protection data mining. Cheng et al. proposed a new lossless privacy protection algorithm secureboost, which is used to train high-quality tree enhancement models when the training data are kept secret on multiple parties [
7]. Chandra et al. proposed a new distributed machine learning method, called splitfed. Due to network fragmentation, it provides a higher level of privacy than federated learning. It is suitable for machine learning (ML) with low computational resources, fast training, and analysis of private and sensitive data [
37]. Mahmood et al. studied the feasibility of data and model poisoning attacks under the blockchain based FL system built next to the Ethereum network and the traditional FL system (no blockchain) [
38]. They proposed a transparent incentive mechanism to fill the knowledge gap, which can encourage good behavior among participating decentralized nodes, avoid common problems, and provide knowledge for FL security literature by studying the current FL system.
In addition, the time required for computing is also an important factor. Federated learning (FL) is largely affected by the heterogeneity of clients. The training speed of clients is inconsistent. In order to reduce the impact of slow clients, Cox et al. proposed a new FL algorithm using model freezing and unloading, which uses a simple, fast, and scalable scheduling algorithm to periodically organize model unloading between clients [
39]. To reduce the communication and computation delays of model parameters, Liu et al. proposed a new iterative algorithm and minimized the total model parameter communication and computation delays by optimizing the number of local iterations and edge iterations [
40].
Recent research has focused on transferring knowledge from decentralized nodes to new nodes with different data domains. Liu et al. introduced new technology and framework called federated transfer learning (FTL) to improve the statistical model under the data alliance. The alliance allows knowledge sharing without compromising users’ privacy and allows supplementary knowledge to be transmitted in the network [
8]. Therefore, the target domain can build a more flexible and robust model using rich tags from the source domain. Peng et al. put forward a principled method to solve the federated domain adaptation problem, which aims to make the representations learned between different nodes consistent with the data distribution of target nodes [
41].
However, we do not think Peng’s method extends to the constraint of federated learning. It cannot be applied to the general environment of federated learning. Therefore, in this paper, we combine domain adaptation with secure federated learning and propose a new federated transfer learning framework based on domain adaptation to solve knowledge transfer in federated transfer learning.
3. Method
In this section, we introduce the details of our proposed method. The network we use is a deep neural network. The network structure we propose is a deep neural network with three layers whose structure is shown in
Figure 1. The input data are passed through a feature extractor to match the feature distribution of each party in the feature space. Then, the labels and domains are classified on each side, the model is trained, and the target model is updated by gradient aggregation.
3.1. Problem Definition
In the actual educational data mining, we often encounter the situation of small data scale and lack of data features and need to manually mark them. However, due to the data privacy protection, in many cases, the data owners can not carry out collaborative training. To solve this problem, we simulated a federated training scenario between Party A and Party B.
Consider two data owners, a source domain where , is the ith label and is the domain label, and a target domain where , is the domain label. represents the number of samples of the source domain and target domain, respectively. The total number of samples is .
In our definition, Party A and Party B represent two platforms with educational data. Among them, Party A has complete data, label information, and class labels. Party A is the leading party in federated learning and the owner of the source domain data in federated transfer learning. Party B only has some data and class labels and has no data labels. Play the role of the customer in federal migration learning. Party A and Party B use the knowledge gained from federal migration learning to build a model and predict their users.
Our goal is to establish a transfer learning model, accurately predict the category of students of Party B, and improve the classification effect of Party A. At the same time, their data will not be exposed. In this paper, domain adaptation is introduced into federated migration learning, which minimizes and maximizes the classification errors of the source domain, and then exchanges knowledge between the source domain and the target domain. Then, this federated learning problem can be formulated as follows:
where
is the parameter of the model,
f represents the loss function of the label classifier, and
represents the loss function of the domain classifier.
Our goal is to exchange only intermediate parameters to achieve the following goals on privacy protection:
3.2. Approach Statement
We assume that there are two data owners training model A and model B. Firstly, the model needs feature extraction to generate two implicit representations and , where and , and d represent the dimensions of the hidden layer. is the feature transformation function, where matrix-vector pair . The neural network and transform their features, and project the features of and into a common feature subspace, where they can exchange knowledge.
Then, the source domain data of feature space is classified by the prediction layer, and the correct labels are separated as far as possible. The prediction function ,
where
. Thus, we obtain the first optimization goal:
where
are training parameters of layers
.
is classification loss, using the negative log-probability of the correct labels.
The network also needs a regularization term, that is, a metric criterion so that the source domain and the target domain in the feature space are as close as possible. We use H-divergence to measure [
42,
43,
44]:
where
is a hypothetical class, binary classifiers
.
According to formula (
5), the empirical H-divergence between the source domain and the target domain can be calculated by the following formula:
It is usually difficult to calculate
. According to the inspiration of Ben-David’s research [
42,
43], we can define the approximate solution of the classification problem of the source domain and target domain, that is, domain classifier. It classifies data in feature space and tries to distinguish which domain the data comes from. Domain classifier
, and we can make the feature distribution of the data in the source domain and the target domain as close as possible by approximately calculating the empirical H-divergence between the two samples. We can obtain the second optimization target:
where
are training parameters of layers
.
is the domain classification label.
is the loss of domain classification layer, and the loss of feature alignment is calculated by cross entropy.
The final objective function is:
where
is the weight parameters. For
, the stochastic gradients can be computed as the following:
3.3. Federated Domain Adaptation
In order to realize secure computation without exchanging data, additively homomorphic encryption is adopted. We define a number
x, which becomes
after additively homomorphic encryption. According to the properties of homomorphic encryption, for any two numbers
x and
y,
and
is valid. However, the loss function and gradient can not be directly calculated by additively homomorphic encryption. In order to solve this problem, we use the second-order Taylor approximation for the logistic loss
and the gradient is
In this way, we obtain the encrypted loss and gradient.
Then, update the parameters according to the gradient direction
where
is the learning rate. With these parameters, we now design a federated Domain Adaptation network for solving the Domain Adaptation problem.As shown in
Figure 2. First, Party A and Party B send public keys to each other, then initialize parameters
, calculate implicit representations
and
, and then calculate, encrypt, and exchange intermediate results
g and
.
To prevent exposure of the gradients of A and B, A and B further mask each gradient with encrypted random values. Then, they send the gradient and loss of the mask to each other and locally decrypt these values. Once the loss convergence condition is met, Party A can send a termination signal to Party B. Otherwise, they will unmask the gradients, update the model parameters with their gradients, and then continue the next iteration. Algorithm 1 summarizes the process.
Algorithm 1 Federated domain adaptation |
Input: learning rate: , adaptation parameter: Output: Model parameters Initialization:Patty A&B initialize while Convergence conditions is not met do for all each iteration do Part A&B compute: A&B create a random mask: A compute and encrypt: and send to B B compute and encrypt: and send to A A&B decrypt gradients and exchange, unmask and update model locally end for end while return
|
3.4. Security Analysis
In this section, we aim to determine whether one party participating in federal training can learn the other party’s data from the parameter set
exchanged during training. Previous studies mainly introduced the leakage of the whole set of model parameters or gradient in the training process [
45,
46,
47]. However, in our framework, the model’s parameters are confidential, and only the intermediate results in the tag prediction layer may reveal information.
Theorem 1. According to our definition, the federated transfer learning framework based on domain heterogeneous adaptation is secure, provided that the underlying encryption parameter sharing scheme is secure.
Proof. In our framework, the model’s parameters are confidential, and only the intermediate results in the label classifier layer may reveal information. In the domain classifier layer, each party can only receive the information of the domain but not the information of the data. Let
be the set of data points sampled in the
t-th iteration and
be the
i-th sample in the
j-th iteration.
refers to the contribution of the
i-th sample to other parties. In the
-th iteration, we can obtain the weight update of the label classifier according to Equation (
19):
where
.For any participant
k with unpublished datasets
and training parameters
following our defined framework, there is an infinite solution of
that produces the same contribution set
. That is to say, no matter the number of iterations, the data of participant
k cannot be uniquely determined from its exchanged information
. This definition of security is consistent with the previous definition of security proposed in privacy protection and secure multiple computation of machine learning [
48]. Under this definition, when some prior knowledge about data are known, the other party may exclude some alternative solutions but cannot infer accurate original data. This model provides a flexible trade-off between privacy and efficiency. □
Theorem 2. According to our definition, the federated transfer learning framework based on domain heterogeneous adaptation is secure, provided that the homomorphic encryption scheme used is secure.
Proof. In our scheme, all the learning and training data are encrypted by a homomorphic encryption mechanism. The parameters and intermediate data of the model are encrypted in the whole federated learning process. In each iteration, both sides will create a new random encryption mask m. This randomness and confidentiality also ensure the security of the whole training and learning process. Therefore, as long as the encryption scheme is secure, the federated learning framework we defined is also secure. □
4. Experiments
In this section, we mainly introduce the usage of our proposed framework and verify it by constraining different features, quantity, and the correlation between features on the datasets of two courses.
4.1. Datasets and Experimental Settings
The data used in the experiment come from the students’ achievements in secondary education in two Portuguese schools. Data attributes include student achievement, demographic, social, and school-related characteristics collected using school reports and questionnaires. The two datasets provide the performance of two different subjects: mathematics and Portuguese. There are 396 students in the mathematics dataset and 650 students in the Portuguese dataset. We selected 16 attributes, which are described in
Table 1. Students are classified in a binary system, and the G3 grades provide the classification labels of students. Students with scores greater than ten are classified into one category, while others are classified.
In our experiments, we set the depth of DNN layers as three in the model architecture. The learning rates are set to 0.0001, and batch size is set to 128. To avoid overfitting, we apply dropout on the DNN layers with a drop rate of 0.3.
4.2. Privacy Protection
In order to ensure the privacy of the framework, the experiment simulates the distributed training scenario of multiple nodes, and the data of Party A and Party B are stored separately in each node to participate in the federated learning. In order to promote secure computing between the two parties, a semi-honest third-party server is introduced, and it is assumed that it does not collude with other participants. After learning and training, each participant can only hold model parameters associated with their characteristics. Therefore, when reasoning, both parties need to work together to generate the output. The third-party server does not obtain the data of all parties but only distributes the secret key and exchanges the encryption parameters.
4.3. Design
In the experiment, according to the datasets size and data feature distribution of Party A and Party B, four scenarios were established to simulate the actual situation. They are as follows:
Scenario 1: Both Party A and Party B have the same feature distribution and data scales;
Scenario 2: Party A and Party B have the same characteristics but different data scales;
Scenario 3: Party A and Party B have different characteristics but the same data scales;
Scenario 4: Both Party A and Party B have different feature distribution and data scales.
To correspond to these four scenarios, we processed the data and selected some data and attributes. In addition, in the scene with different feature distribution, the correlation analysis of attributes is carried out to ensure that the correlation with domain invariant features does not exceed the threshold. We carry out a factor analysis on the features contained in the datasets. The calculated factor scores are shown in
Table 2, indicating eight principal components in total.
Given the scenarios with consistent feature distribution, we selected eight variables that contributed the most to the principal components as the input features of the model, namely walc, fedu, paid, address, romantic, famrel, famsize, and activities. The source domain still adopts the eight features selected from previous scenarios for scenarios with inconsistent features, while the target domain adopts the new eight features. They are dalc, absences, medu, goout, higher, freetime, studytime and internet. We visualize these two groups of features, use PCA to reduce the dimension to three-dimensional space, and check the distribution of data points.
Figure 3 respectively shows the distribution of the features of the source domain and the target domain in the three-dimensional space. To ensure the domain invariant features in the transfer process, we also conducted a pairwise correlation analysis on a total of 16 features and tested the significance of the correlation results. The results are shown in
Table 3.
It can be seen from
Figure 3 that the feature spaces of the source domain and the target domain are inconsistent, so the two datasets are heterogeneous in features, which guarantees our hypothesis. At the same time, to ensure the transfer of valuable knowledge in the transfer process, the correlation between features is analyzed. From
Table 3, there is a specific correlation between the features. The first row of each feature in
Table 3 represents the Pearson correlation coefficient between variables, and the second row represents the
p-value of the significance hypothesis. From the significance test results, there are significantly associated features between the two domains, which ensures the extraction of domain invariant features in the process of domain transfer and the effectiveness of transfer learning.
For the scenario with the same data scale, both the source domain and the target domain used the data of 200 students. In the case of an inconsistent data scale, the source domain uses the data of 200 students, and the target domain uses the data of 80 students.
From the results, we can see that both the source domain and the target domain can make classification predictions when they are not trained. The accuracy is 0.5, which accords with the common sense of binary classification. The accuracy of the domain classifier in each scene is between 0.5 and 0.6, which indicates that the machine can not accurately distinguish the features of the source domain or the target domain, displaying that knowledge has transferred.
5. Results and Discussion
In the experiment, we use the logistic regression model to classify the students’ academic performance in binary, the network structure is shown in
Figure 1, and we use the federated transfer learning framework based on heterogeneous domain adaptation defined by this paper.
In each scenario, we conducted three pieces of training, namely, training only for the source domain, training only for the target domain, and transfer training. We give the classification accuracy of the source domain and target domain under each training mode and give the classification accuracy of the domain under the transfer training mode. The results are shown in
Table 4.
We can find that, in Scenarios 2–4, the results of target-only training are generally a little worse than source-only training due to the difference in the data size of the source and target domains and the different contributions of the included features to the principal components. Moreover, in Scenario 1, after using domain adaptation training, the accuracy results are not much different from the results of individual training in each domain, which indicates that using domain adaptation training will not improve the model effect in this case. In Scenario 2, after using domain adaptation, the classification accuracy of the target domain is greatly enhanced. In Scenario 3, both the source domain’s classification accuracy and the target domain have dramatically increased. However, in Scenario 4, the classification accuracy of the source domain and the target domain increased less than that of Scenario 3.
The result analysis shows that the domain adaptation model has little effect when the feature distribution and scale of the source domain and the target domain are consistent. However, when one party’s data scale is small, it can benefit from federated learning. Especially when the feature distribution of both parties is inconsistent, both parties can learn new knowledge from each other and improve the classification accuracy. In Scenario 4, although we project the source and target domains into a common feature subspace, since the source and target domains have different feature distributions and data scales, the training knowledge on the source domain is directly applied to the target domain, and the resulting effect will be inferior to Scenario 2 and Scenario 3.
From the experimental results, the federated transfer learning framework designed by us can effectively solve the problem of federated learning with small unilateral data and missing unilateral features in students’ grades classification, and all participants can obtain new knowledge from each other while maintaining privacy. Experiments show that applying a federated learning framework can solve the problem of data islands in educational data mining, which can be used for reference in the field of intelligent education.
The study has potential limitations. Since the experimental part of this paper only verifies the data exchange between two different education platforms, and the data differences generated by different education platforms may be large, how to transfer privacy protection knowledge between more online education platforms still needs further exploration.
6. Conclusions
This paper proposes a framework based on federated transfer learning for students’ grades classification with privacy protection. By introducing feature extractors and domain classifiers, the framework matches the feature distribution of all parties in collaborative learning in some feature spaces. It uses homomorphic encryption and random masking to ensure data privacy during transmission. The experimental results show that both parties with different data feature spaces and inconsistent data scales can effectively transfer knowledge under this framework. It can help the target data parties with missing data labels, or fewer feature data train the students’ grades classification model and improve the generalization ability and classification effect of the students’ grades classification model of the primary data participants.
In future work, we will further study privacy protection in the framework. This includes studying privacy protection in the federated learning framework in the presence of malicious participants, and more efficient feature extraction methods. We are committed to expanding federated learning technology to more fields and exploring more effective federated learning methods. These methods can include studying and further improving this method’s communication efficiency and effectiveness in more complex and asynchronous federated learning systems.
Author Contributions
Funding acquisition, B.X.; methodology, S.Y. and B.X.; conceptualization, S.Y.; software, S.Y.; validation, S.Y. and B.X.; writing, review and editing, S.Y., S.L. and Y.D.; supervision, S.L. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the National Natural Science Foundation of China, grant (U1811261), and the Fundamental Research Funds for the Central Universities of China, grant (N2116019), and the Liaoning Natural Science Foundation, grant (2022-MS-119).
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Not applicable.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Yang, Q.; Liu, Y.; Chen, T.; Tong, Y. Federated Machine Learning: Concept and Applications. ACM Trans. Intell. Syst. Technol. 2019, 10, 1–19. [Google Scholar] [CrossRef]
- Smith, V.; Chiang, C.K.; Sanjabi, M.; Talwalkar, A. Federated Multi-Task Learning. arXiv 2018, arXiv:1705.10467. [Google Scholar]
- Konečný, J.; McMahan, H.B.; Yu, F.X.; Richtárik, P.; Suresh, A.T.; Bacon, D. Federated Learning: Strategies for Improving Communication Efficiency. arXiv 2017, arXiv:1610.05492. [Google Scholar]
- McMahan, H.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Proceedings of the AISTATS, Fort Lauderdale, FL, USA, 20–22 April 2017. [Google Scholar]
- Konečný, J.; McMahan, H.B.; Ramage, D.; Richtárik, P. Federated Optimization: Distributed Machine Learning for On-Device Intelligence. arXiv 2016, arXiv:1610.02527. [Google Scholar]
- Hardy, S.; Henecka, W.; Ivey-Law, H.; Nock, R.; Patrini, G.; Smith, G.; Thorne, B. Private federated learning on vertically partitioned data via entity resolution and additively homomorphic encryption. arXiv 2017, arXiv:1711.10677. [Google Scholar]
- Cheng, K.; Fan, T.; Jin, Y.; Liu, Y.; Chen, T.; Papadopoulos, D.; Yang, Q. Secureboost: A lossless federated learning framework. IEEE Intell. Syst. 2021, 36, 87–98. [Google Scholar] [CrossRef]
- Liu, Y.; Kang, Y.; Xing, C.; Chen, T.; Yang, Q. A Secure Federated Transfer Learning Framework. IEEE Intell. Syst. 2020, 35, 70–82. [Google Scholar] [CrossRef]
- Quiñonero-Candela, J.; Sugiyama, M.; Lawrence, N.D.; Schwaighofer, A. Dataset Shift in Machine Learning; MIT Press: Cambridge, CA, USA, 2009. [Google Scholar]
- Huang, J.; Smola, A.J.; Gretton, A.; Borgwardt, K.M.; Scholkopf, B. Correcting Sample Selection Bias by Unlabeled Data. In Proceedings of the 19th International Conference on Neural Information Processing Systems, NIPS’06, Vancouver, BC, Canada, 4–7 December 2006; MIT Press: Cambridge, MA, USA, 2006; pp. 601–608. [Google Scholar]
- Daumé, H., III. Frustratingly Easy Domain Adaptation. In Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics, Prague, Czech Republic, 23–30 June 2007; Association for Computational Linguistics: Prague, Czech Republic, 2007; pp. 256–263. [Google Scholar]
- Gong, B.; Shi, Y.; Sha, F.; Grauman, K. Geodesic flow kernel for unsupervised domain adaptation. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2066–2073. [Google Scholar] [CrossRef] [Green Version]
- Gopalan, R.; Li, R.; Chellappa, R. Domain adaptation for object recognition: An unsupervised approach. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 999–1006. [Google Scholar] [CrossRef]
- Peng, X.; Saenko, K. Synthetic to Real Adaptation with Generative Correlation Alignment Networks. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 4–8 January 2022; pp. 1982–1991. [Google Scholar] [CrossRef] [Green Version]
- Tzeng, E.; Hoffman, J.; Saenko, K.; Darrell, T. Adversarial discriminative domain adaptation. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7167–7176. [Google Scholar]
- Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M.; Lempitsky, V. Domain-Adversarial Training of Neural Networks. J. Mach. Learn. Res. 2016, 17, 1–35. [Google Scholar]
- Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. In Proceedings of the 27th International Conference on Neural Information Processing Systems—Volume 2, NIPS’14, Montreal, QC, Canada, 8–13 December 2014; MIT Press: Cambridge, MA, USA, 2014; pp. 2672–2680. [Google Scholar]
- Ganin, Y.; Lempitsky, V. Unsupervised domain adaptation by backpropagation. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; PMLR: Westminster, UK, 2015; pp. 1180–1189. [Google Scholar]
- Cao, M.; Zhou, X.; Xu, Y.; Pang, Y.; Yao, B. Adversarial Domain Adaptation with Semantic Consistency for Cross-Domain Image Classification. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, CIKM ’19, Beijing, China, 3–7 November 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 259–268. [Google Scholar] [CrossRef] [Green Version]
- Hu, H.; Tang, M.; Bai, C. DATSING: Data Augmented Time Series Forecasting with Adversarial Domain Adaptation. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, CIKM ’20, Virtual, 19–23 October 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 2061–2064. [Google Scholar] [CrossRef]
- Su, L.; Guo, J.; Zhang, R.; Fan, Y.; Lan, Y.; Cheng, X. Continual Domain Adaptation for Machine Reading Comprehension. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, CIKM ’20, Virtual, 19–23 October 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 1395–1404. [Google Scholar] [CrossRef]
- Shen, J.; Qu, Y.; Zhang, W.; Yu, Y. Wasserstein distance guided representation learning for domain adaptation. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
- Hoffman, J.; Tzeng, E.; Park, T.; Zhu, J.Y.; Isola, P.; Saenko, K.; Efros, A.; Darrell, T. Cycada: Cycle-consistent adversarial domain adaptation. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; PMLR: Westminster, UK, 2018; pp. 1989–1998. [Google Scholar]
- Long, M.; Cao, Z.; Wang, J.; Jordan, M.I. Conditional Adversarial Domain Adaptation. In Proceedings of the Advances in Neural Information Processing Systems; Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2018; Volume 31. [Google Scholar]
- Zhang, Y.; Tang, H.; Jia, K.; Tan, M. Domain-symmetric networks for adversarial domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5031–5040. [Google Scholar]
- Zhao, H.; Zhang, S.; Wu, G.; Moura, J.M.; Costeira, J.P.; Gordon, G.J. Adversarial multiple source domain adaptation. Adv. Neural Inf. Process. Syst. 2018, 31, 8559–8570. [Google Scholar]
- Yu, C.; Wang, J.; Chen, Y.; Huang, M. Transfer learning with dynamic adversarial adaptation network. In Proceedings of the 2019 IEEE International Conference on Data Mining (ICDM), Beijing, China, 8–11 November 2019; pp. 778–786. [Google Scholar]
- Chen, J.; Wu, X.; Duan, L.; Gao, S. Domain adversarial reinforcement learning for partial domain adaptation. IEEE Trans. Neural Netw. Learn. Syst. 2020, 33, 539–553. [Google Scholar] [CrossRef]
- Chen, C.; Fu, Z.; Chen, Z.; Cheng, Z.; Jin, X. Towards self-similarity consistency and feature discrimination for unsupervised domain adaptation. Signal Process. Image Commun. 2021, 94, 116232. [Google Scholar] [CrossRef]
- Tran, H.H.; Ahn, S.; Lee, T.; Yi, Y. Enlarging Discriminative Power by Adding an Extra Class in Unsupervised Domain Adaptation. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 1812–1819. [Google Scholar]
- Guo, D.; Tian, L.; Zhao, H.; Zhou, M.; Zha, H. Adaptive Distribution Calibration for Few-Shot Learning with Hierarchical Optimal Transport. arXiv 2022, arXiv:2210.04144. [Google Scholar]
- Deng, Z.; Li, D.; Song, Y.Z.; Xiang, T. Robust Target Training for Multi-Source Domain Adaptation. arXiv 2022, arXiv:2210.01676. [Google Scholar]
- Zhao, Y.; Li, M.; Lai, L.; Suda, N.; Civin, D.; Chandra, V. Federated Learning with Non-IID Data. arXiv 2018, arXiv:1806.00582. [Google Scholar] [CrossRef]
- Bonawitz, K.; Ivanov, V.; Kreuter, B.; Marcedone, A.; McMahan, H.B.; Patel, S.; Ramage, D.; Segal, A.; Seth, K. Practical Secure Aggregation for Privacy-Preserving Machine Learning. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, CCS ’17, Dallas, TX, USA, 30 October–3 November 2017; Association for Computing Machinery: New York, NY, USA, 2017; pp. 1175–1191. [Google Scholar] [CrossRef] [Green Version]
- Geyer, R.C.; Klein, T.; Nabi, M. Differentially Private Federated Learning: A Client Level Perspective. arXiv 2018, arXiv:1712.07557. [Google Scholar]
- Chen, F.; Luo, M.; Dong, Z.; Li, Z.; He, X. Federated Meta-Learning with Fast Convergence and Efficient Communication. arXiv 2019, arXiv:1802.07876. [Google Scholar]
- Thapa, C.; Arachchige, P.C.M.; Camtepe, S.; Sun, L. Splitfed: When federated learning meets split learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 24–26 February 2022; Volume 36, pp. 8485–8493. [Google Scholar]
- Mahmood, Z.; Jusas, V. Blockchain-Enabled: Multi-Layered Security Federated Learning Platform for Preserving Data Privacy. Electronics 2022, 11, 1624. [Google Scholar] [CrossRef]
- Cox, B.; Chen, L.Y.; Decouchant, J. Aergia: Leveraging Heterogeneity in Federated Learning Systems. arXiv 2022, arXiv:2210.06154. [Google Scholar]
- Liu, C.; Chua, T.J.; Zhao, J. Time Minimization in Hierarchical Federated Learning. arXiv 2022, arXiv:2210.04689. [Google Scholar]
- Peng, X.; Huang, Z.; Zhu, Y.; Saenko, K. Federated Adversarial Domain Adaptation. arXiv 2019, arXiv:1911.02054. [Google Scholar]
- Ben-David, S.; Blitzer, J.; Crammer, K.; Pereira, F. Analysis of Representations for Domain Adaptation. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–9 December 2006; Schölkopf, B., Platt, J., Hoffman, T., Eds.; MIT Press: Cambridge, CA, USA, 2006; Volume 19. [Google Scholar]
- Ben-David, S.; Blitzer, J.; Crammer, K.; Kulesza, A.; Pereira, F.; Vaughan, J.W. A Theory of Learning from Different Domains. Mach. Learn. 2010, 79, 151–175. [Google Scholar] [CrossRef]
- Nascimento, M.A.; Özsu, M.T.; Kossmann, D.; Miller, R.J.; Blakeley, J.A.; Schiefer, K.B. (Eds.) Proceedings of the Thirtieth International Conference on Very Large Data Bases, VLDB 2004, Toronto, ON, Canada, 31 August–3 September 2004; Morgan Kaufmann: Burlington, MA, USA, 2004. [Google Scholar]
- Zhu, L.; Han, S. Deep leakage from gradients. In Federated Learning; Springer: Berlin/Heidelberg, Germany, 2020; pp. 17–31. [Google Scholar]
- Hitaj, B.; Ateniese, G.; Perez-Cruz, F. Deep models under the GAN: Information leakage from collaborative deep learning. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, Dallas, TX, USA, 30 October–3 November 2017; pp. 603–618. [Google Scholar]
- Melis, L.; Song, C.; De Cristofaro, E.; Shmatikov, V. Exploiting unintended feature leakage in collaborative learning. In Proceedings of the 2019 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 19–23 May 2019; pp. 691–706. [Google Scholar]
- Li, Q.; Wen, Z.; He, B. Practical federated gradient boosting decision trees. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 4642–4649. [Google Scholar]
| Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).