Next Article in Journal
Design and Experiment of High-Resolution Multispectral Polarization Imaging System
Next Article in Special Issue
Social Recommendation Based on Quantified Trust and User’s Primary Preference Space
Previous Article in Journal
Efficient Realization for Third-Order Volterra Filter Based on Singular Value Decomposition
Previous Article in Special Issue
Application of Extension Engineering in Safety Evaluation of Chemical Enterprises
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Federated Transfer Learning Framework Based on Heterogeneous Domain Adaptation for Students’ Grades Classification

School of Computer Science and Engineering, Northeastern University, Shenyang 110169, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(21), 10711; https://doi.org/10.3390/app122110711
Submission received: 10 September 2022 / Revised: 18 October 2022 / Accepted: 20 October 2022 / Published: 22 October 2022

Abstract

:
In the field of educational data mining, the classification of students’ grades is a subject that receives widespread attention. However, solving this problem based on machine learning algorithms and deep learning algorithms is usually limited by large datasets. The privacy problem of educational data platforms also limits the possibility of building an extensive dataset of students’ information and behavior by gathering small datasets and then carrying out the federated training model. Therefore, the balance of educational data and the inconsistency of feature distribution are the critical problems that need to be solved urgently in educational data mining. Federated learning technology enables multiple participants to continue machine learning and deep learning in protecting data privacy and meeting legal compliance requirements to solve the data island problem. However, these methods are only applicable to the data environment with common characteristics or common samples under the alliance. This results in domain transfer between nodes. Therefore, in this paper, we propose a framework based on federated transfer learning for student classification with privacy protection. This framework introduces the domain adaptation method and extends the domain adaptation to the constraint of federated learning. Through the feature extractor, this method matches the feature distribution of each party in the feature space. Then, labels and domains are classified on each side, the model is trained, and the target model is updated by gradient aggregation. The federated learning framework based on this method can effectively solve the federated transfer learning on heterogeneous datasets. We evaluated the performance of the proposed framework for student classification on the datasets of two courses. We simulated four scenarios according to different situations in reality. Then, the results of only source domain training, only target domain training, and federated migration training are compared. The experimental results show that the heterogeneous federated transfer framework based on domain adaptation can solve federated learning and knowledge transfer problems when there are little data at the data source and can be used for students’ grades classification in small datasets.

1. Introduction

With the development of artificial intelligence, especially the emergence of neural networks, artificial intelligence technology driven by big data has played its role in many complex fields, such as automatic driving, medical care, and finance. People want this technology to be realized in all walks of life. However, the actual situation is very disappointing. Except for a few limited fields, many industries have limited data and poor data quality, which is insufficient to support the implementation of artificial intelligence technology [1]. Even if there are many data in more application fields, these data are distributed among different data sources and marked unevenly. For example, there are 537,100 schools of all levels and types in China, with 289 million students. These students’ data are stored in different schools and on various online education platforms. These data sources are not connected to each other, and the data exist in the form of isolated islands. It is almost impossible to integrate these data, or the cost required is enormous.
On the other hand, with the further development of big data, data privacy and security have become a worldwide trend. Because of IoT devices’ increasing storage and computing capabilities, it is common to save data and models on local devices [2]. Federated learning is a new frontier field [1,3,4]. Google has launched a federated learning system [3,4,5], in which a federated of distributed participants updates the global machine learning model and their data are stored locally. Their framework requires all contributors to share the same feature space. On the other hand, secure machine learning for partitioning data in feature space is also studied [6,7,8], which finds that, in machine learning, each participant can make joint modeling with the help of other parties’ data. However, all parties do not need to share data resources. They can conduct federated data training and establish a shared machine learning model when the data do not come out locally. Federated learning enables multiple participants to continue machine learning and deep learning in protecting data privacy and meeting legal compliance requirements to solve the data island problem. Federated learning technology can be used among mobile terminals, wearable devices, and different platforms to realize a training machine learning model that pays more attention to privacy and efficiency. However, when one party collects a lot of unlabeled data, even when the features are statistically different from those collected by the other party, domain shift will arise [9]. This makes the model obtained by federated learning and training unable to be extended to new equipment. For example, in the field of education, two different online education platforms save local student data. One platform may collect more student behavior logs, while the other platform saves students’ personal information.
In the field of educational data mining, the classification of students’ grades is usually limited by the size of datasets and the number of data features. Moreover, because of data privacy protection, each data platform cannot carry out practical federated training. Therefore, it is an urgent problem to solve the imbalance of data and the inconsistency of characteristics among data owners to train student models.
This paper proposes a framework based on federated transfer learning for students’ grades classification with privacy protection. In this paper, heterogeneous domain adaptive technology is introduced into federated transfer learning, and knowledge transfer between different nodes is solved by exchanging encryption parameters between data. In this framework, each data owner uses the network of label classifier and domain classifier for training and aligns the features of all parties in the new feature space for transfer learning. Furthermore, according to the actual situation, we also simulate the following four scenarios by constraining different characteristics and data scales:
  • Scenario 1: Both Party A and Party B have the same feature distribution and data scales;
  • Scenario 2: Party A and Party B have the same characteristics but different data scales;
  • Scenario 3: Party A and Party B have different characteristics but the same data scales;
  • Scenario 4: Both Party A and Party B have different feature distribution and data scales.
We conducted experiments on datasets from two middle schools in Portugal and obtained the training results of student models in these four situations. Experimental results show that both parties with different data feature spaces and inconsistent data scales can effectively transfer knowledge under this framework. It can help the target data party with missing data labels or fewer feature data to train and improve the generalization ability of the model of main data participants.
The rest of this paper is organized as follows: In Section 2, we outline related work. In Section 3, we present the details of the model proposed in this paper. Then, in Section 4, we describe the experimental details. The analysis of the experimental results is in Section 5. Finally, in Section 6, we summarize our work and present some ideas for future work.

2. Related Work

2.1. Domain Adaptation

Domain adaptation is an active research topic in many fields of artificial intelligence, including machine learning, natural language processing, and computer vision. The methods of early domain adaptation mainly focus on constructing domain-invariant features. This can be achieved by the weight and selection mechanism of features  [10,11] or by learning the explicit feature transformation that aligns the distribution of source domain and target domain [12,13,14]. In recent work, deep neural networks have been used to learn powerful representation and perform unsupervised domain adaptation [15]. Revgrad [16] uses a domain classification network, which aims to distinguish between source and target embedding. The goal of the feature extraction network is to produce embedding that maximizes the loss of domain classifiers and minimizes the loss of tag prediction. This is achieved by negating the gradient from the domain classification network.
Driven by GAN [17], many transfer learning methods are based on the assumption that a good feature representation contains almost no specific information about the original domain of an instance. For example, Ganin et al. proposed a depth structure called Domain-Adversarial Neural Network (DANN) for domain adaptation [16,18]. Its architecture consists of a feature extractor, tag predictor, and domain classifier. DANN can be trained by inserting a particular gradient reversal layer (GRL). After the training of the whole system, the feature extractor learns the deep features of instances, and the output classifies the prediction tags of unmarked target domain instances.
In the field of domain adaptation, some researchers have improved and applied domain adaptation technology. Cao et al. put forward an adaptive model with semantic consistency (ADASC), which gradually and effectively aligns the distinguishing features across domains using the class hierarchy relationship between domains [19]. Hu et al. developed a framework DATSING based on transfer learning, which effectively utilized the potential representation of cross-domain time series to enhance target domain prediction [20]. Su et al. introduced the task of continuous domain adaptation in machine reading comprehension [21].
There are some other related impressive works. Zeng et al.’ s work puts forward a unified framework for adaptation in adversarial fields [15]. Shen et al. used Wasserstein distance for domain adaptation [22]. Hoffman et al. adopted cycle-consistency loss to ensure the consistency of structure and semantics [23]. Long et al. put forward the condition domain adversarial network, which uses the condition domain discriminator to assist the adversarial adaptation [24]. Zhang et al. adopted the symmetrical design of source classifier and target classifier [25]. Zhao et al. used a domain confrontation network to solve the problem of multi-source transfer learning [26]. Yu et al. proposed a dynamic adversarial adaptation network [27]. Chen et al. proposed a reinforcement learning (DARL) framework combined with adversarial learning for partial domain adaptation [28]. The DARL framework adopts the deep Q-learning method to learn the strategy by approximating the action value function, so that the agent can make a choice decision. In order to minimize the domain differences in cross domain knowledge transfer, Chen et al. used self similarity consistency (SSC) constraints to match the deep feature structures in the source and target domains, rather than aligning the global distribution statistics across domains [29]. Tran et al. proposed a generative adversary domain adaptive (GADA) algorithm, which significantly improved the process of distinguishing feature extraction by injecting additional classes and using the generated samples for training [30]. Guo et al. proposed a new distribution calibration method by learning an adaptive weight matrix between new samples and base classes, which built on the Hierarchical Optimal Transport (H-OT) framework by minimizing the high-level OT distance between the new sample and the base class to convey the adaptive weight information of the base class [31]. In the multi-source domain adaptation task, Deng et al. proposed a two-layer optimization-based robust object training (BORT2) [32]. BORT2 first learns a labeling function using source and target data, and then trains a noise-robust model only on the pseudo-labeled target domain. Noise-robust models exploit feature uncertainty to detect label noise and mitigate its negative effects.
To sum up, although there are many kinds of research in the field of domain adaptation in recent years, this research is all based on typical scene training. Their data are stored on the same server, which leads to incompatibility under the condition of federated learning. Therefore, our research hopes to achieve domain adaptation under the framework of federated learning.

2.2. Federated Learning

Google recently proposed the concept of federated learning [3,4,5]. Their main idea is to build a machine learning model based on datasets distributed on multiple devices while preventing data leakage. Recent improvements focus on overcoming statistical challenges [2,33] and improving the security of federated learning [34,35], and there are some research efforts to make federated learning more personalized [36]. All of the above work focuses on device federated learning, involving distributed mobile user interaction and communication cost, unbalanced data distribution, and device reliability in large-scale distribution are some of the main factors of optimization. To extend the concept of federated learning to cover collaborative learning scenarios among organizations, Yang et al. expanded the original federated learning into the general idea of all decentralized collaborative machine learning technologies for privacy protection [1]. They made a comprehensive overview of federated learning and federated transfer learning technologies. At the same time, they further investigated the related security foundation and discussed the relationship with other associated fields, such as multi-agent theory and privacy protection data mining. Cheng et al. proposed a new lossless privacy protection algorithm secureboost, which is used to train high-quality tree enhancement models when the training data are kept secret on multiple parties [7]. Chandra et al. proposed a new distributed machine learning method, called splitfed. Due to network fragmentation, it provides a higher level of privacy than federated learning. It is suitable for machine learning (ML) with low computational resources, fast training, and analysis of private and sensitive data [37]. Mahmood et al. studied the feasibility of data and model poisoning attacks under the blockchain based FL system built next to the Ethereum network and the traditional FL system (no blockchain) [38]. They proposed a transparent incentive mechanism to fill the knowledge gap, which can encourage good behavior among participating decentralized nodes, avoid common problems, and provide knowledge for FL security literature by studying the current FL system.
In addition, the time required for computing is also an important factor. Federated learning (FL) is largely affected by the heterogeneity of clients. The training speed of clients is inconsistent. In order to reduce the impact of slow clients, Cox et al. proposed a new FL algorithm using model freezing and unloading, which uses a simple, fast, and scalable scheduling algorithm to periodically organize model unloading between clients [39]. To reduce the communication and computation delays of model parameters, Liu et al. proposed a new iterative algorithm and minimized the total model parameter communication and computation delays by optimizing the number of local iterations and edge iterations [40].
Recent research has focused on transferring knowledge from decentralized nodes to new nodes with different data domains. Liu et al. introduced new technology and framework called federated transfer learning (FTL) to improve the statistical model under the data alliance. The alliance allows knowledge sharing without compromising users’ privacy and allows supplementary knowledge to be transmitted in the network [8]. Therefore, the target domain can build a more flexible and robust model using rich tags from the source domain. Peng et al. put forward a principled method to solve the federated domain adaptation problem, which aims to make the representations learned between different nodes consistent with the data distribution of target nodes [41].
However, we do not think Peng’s method extends to the constraint of federated learning. It cannot be applied to the general environment of federated learning. Therefore, in this paper, we combine domain adaptation with secure federated learning and propose a new federated transfer learning framework based on domain adaptation to solve knowledge transfer in federated transfer learning.

3. Method

In this section, we introduce the details of our proposed method. The network we use is a deep neural network. The network structure we propose is a deep neural network with three layers whose structure is shown in Figure 1. The input data are passed through a feature extractor to match the feature distribution of each party in the feature space. Then, the labels and domains are classified on each side, the model is trained, and the target model is updated by gradient aggregation.

3.1. Problem Definition

In the actual educational data mining, we often encounter the situation of small data scale and lack of data features and need to manually mark them. However, due to the data privacy protection, in many cases, the data owners can not carry out collaborative training. To solve this problem, we simulated a federated training scenario between Party A and Party B.
Consider two data owners, a source domain D A = { ( x i A , y i A , z A ) i = 1 n } where x i A R n , y i A { 0 , 1 } is the ith label and z A = 1 is the domain label, and a target domain D B = { ( x i B , z B ) i = 1 n } where x i B R n , z B = 0 is the domain label. n , n represents the number of samples of the source domain and target domain, respectively. The total number of samples is N = n + n .
In our definition, Party A and Party B represent two platforms with educational data. Among them, Party A has complete data, label information, and class labels. Party A is the leading party in federated learning and the owner of the source domain data in federated transfer learning. Party B only has some data and class labels and has no data labels. Play the role of the customer in federal migration learning. Party A and Party B use the knowledge gained from federal migration learning to build a model and predict their users.
Our goal is to establish a transfer learning model, accurately predict the category of students of Party B, and improve the classification effect of Party A. At the same time, their data will not be exposed. In this paper, domain adaptation is introduced into federated migration learning, which minimizes and maximizes the classification errors of the source domain, and then exchanges knowledge between the source domain and the target domain. Then, this federated learning problem can be formulated as follows:
minimize w R N L 1 w , D = 1 N i = 1 N f w ; x i , y i
maximize w R N L 2 w , D = i = 1 N γ w ; x i , z i
where w is the parameter of the model, f represents the loss function of the label classifier, and γ represents the loss function of the domain classifier.
Our goal is to exchange only intermediate parameters to achieve the following goals on privacy protection:
minimize w R N L = L 1 w , D L 2 w , D

3.2. Approach Statement

We assume that there are two data owners P a r t y A & B training model A and model B. Firstly, the model needs feature extraction to generate two implicit representations u i A = A x i A and u i B = B x i B , where u A R N A × d and u B R N B × d , and d represent the dimensions of the hidden layer. φ ( x ; w , b ) = 1 1 + exp ( w T x ) is the feature transformation function, where matrix-vector pair ( w , b ) R N × R . The neural network φ A and φ B transform their features, and project the features of P a r t y A and P a r t y B into a common feature subspace, where they can exchange knowledge.
Then, the source domain data of feature space is classified by the prediction layer, and the correct labels are separated as far as possible. The prediction function ψ ( u i B ) = ψ ( u 1 A , y 1 A . . . u N A A , y N A A , u j B ) ,
where ψ ( a ) = exp a i j = 1 | a | exp a j i = 1 | a | . Thus, we obtain the first optimization goal:
argmin Θ x , Θ y L 1 = 1 n i = 1 T 1 y i A , ψ u i B
where Θ x , Θ y are training parameters of layers φ A , ψ . 1 is classification loss, using the negative log-probability of the correct labels.
The network also needs a regularization term, that is, a metric criterion so that the source domain and the target domain in the feature space are as close as possible. We use H-divergence to measure [42,43,44]:
d ^ H ( S , T ) = 2 1 min η H 1 n i = 1 n I η x i = 0 + 1 n i = n + 1 N I η x i = 1
where H is a hypothetical class, binary classifiers η : X { 0 , 1 } .
According to formula (5), the empirical H-divergence between the source domain and the target domain can be calculated by the following formula:
d ^ H D A , D B = 2 1 min η H 1 n i = 1 n I η ω u i B = 0 + 1 n i = n + 1 N I η ω u i B = 1
It is usually difficult to calculate d ^ H D A , D B . According to the inspiration of Ben-David’s research [42,43], we can define the approximate solution of the classification problem of the source domain and target domain, that is, domain classifier. It classifies data in feature space and tries to distinguish which domain the data comes from. Domain classifier ω ( u j B ) = 1 1 + exp ( u j B ) , and we can make the feature distribution of the data in the source domain and the target domain as close as possible by approximately calculating the empirical H-divergence between the two samples. We can obtain the second optimization target:
argmax Θ y , Θ z L 2 = 1 n i = 1 n 2 z i A , ω u i B = 0 1 n i = n + 1 N 2 z i A , ω u i B = 1
where Θ y , Θ z are training parameters of layers ψ , ω . z i is the domain classification label. 2 is the loss of domain classification layer, and the loss of feature alignment is calculated by cross entropy.
The final objective function is:
argmin Θ x , Θ y , Θ z L = 1 n i = 1 n 1 y i A , ψ u i B + λ 1 n i = 1 n 2 z i A , ω u i B = 0 + 1 n i = n + 1 N 2 z i A , ω u i B = 1
where λ is the weight parameters. For i { A , B } , the stochastic gradients can be computed as the following:
g x i = L 1 i Θ i x λ L 2 i Θ i x
g y i = L 1 i Θ i y
g z i = L 2 i Θ i z

3.3. Federated Domain Adaptation

In order to realize secure computation without exchanging data, additively homomorphic encryption is adopted. We define a number x, which becomes x after additively homomorphic encryption. According to the properties of homomorphic encryption, for any two numbers x and y, x + y = x + y and x   × y = x × y is valid. However, the loss function and gradient can not be directly calculated by additively homomorphic encryption. In order to solve this problem, we use the second-order Taylor approximation for the logistic loss
( y , φ ) ( y , 0 ) 1 2 y φ + 1 8 y 2 φ 2
and the gradient is
φ = 1 2 y + 1 4 y 2 φ
In this way, we obtain the encrypted loss and gradient.
L 1 = 1 n i = 1 n 1 ( y i A , 0 ) 1 2 y i A ψ u i B + 1 8 y i A 2 ψ u i B ψ u i B
L 2 = 1 n i = 1 n 2 ( z i A , 0 ) 1 2 z i A ω u i B + 1 8 z i A 2 ω u i B ω u i B 1 n i = n + 1 n 2 ( z i A , 0 ) 1 2 z i A ω u i B + 1 8 z i A 2 ω u i B ω u i B
L 1 i Θ i = 1 n i = 1 n ψ u i B ψ u i B u i B 1 8 y i A 2 ψ u i B ψ u i B u i B Θ i + 1 n i = 1 n 1 2 y i A ψ u i B ψ u i B u i B u i B Θ i
L 2 i Θ i = 1 n i = 1 n ω u i B ω u i B u i B 1 8 z i A 2 ω u i B ω u i B u i B Θ i 1 n i = 1 n 1 2 z i A ω u i B ω u i B u i B u i B Θ i 1 n i = 1 n ω u i B ω u i B u i B 1 8 z i A 2 ω u i B ω u i B u i B Θ i 1 n i = 1 n 1 2 z i A ω u i B ω u i B u i B u i B Θ i
Then, update the parameters according to the gradient direction
Θ x Θ x η g x
Θ y Θ y η g y
Θ z Θ z η λ g z
where η is the learning rate. With these parameters, we now design a federated Domain Adaptation network for solving the Domain Adaptation problem.As shown in Figure 2. First, Party A and Party B send public keys to each other, then initialize parameters Θ x , Θ y , Θ z , calculate implicit representations u A and u B , and then calculate, encrypt, and exchange intermediate results g and L .
To prevent exposure of the gradients of A and B, A and B further mask each gradient with encrypted random values. Then, they send the gradient and loss of the mask to each other and locally decrypt these values. Once the loss convergence condition is met, Party A can send a termination signal to Party B. Otherwise, they will unmask the gradients, update the model parameters with their gradients, and then continue the next iteration. Algorithm 1 summarizes the process.
Algorithm 1 Federated domain adaptation
  • Input: learning rate: η , adaptation parameter: λ
  • Output: Model parameters Θ x , Θ y , Θ z  Initialization:Patty A&B initialize Θ x , Θ y , Θ z
  • while Convergence conditions is not met do
  •   for all each iteration k = 1 , 2 , 3 . . . do
  •    Part A&B compute: u i A , u i B
  •    A&B create a random mask: m A , m B
  •    A compute and encrypt: L A , g x A + m A , g y A + m A , g z A + m A and send to B
  •    B compute and encrypt: g x B + m B , g y B + m B , g z B + m B and send to A
  •    A&B decrypt gradients and exchange, unmask and update model locally
  •      Θ x Θ x η g x
  •      Θ y Θ y η g y
  •      Θ z Θ z η λ g z
  •   end for
  • end while
  • return Θ x , Θ y , Θ z

3.4. Security Analysis

In this section, we aim to determine whether one party participating in federal training can learn the other party’s data from the parameter set H k exchanged during training. Previous studies mainly introduced the leakage of the whole set of model parameters or gradient in the training process [45,46,47]. However, in our framework, the model’s parameters are confidential, and only the intermediate results in the tag prediction layer may reveal information.
Theorem 1.
According to our definition, the federated transfer learning framework based on domain heterogeneous adaptation is secure, provided that the underlying encryption parameter sharing scheme is secure.
Proof. 
In our framework, the model’s parameters are confidential, and only the intermediate results in the label classifier layer may reveal information. In the domain classifier layer, each party can only receive the information of the domain but not the information of the data. Let S j be the set of data points sampled in the t-th iteration and i j be the i-th sample in the j-th iteration. H i j refers to the contribution of the i-th sample to other parties. In the j + 1 -th iteration, we can obtain the weight update of the label classifier according to Equation (19):
Θ j + 1 y = Θ j y η j 1 N S j i j S j 1 2 y i j + 1 4 y i j 2 H i j
where H i j = φ ( x i j ) Θ j y .For any participant k with unpublished datasets D k and training parameters Θ k y following our defined framework, there is an infinite solution of { x i j k } i S j , j = 0 , 1 that produces the same contribution set { H i j k } j = 0 , 1 . That is to say, no matter the number of iterations, the data of participant k cannot be uniquely determined from its exchanged information { H i j k } j = 0 , 1 . This definition of security is consistent with the previous definition of security proposed in privacy protection and secure multiple computation of machine learning [48]. Under this definition, when some prior knowledge about data are known, the other party may exclude some alternative solutions but cannot infer accurate original data. This model provides a flexible trade-off between privacy and efficiency. □
Theorem 2.
According to our definition, the federated transfer learning framework based on domain heterogeneous adaptation is secure, provided that the homomorphic encryption scheme used is secure.
Proof. 
In our scheme, all the learning and training data are encrypted by a homomorphic encryption mechanism. The parameters and intermediate data of the model are encrypted in the whole federated learning process. In each iteration, both sides will create a new random encryption mask m. This randomness and confidentiality also ensure the security of the whole training and learning process. Therefore, as long as the encryption scheme is secure, the federated learning framework we defined is also secure. □

4. Experiments

In this section, we mainly introduce the usage of our proposed framework and verify it by constraining different features, quantity, and the correlation between features on the datasets of two courses.

4.1. Datasets and Experimental Settings

The data used in the experiment come from the students’ achievements in secondary education in two Portuguese schools. Data attributes include student achievement, demographic, social, and school-related characteristics collected using school reports and questionnaires. The two datasets provide the performance of two different subjects: mathematics and Portuguese. There are 396 students in the mathematics dataset and 650 students in the Portuguese dataset. We selected 16 attributes, which are described in Table 1. Students are classified in a binary system, and the G3 grades provide the classification labels of students. Students with scores greater than ten are classified into one category, while others are classified.
In our experiments, we set the depth of DNN layers as three in the model architecture. The learning rates are set to 0.0001, and batch size is set to 128. To avoid overfitting, we apply dropout on the DNN layers with a drop rate of 0.3.

4.2. Privacy Protection

In order to ensure the privacy of the framework, the experiment simulates the distributed training scenario of multiple nodes, and the data of Party A and Party B are stored separately in each node to participate in the federated learning. In order to promote secure computing between the two parties, a semi-honest third-party server is introduced, and it is assumed that it does not collude with other participants. After learning and training, each participant can only hold model parameters associated with their characteristics. Therefore, when reasoning, both parties need to work together to generate the output. The third-party server does not obtain the data of all parties but only distributes the secret key and exchanges the encryption parameters.

4.3. Design

In the experiment, according to the datasets size and data feature distribution of Party A and Party B, four scenarios were established to simulate the actual situation. They are as follows:
  • Scenario 1: Both Party A and Party B have the same feature distribution and data scales;
  • Scenario 2: Party A and Party B have the same characteristics but different data scales;
  • Scenario 3: Party A and Party B have different characteristics but the same data scales;
  • Scenario 4: Both Party A and Party B have different feature distribution and data scales.
To correspond to these four scenarios, we processed the data and selected some data and attributes. In addition, in the scene with different feature distribution, the correlation analysis of attributes is carried out to ensure that the correlation with domain invariant features does not exceed the threshold. We carry out a factor analysis on the features contained in the datasets. The calculated factor scores are shown in Table 2, indicating eight principal components in total.
Given the scenarios with consistent feature distribution, we selected eight variables that contributed the most to the principal components as the input features of the model, namely walc, fedu, paid, address, romantic, famrel, famsize, and activities. The source domain still adopts the eight features selected from previous scenarios for scenarios with inconsistent features, while the target domain adopts the new eight features. They are dalc, absences, medu, goout, higher, freetime, studytime and internet. We visualize these two groups of features, use PCA to reduce the dimension to three-dimensional space, and check the distribution of data points. Figure 3 respectively shows the distribution of the features of the source domain and the target domain in the three-dimensional space. To ensure the domain invariant features in the transfer process, we also conducted a pairwise correlation analysis on a total of 16 features and tested the significance of the correlation results. The results are shown in Table 3.
It can be seen from Figure 3 that the feature spaces of the source domain and the target domain are inconsistent, so the two datasets are heterogeneous in features, which guarantees our hypothesis. At the same time, to ensure the transfer of valuable knowledge in the transfer process, the correlation between features is analyzed. From Table 3, there is a specific correlation between the features. The first row of each feature in Table 3 represents the Pearson correlation coefficient between variables, and the second row represents the p-value of the significance hypothesis. From the significance test results, there are significantly associated features between the two domains, which ensures the extraction of domain invariant features in the process of domain transfer and the effectiveness of transfer learning.
For the scenario with the same data scale, both the source domain and the target domain used the data of 200 students. In the case of an inconsistent data scale, the source domain uses the data of 200 students, and the target domain uses the data of 80 students.
From the results, we can see that both the source domain and the target domain can make classification predictions when they are not trained. The accuracy is 0.5, which accords with the common sense of binary classification. The accuracy of the domain classifier in each scene is between 0.5 and 0.6, which indicates that the machine can not accurately distinguish the features of the source domain or the target domain, displaying that knowledge has transferred.

5. Results and Discussion

In the experiment, we use the logistic regression model to classify the students’ academic performance in binary, the network structure is shown in Figure 1, and we use the federated transfer learning framework based on heterogeneous domain adaptation defined by this paper.
In each scenario, we conducted three pieces of training, namely, training only for the source domain, training only for the target domain, and transfer training. We give the classification accuracy of the source domain and target domain under each training mode and give the classification accuracy of the domain under the transfer training mode. The results are shown in Table 4.
We can find that, in Scenarios 2–4, the results of target-only training are generally a little worse than source-only training due to the difference in the data size of the source and target domains and the different contributions of the included features to the principal components. Moreover, in Scenario 1, after using domain adaptation training, the accuracy results are not much different from the results of individual training in each domain, which indicates that using domain adaptation training will not improve the model effect in this case. In Scenario 2, after using domain adaptation, the classification accuracy of the target domain is greatly enhanced. In Scenario 3, both the source domain’s classification accuracy and the target domain have dramatically increased. However, in Scenario 4, the classification accuracy of the source domain and the target domain increased less than that of Scenario 3.
The result analysis shows that the domain adaptation model has little effect when the feature distribution and scale of the source domain and the target domain are consistent. However, when one party’s data scale is small, it can benefit from federated learning. Especially when the feature distribution of both parties is inconsistent, both parties can learn new knowledge from each other and improve the classification accuracy. In Scenario 4, although we project the source and target domains into a common feature subspace, since the source and target domains have different feature distributions and data scales, the training knowledge on the source domain is directly applied to the target domain, and the resulting effect will be inferior to Scenario 2 and Scenario 3.
From the experimental results, the federated transfer learning framework designed by us can effectively solve the problem of federated learning with small unilateral data and missing unilateral features in students’ grades classification, and all participants can obtain new knowledge from each other while maintaining privacy. Experiments show that applying a federated learning framework can solve the problem of data islands in educational data mining, which can be used for reference in the field of intelligent education.
The study has potential limitations. Since the experimental part of this paper only verifies the data exchange between two different education platforms, and the data differences generated by different education platforms may be large, how to transfer privacy protection knowledge between more online education platforms still needs further exploration.

6. Conclusions

This paper proposes a framework based on federated transfer learning for students’ grades classification with privacy protection. By introducing feature extractors and domain classifiers, the framework matches the feature distribution of all parties in collaborative learning in some feature spaces. It uses homomorphic encryption and random masking to ensure data privacy during transmission. The experimental results show that both parties with different data feature spaces and inconsistent data scales can effectively transfer knowledge under this framework. It can help the target data parties with missing data labels, or fewer feature data train the students’ grades classification model and improve the generalization ability and classification effect of the students’ grades classification model of the primary data participants.
In future work, we will further study privacy protection in the framework. This includes studying privacy protection in the federated learning framework in the presence of malicious participants, and more efficient feature extraction methods. We are committed to expanding federated learning technology to more fields and exploring more effective federated learning methods. These methods can include studying and further improving this method’s communication efficiency and effectiveness in more complex and asynchronous federated learning systems.

Author Contributions

Funding acquisition, B.X.; methodology, S.Y. and B.X.; conceptualization, S.Y.; software, S.Y.; validation, S.Y. and B.X.; writing, review and editing, S.Y., S.L. and Y.D.; supervision, S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant (U1811261), and the Fundamental Research Funds for the Central Universities of China, grant (N2116019), and the Liaoning Natural Science Foundation, grant (2022-MS-119).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yang, Q.; Liu, Y.; Chen, T.; Tong, Y. Federated Machine Learning: Concept and Applications. ACM Trans. Intell. Syst. Technol. 2019, 10, 1–19. [Google Scholar] [CrossRef]
  2. Smith, V.; Chiang, C.K.; Sanjabi, M.; Talwalkar, A. Federated Multi-Task Learning. arXiv 2018, arXiv:1705.10467. [Google Scholar]
  3. Konečný, J.; McMahan, H.B.; Yu, F.X.; Richtárik, P.; Suresh, A.T.; Bacon, D. Federated Learning: Strategies for Improving Communication Efficiency. arXiv 2017, arXiv:1610.05492. [Google Scholar]
  4. McMahan, H.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Proceedings of the AISTATS, Fort Lauderdale, FL, USA, 20–22 April 2017. [Google Scholar]
  5. Konečný, J.; McMahan, H.B.; Ramage, D.; Richtárik, P. Federated Optimization: Distributed Machine Learning for On-Device Intelligence. arXiv 2016, arXiv:1610.02527. [Google Scholar]
  6. Hardy, S.; Henecka, W.; Ivey-Law, H.; Nock, R.; Patrini, G.; Smith, G.; Thorne, B. Private federated learning on vertically partitioned data via entity resolution and additively homomorphic encryption. arXiv 2017, arXiv:1711.10677. [Google Scholar]
  7. Cheng, K.; Fan, T.; Jin, Y.; Liu, Y.; Chen, T.; Papadopoulos, D.; Yang, Q. Secureboost: A lossless federated learning framework. IEEE Intell. Syst. 2021, 36, 87–98. [Google Scholar] [CrossRef]
  8. Liu, Y.; Kang, Y.; Xing, C.; Chen, T.; Yang, Q. A Secure Federated Transfer Learning Framework. IEEE Intell. Syst. 2020, 35, 70–82. [Google Scholar] [CrossRef]
  9. Quiñonero-Candela, J.; Sugiyama, M.; Lawrence, N.D.; Schwaighofer, A. Dataset Shift in Machine Learning; MIT Press: Cambridge, CA, USA, 2009. [Google Scholar]
  10. Huang, J.; Smola, A.J.; Gretton, A.; Borgwardt, K.M.; Scholkopf, B. Correcting Sample Selection Bias by Unlabeled Data. In Proceedings of the 19th International Conference on Neural Information Processing Systems, NIPS’06, Vancouver, BC, Canada, 4–7 December 2006; MIT Press: Cambridge, MA, USA, 2006; pp. 601–608. [Google Scholar]
  11. Daumé, H., III. Frustratingly Easy Domain Adaptation. In Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics, Prague, Czech Republic, 23–30 June 2007; Association for Computational Linguistics: Prague, Czech Republic, 2007; pp. 256–263. [Google Scholar]
  12. Gong, B.; Shi, Y.; Sha, F.; Grauman, K. Geodesic flow kernel for unsupervised domain adaptation. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2066–2073. [Google Scholar] [CrossRef] [Green Version]
  13. Gopalan, R.; Li, R.; Chellappa, R. Domain adaptation for object recognition: An unsupervised approach. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 999–1006. [Google Scholar] [CrossRef]
  14. Peng, X.; Saenko, K. Synthetic to Real Adaptation with Generative Correlation Alignment Networks. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 4–8 January 2022; pp. 1982–1991. [Google Scholar] [CrossRef] [Green Version]
  15. Tzeng, E.; Hoffman, J.; Saenko, K.; Darrell, T. Adversarial discriminative domain adaptation. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7167–7176. [Google Scholar]
  16. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M.; Lempitsky, V. Domain-Adversarial Training of Neural Networks. J. Mach. Learn. Res. 2016, 17, 1–35. [Google Scholar]
  17. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. In Proceedings of the 27th International Conference on Neural Information Processing Systems—Volume 2, NIPS’14, Montreal, QC, Canada, 8–13 December 2014; MIT Press: Cambridge, MA, USA, 2014; pp. 2672–2680. [Google Scholar]
  18. Ganin, Y.; Lempitsky, V. Unsupervised domain adaptation by backpropagation. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; PMLR: Westminster, UK, 2015; pp. 1180–1189. [Google Scholar]
  19. Cao, M.; Zhou, X.; Xu, Y.; Pang, Y.; Yao, B. Adversarial Domain Adaptation with Semantic Consistency for Cross-Domain Image Classification. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, CIKM ’19, Beijing, China, 3–7 November 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 259–268. [Google Scholar] [CrossRef] [Green Version]
  20. Hu, H.; Tang, M.; Bai, C. DATSING: Data Augmented Time Series Forecasting with Adversarial Domain Adaptation. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, CIKM ’20, Virtual, 19–23 October 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 2061–2064. [Google Scholar] [CrossRef]
  21. Su, L.; Guo, J.; Zhang, R.; Fan, Y.; Lan, Y.; Cheng, X. Continual Domain Adaptation for Machine Reading Comprehension. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, CIKM ’20, Virtual, 19–23 October 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 1395–1404. [Google Scholar] [CrossRef]
  22. Shen, J.; Qu, Y.; Zhang, W.; Yu, Y. Wasserstein distance guided representation learning for domain adaptation. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  23. Hoffman, J.; Tzeng, E.; Park, T.; Zhu, J.Y.; Isola, P.; Saenko, K.; Efros, A.; Darrell, T. Cycada: Cycle-consistent adversarial domain adaptation. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; PMLR: Westminster, UK, 2018; pp. 1989–1998. [Google Scholar]
  24. Long, M.; Cao, Z.; Wang, J.; Jordan, M.I. Conditional Adversarial Domain Adaptation. In Proceedings of the Advances in Neural Information Processing Systems; Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2018; Volume 31. [Google Scholar]
  25. Zhang, Y.; Tang, H.; Jia, K.; Tan, M. Domain-symmetric networks for adversarial domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5031–5040. [Google Scholar]
  26. Zhao, H.; Zhang, S.; Wu, G.; Moura, J.M.; Costeira, J.P.; Gordon, G.J. Adversarial multiple source domain adaptation. Adv. Neural Inf. Process. Syst. 2018, 31, 8559–8570. [Google Scholar]
  27. Yu, C.; Wang, J.; Chen, Y.; Huang, M. Transfer learning with dynamic adversarial adaptation network. In Proceedings of the 2019 IEEE International Conference on Data Mining (ICDM), Beijing, China, 8–11 November 2019; pp. 778–786. [Google Scholar]
  28. Chen, J.; Wu, X.; Duan, L.; Gao, S. Domain adversarial reinforcement learning for partial domain adaptation. IEEE Trans. Neural Netw. Learn. Syst. 2020, 33, 539–553. [Google Scholar] [CrossRef]
  29. Chen, C.; Fu, Z.; Chen, Z.; Cheng, Z.; Jin, X. Towards self-similarity consistency and feature discrimination for unsupervised domain adaptation. Signal Process. Image Commun. 2021, 94, 116232. [Google Scholar] [CrossRef]
  30. Tran, H.H.; Ahn, S.; Lee, T.; Yi, Y. Enlarging Discriminative Power by Adding an Extra Class in Unsupervised Domain Adaptation. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 1812–1819. [Google Scholar]
  31. Guo, D.; Tian, L.; Zhao, H.; Zhou, M.; Zha, H. Adaptive Distribution Calibration for Few-Shot Learning with Hierarchical Optimal Transport. arXiv 2022, arXiv:2210.04144. [Google Scholar]
  32. Deng, Z.; Li, D.; Song, Y.Z.; Xiang, T. Robust Target Training for Multi-Source Domain Adaptation. arXiv 2022, arXiv:2210.01676. [Google Scholar]
  33. Zhao, Y.; Li, M.; Lai, L.; Suda, N.; Civin, D.; Chandra, V. Federated Learning with Non-IID Data. arXiv 2018, arXiv:1806.00582. [Google Scholar] [CrossRef]
  34. Bonawitz, K.; Ivanov, V.; Kreuter, B.; Marcedone, A.; McMahan, H.B.; Patel, S.; Ramage, D.; Segal, A.; Seth, K. Practical Secure Aggregation for Privacy-Preserving Machine Learning. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, CCS ’17, Dallas, TX, USA, 30 October–3 November 2017; Association for Computing Machinery: New York, NY, USA, 2017; pp. 1175–1191. [Google Scholar] [CrossRef] [Green Version]
  35. Geyer, R.C.; Klein, T.; Nabi, M. Differentially Private Federated Learning: A Client Level Perspective. arXiv 2018, arXiv:1712.07557. [Google Scholar]
  36. Chen, F.; Luo, M.; Dong, Z.; Li, Z.; He, X. Federated Meta-Learning with Fast Convergence and Efficient Communication. arXiv 2019, arXiv:1802.07876. [Google Scholar]
  37. Thapa, C.; Arachchige, P.C.M.; Camtepe, S.; Sun, L. Splitfed: When federated learning meets split learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 24–26 February 2022; Volume 36, pp. 8485–8493. [Google Scholar]
  38. Mahmood, Z.; Jusas, V. Blockchain-Enabled: Multi-Layered Security Federated Learning Platform for Preserving Data Privacy. Electronics 2022, 11, 1624. [Google Scholar] [CrossRef]
  39. Cox, B.; Chen, L.Y.; Decouchant, J. Aergia: Leveraging Heterogeneity in Federated Learning Systems. arXiv 2022, arXiv:2210.06154. [Google Scholar]
  40. Liu, C.; Chua, T.J.; Zhao, J. Time Minimization in Hierarchical Federated Learning. arXiv 2022, arXiv:2210.04689. [Google Scholar]
  41. Peng, X.; Huang, Z.; Zhu, Y.; Saenko, K. Federated Adversarial Domain Adaptation. arXiv 2019, arXiv:1911.02054. [Google Scholar]
  42. Ben-David, S.; Blitzer, J.; Crammer, K.; Pereira, F. Analysis of Representations for Domain Adaptation. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–9 December 2006; Schölkopf, B., Platt, J., Hoffman, T., Eds.; MIT Press: Cambridge, CA, USA, 2006; Volume 19. [Google Scholar]
  43. Ben-David, S.; Blitzer, J.; Crammer, K.; Kulesza, A.; Pereira, F.; Vaughan, J.W. A Theory of Learning from Different Domains. Mach. Learn. 2010, 79, 151–175. [Google Scholar] [CrossRef]
  44. Nascimento, M.A.; Özsu, M.T.; Kossmann, D.; Miller, R.J.; Blakeley, J.A.; Schiefer, K.B. (Eds.) Proceedings of the Thirtieth International Conference on Very Large Data Bases, VLDB 2004, Toronto, ON, Canada, 31 August–3 September 2004; Morgan Kaufmann: Burlington, MA, USA, 2004. [Google Scholar]
  45. Zhu, L.; Han, S. Deep leakage from gradients. In Federated Learning; Springer: Berlin/Heidelberg, Germany, 2020; pp. 17–31. [Google Scholar]
  46. Hitaj, B.; Ateniese, G.; Perez-Cruz, F. Deep models under the GAN: Information leakage from collaborative deep learning. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, Dallas, TX, USA, 30 October–3 November 2017; pp. 603–618. [Google Scholar]
  47. Melis, L.; Song, C.; De Cristofaro, E.; Shmatikov, V. Exploiting unintended feature leakage in collaborative learning. In Proceedings of the 2019 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 19–23 May 2019; pp. 691–706. [Google Scholar]
  48. Li, Q.; Wen, Z.; He, B. Practical federated gradient boosting decision trees. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 4642–4649. [Google Scholar]
Figure 1. Structure diagram of a neural network in the experiment.
Figure 1. Structure diagram of a neural network in the experiment.
Applsci 12 10711 g001
Figure 2. System Architecture of Federated Domain Adaptation.
Figure 2. System Architecture of Federated Domain Adaptation.
Applsci 12 10711 g002
Figure 3. Visualization results of features of the source domain and target domain in three-dimensional space.
Figure 3. Visualization results of features of the source domain and target domain in three-dimensional space.
Applsci 12 10711 g003
Table 1. Preprocessed student-related variables and their descriptions.
Table 1. Preprocessed student-related variables and their descriptions.
No.AttributeDescription (Domain)
1addressstudent’s home address type (binary: “U”—urban or “R”—rural)
2famsizefamily size (binary: “LE3”—less or equal to 3 or “GT3”—greater than 3)
3Medumother’s education (numeric: 0—none, 1—primary education (4th grade), 2—5th to 9th grade,
3—secondary education or 4—higher education)
4Fedufather’s education (numeric: 0—none, 1—primary education (4th grade), 2—5th to 9th grade,
3—secondary education or 4—higher education)
5studytimeweekly study time (numeric: 1—<2 h, 2—2 to 5 h, 3—5 to 10 h, or 4—>10 h)
6paidextra paid classes within the course subject (Math or Portuguese) (binary: yes or no)
7activitiesextra-curricular activities (binary: yes or no)
8higherwants to take higher education (binary: yes or no)
9internetInternet access at home (binary: yes or no)
10romanticwith a romantic relationship (binary: yes or no)
11famrelquality of family relationships (numeric: from 1—very bad to 5—excellent)
12freetimefree time after school (numeric: from 1—very low to 5—very high)
13gooutgoing out with friends (numeric: from 1—very low to 5—very high)
14Dalcworkday alcohol consumption (numeric: from 1—very low to 5—very high)
15Walcweekend alcohol consumption (numeric: from 1—very low to 5—very high)
16absencesnumber of school absences (numeric: from 0 to 93)
17G1first period grade (numeric: from 0 to 20)
18G2second period grade (numeric: from 0 to 20)
19G3final grade (numeric: from 0 to 20, output target)
Table 2. Results of factor analysis.
Table 2. Results of factor analysis.
Factor1Factor2Factor3Factor4Factor5Factor6Factor7Factor8
address−0.0174−0.04261−0.029720.55018−0.038830.029520.06146−0.11739
famsize0.05361−0.122590.003060.04264−0.02741−0.005220.59340.0815
pstatus0.02567−0.127430.098570.03471−0.06984−0.02377−0.40380.25312
medu0.020690.43536−0.047740.012690.05549−0.01142−0.021680.02357
fedu0.0210.48458−0.12044−0.05813−0.02211−0.00722−0.08667−0.01257
traveltime0.06977−0.078990.08823−0.427540.05173−0.003240.099230.11188
studytime−0.15099−0.126720.32608−0.050970.073690.017510.077210.21671
failures0.03835−0.14285−0.092680.029180.303490.11095−0.12298−0.1072
schoolsup0.00016−0.03632−0.012820.00861−0.05149−0.022550.000630.06034
famsup−0.00040.056220.4013−0.082190.131350.15243−0.18615−0.20652
paid0.10079−0.078940.531660.02635−0.04372−0.030440.02374−0.10245
activities−0.031360.1021−0.13864−0.133760.047470.03055−0.02090.65063
nursery−0.086740.07820.11152−0.067790.088350.179890.38605−0.06069
higher0.012060.086490.15443−0.06712−0.32819−0.051690.120330.10186
internet0.05957−0.026260.15020.32490.20655−0.07449−0.088870.20321
romantic−0.094280.01830.07294−0.055780.562480.020870.066260.09132
famrel−0.10208−0.070480.068920.0183−0.024250.533520.111730.06412
freetime0.12496−0.01853−0.043920.11156−0.007420.419240.004990.15355
goout0.26257−0.033280.0410.128160.007520.169570.055860.19743
dalc0.376710.027680.07301−0.05661−0.04551−0.045090.01312−0.05814
walc0.402380.014660.03822−0.04011−0.06415−0.08516−0.00015−0.01715
health−0.040720.13569−0.01753−0.189460.035890.44102−0.10537−0.31336
absences0.053410.08776−0.01319−0.016970.38273−0.170250.070090.01427
Table 3. Results of significance test of pairwise correlation.
Table 3. Results of significance test of pairwise correlation.
Variables(1)(2)(3)(4)(5)(6)(7)(8)(9)(10)(11)(12)(13)(14)(15)(16)
(1) address1
(2) famsize−0.1041
−0.302
(3) fedu−0.0090.1121
−0.931−0.268
(4) paid−0.0960.0140.0911
−0.343−0.888−0.366
(5) activities−0.1250.0260.1040.1011
−0.214−0.798−0.305−0.32
(6) romantic−0.0850.09800.0170.0581
−0.401−0.331−0.996−0.864−0.57
(7) famrel0.236 *−0.155−0.1010.1530.024−0.1031
−0.018−0.123−0.318−0.129−0.812−0.306
(8) walc0.078−0.1770.0660.03−0.0980.220 *−0.151
−0.441−0.078−0.516−0.768−0.334−0.028−0.137
(9) medu−0.165−0.0960.023−0.134−0.0130.0260.0790.0561
−0.1−0.34−0.82−0.184−0.899−0.8−0.437−0.578
(10) studytime−0.0710.0050.1100.106−0.0240.0540.0390.0921
−0.482−0.96−0.278−1−0.294−0.81−0.596−0.699−0.363
(11) higher−0.050.208 *0.015−0.190.0140.067−0.159−0.1640.253 *0.11
−0.62−0.038−0.883−0.058−0.887−0.508−0.114−0.102−0.011−0.321
(12) internet−0.050.0940.151−0.058−0.0170.189−0.009−0.206 *0.110.1150.1481
−0.623−0.353−0.134−0.569−0.864−0.06−0.931−0.04−0.278−0.256−0.141
(13) freetime−0.024−0.029−0.0080.1040.075−0.05−0.019−0.0340.032−0.278 *−0.198 *−0.0641
−0.815−0.771−0.935−0.304−0.461−0.623−0.855−0.736−0.752−0.005−0.048−0.528
(14) goout−0.1360.069−0.1420.102−0.029−0.083−0.015−0.1140.07−0.080.182−0.0550.1531
−0.177−0.493−0.159−0.312−0.776−0.413−0.88−0.26−0.49−0.43−0.07−0.585−0.129
(15) dalc−0.025−0.1660.015−0.017−0.056−0.093−0.1380.0810.0360.0140.055−0.0380.0930.228 *1
−0.804−0.098−0.88−0.869−0.583−0.356−0.172−0.423−0.719−0.892−0.587−0.71−0.358−0.022
(16) absences0.0190.0450.0490.1420.069−0.1640.036−0.189−0.0040.065−0.043−0.157−0.225 *0.1040.051
−0.848−0.653−0.626−0.158−0.496−0.104−0.723−0.059−0.965−0.52−0.674−0.118−0.024−0.301−0.624
1* shows that the correlation between the two variables is significant.
Table 4. Training accuracy results of datasets.
Table 4. Training accuracy results of datasets.
Source Only TrainingTarget Only TrainingDomain Adaptation Training
SourceTargetSourceTargetSourceTargetDomain
Scenario 10.90450.51750.5450.87750.920.870.6
Scenario 20.90450.5050.50.50.90350.8750.55
Scenario 30.750.50450.50.550.84550.9050.545
Scenario 40.750.50.450.4750.80350.850.6
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xu, B.; Yan, S.; Li, S.; Du, Y. A Federated Transfer Learning Framework Based on Heterogeneous Domain Adaptation for Students’ Grades Classification. Appl. Sci. 2022, 12, 10711. https://doi.org/10.3390/app122110711

AMA Style

Xu B, Yan S, Li S, Du Y. A Federated Transfer Learning Framework Based on Heterogeneous Domain Adaptation for Students’ Grades Classification. Applied Sciences. 2022; 12(21):10711. https://doi.org/10.3390/app122110711

Chicago/Turabian Style

Xu, Bin, Sheng Yan, Shuai Li, and Yidi Du. 2022. "A Federated Transfer Learning Framework Based on Heterogeneous Domain Adaptation for Students’ Grades Classification" Applied Sciences 12, no. 21: 10711. https://doi.org/10.3390/app122110711

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop