Next Article in Journal
Quantum Dynamical Interpretation of the Mean Strategy
Next Article in Special Issue
Research on a Transformer Vibration Fault Diagnosis Method Based on Time-Shift Multiscale Increment Entropy and CatBoost
Previous Article in Journal
A Hierarchical Neural Network for Point Cloud Segmentation and Geometric Primitive Fitting
Previous Article in Special Issue
Fault Diagnosis Method for Wind Turbine Gearbox Based on Ensemble-Refined Composite Multiscale Fluctuation-Based Reverse Dispersion Entropy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Federated Adversarial Fault Diagnosis Method Driven by Fault Information Discrepancy

School of Logistic Engineering, Shanghai Maritime University, Shanghai 201306, China
*
Author to whom correspondence should be addressed.
Entropy 2024, 26(9), 718; https://doi.org/10.3390/e26090718 (registering DOI)
Submission received: 11 July 2024 / Revised: 21 August 2024 / Accepted: 22 August 2024 / Published: 23 August 2024
(This article belongs to the Special Issue Entropy Applications in Condition Monitoring and Fault Diagnosis)

Abstract

:
Federated learning (FL) facilitates the collaborative optimization of fault diagnosis models across multiple clients. However, the performance of the global model in the federated center is contingent upon the effectiveness of the local models. Low-quality local models participating in the federation can result in negative transfer within the FL framework. Traditional regularization-based FL methods can partially mitigate the performance disparity between local models. Nevertheless, they do not adequately address the inconsistency in model optimization directions caused by variations in fault information distribution under different working conditions, thereby diminishing the applicability of the global model. This paper proposes a federated adversarial fault diagnosis method driven by fault information discrepancy (FedAdv_ID) to address the challenge of constructing an optimal global model under multiple working conditions. A consistency evaluation metric is introduced to quantify the discrepancy between local and global average fault information, guiding the federated adversarial training mechanism between clients and the federated center to minimize feature discrepancy across clients. In addition, an optimal aggregation strategy is developed based on the information discrepancies among different clients, which adaptively learns the aggregation weights and model parameters needed to reduce global feature discrepancy, ultimately yielding an optimal global model. Experiments conducted on benchmark and real-world motor-bearing datasets demonstrate that FedAdv_ID achieves a fault diagnosis accuracy of 93.09% under various motor operating conditions, outperforming model regularization-based FL methods by 17.89%.

1. Introduction

As the core driving components of port cranes, motors operate under complex and variable conditions for extended periods, inevitably facing significant risks of fatigue, wear, and failure [1]. The health status of rolling bearings, which are critical load-bearing components, directly affects the safe operation of these motors. However, the limited working conditions and fault types in a single crane, coupled with significant disparities in data distribution across various conditions, severely impact the optimization of fault diagnosis models [2]. Therefore, developing a fault diagnosis model capable of addressing multiple fault classes under complex working conditions within an FL framework is essential for the effective health management of port cranes.
Existing fault diagnosis methods can generally be categorized into three approaches: signal-based analysis; fault mechanism-based modeling; and data-driven methods [3]. Signal-based fault diagnosis methods identify fault types by extracting sparse representations of faults from monitoring data. However, this approach relies heavily on expert knowledge, making the diagnostic results somewhat subjective. Fault mechanism-based modeling methods, on the other hand, struggle to cope with time-varying nonlinear systems [4]. As a prominent data-driven approach, deep learning (DL), with its powerful nonlinear adaptation capabilities, has been widely employed to process and analyze operational state data to extract fault features. [5,6]. In the area of early fault detection, scholars have developed methods based on stochastic resonance analysis to extract weak fault characteristics from noisy signals for diagnosing early-stage faults in rotating machinery. Additionally, a multi-objective optimization coupled neuron model has been proposed for the identification of early mechanical faults [7,8,9]. For fault diagnosis under multiple working conditions, domain adaptive cross-condition diagnostic models have been developed [10,11]. To address the challenge of diagnosing missing fault types, researchers have developed methods for generating missing type samples and global fault models based on federated learning [12,13]. However, the effectiveness of DL models is highly dependent on large amounts of high-quality and identically distributed (i.i.d.) data. In practice, port cranes typically operate most of their lifetime, resulting in a significantly low proportion of fault data [14]. In addition, since hoist motors typically operate under variable speed and load conditions, acquiring high-quality i.i.d. data becomes very challenging [15]. Data from different operating conditions often exhibit significant distributional disparities, severely compromising the feature extraction capabilities of DL models. As a result, individual clients struggle to train robust fault diagnosis models using limited fault samples from limited working conditions, rendering them ineffective in dealing with potentially unknown working conditions and new fault classes.
FL is a collaborative training mechanism that facilitates the complementation and sharing of information across clients [16]. By fully integrating fault data of various classes and from different working conditions across clients, it is possible to establish a robust global fault diagnosis model. Traditional FL methods effectively alleviate the issue of insufficient fault samples by incorporating collaborative training from multiple clients, thereby improving the fault diagnosis performance of the models [17,18,19]. However, due to the frequent and staged variations in the hoisting speeds of port cranes, motors often experience frequent transitions between start and stop states, resulting in significant disparities in motor working conditions across different clients. Consequently, the fault data collected from crane equipment sensors at different clients exhibit significant distribution heterogeneity. This heterogeneity leads to deviations in the optimization directions of local fault models at each client, ultimately affecting the generalization performance of the global model. Furthermore, it is challenging for a single client to experience and collect data on all possible fault types within the service life of its motors [20]. Therefore, during the model training process, the lack of data for certain fault types may result in the trained model having inadequate recognition capabilities for those fault classes, affecting fault diagnosis accuracy.
Existing improved algorithms based on FL primarily focus on devising various aggregation strategies to improve the quality of global models [21,22,23]. While these methods alleviate some of the challenges posed by multi-condition problems, they do not fundamentally address the intrinsic differences in feature distributions across clients. As a result, global models often struggle to adapt to clients with complex data distributions, particularly in scenarios with limited data samples and complicated multi-condition problems, where existing metrics frequently lack the accuracy needed to measure discrepancies between clients. In FL for fault diagnosis modeling, local models from clients with heterogeneous statistical distributions often exhibit inconsistent optimization directions. This statistical distribution heterogeneity prevents global models from meeting the optimization objectives of individual clients, leading to locally optimal solutions that not only degrade the performance of fault models but may also induce negative transfer phenomena [24]. Current approaches to mitigating statistical distribution heterogeneity primarily address the issue from the perspective of model drift, overlooking the fact that deep-seated discrepancies in fault information between different clients are the root cause of statistical distribution heterogeneity [25,26,27]. These discrepancies can result in inconsistent optimization directions between local models, ultimately causing misalignment between the optimization of global and local models.
This study proposes an innovative FedAdv_ID method. This method integrates a consistency evaluation metric into the local model training process to quantify the discrepancies between drifting local and global average fault information. By establishing an adversarial training mechanism between clients and the federated center, this method aims to eliminate information disparities, ensuring alignment between local optimization directions and the global model optimization direction, thereby significantly reducing the probability of negative transfer phenomena. In addition, a loss function-based optimal aggregation weight learning strategy is designed at the federated center to enhance the generalization capability of the global model. The main contributions of this study are as follows:
  • A federated adversarial fault diagnosis framework driven by information discrepancy is proposed, incorporating a feature consistency evaluation metric to quantify the feature distribution discrepancies between local and global models;
  • A probability distribution consistency loss function is developed to facilitate adversarial training between client parameters and the federated central model parameters. This approach aims to eliminate distribution discrepancies among clients and addresses the issue of negative transfer in federated learning;
  • An optimal model aggregation strategy driven by fault information discrepancy is designed to guide the construction of the global model, enhancing its generalization capability;
  • The proposed method demonstrates robust fault diagnosis performance when evaluated on benchmark and real-world datasets under various working conditions, including scenarios with incomplete fault class information.

2. Related Works

2.1. Fault Diagnosis Based on FL

FL facilitates the joint optimization and training of fault diagnosis models across multiple clients while avoiding privacy conflicts through model aggregation [28]. Chen et al. [9] proposed a federated transfer learning (FTL) framework based on difference-based weighted federated averaging (D-WFA) to collaboratively train well-trained global diagnostic models and designed an aggregation strategy using a dynamic weighted averaging algorithm with maximum mean difference (MMD) but ignored the inconsistency of the features between clients. Lu et al. [29] proposed a novel clustering-based FL algorithm. This algorithm involves training a probabilistic deep learning model, allowing each client to assess the prediction uncertainty of other clients’ models on its local dataset. Clients are then clustered based on the similarity of their datasets, thereby reducing the uncertainty across clients with varying data quality and ensuring the effectiveness of the federated process. Wang et al. [22] proposed a novel federated contrastive prototype learning scheme that facilitates collaborative modeling between the central server and multiple clients to establish a global fault diagnosis model while preserving data privacy. Mehta et al. [23] developed a duplex classifier based on 1D-CNN, which separates mixed fault classification tasks into two parallel networks for diagnosis. Zhang et al. [30] introduced a multi-scale recursive FL framework that allows the federated center to prioritize valuable client information. This framework utilizes federated center information through local multi-scale feature fusion, thereby constructing a reliable fault diagnosis model capable of handling multiple working conditions. Chen et al. [31] addressed the challenge of insufficient training data collection by individual industrial entities by incorporating capsule-based fault feature expressions into an FL framework. This approach reduces the transmission burden and enhances data security. Qin et al. [32] proposed an improved FL algorithm that reduces computational costs through wavelet packet decomposition. This algorithm employs SecureBoost to train local diagnostic models during the modeling process, aggregating model parameters from multiple railway lines in each iteration. Du et al. [33] tackled the issue of insufficient generalization capability in fault models with small samples by combining the Forgetting Kalman Filter (FKF) with Cubic Exponential Smoothing (CES) during the model aggregation process. This approach enables high-precision fault diagnosis in mixed fault scenarios across multiple clients.
Existing FL frameworks generally assume that each client is ideally isomorphic. However, different clients have diverse data distributions, and FL needs to deal with the statistical distribution heterogeneity to achieve effective joint optimization. Shi et al. [34] proposed a method that corrects the optimization direction of the model by introducing a sparse regularization term in the optimization function of the local client, thereby inhibiting the drift of the local model. This approach has proven effective in mitigating statistical distribution heterogeneity. However, this algorithm primarily focuses on eliminating discrepancies with the global model at the model level. Therefore, fully leveraging fault information, such as features from different clients under multiple working conditions, is crucial for enhancing the effectiveness of FL-based fault diagnosis in the presence of statistical data distribution heterogeneity.

2.2. Federated Heterogeneity Elimination under Multiple Working Conditions

Existing FL approaches are typically designed for joint optimization under the assumption that all clients operate under ideal homogeneous conditions. However, motors at different client locations function under diverse operational environments, which can significantly impact model optimization across clients and adversely affect the generalization performance of the global model.
Li et al. [35] addressed feature skew in statistical heterogeneity by adding batch normalization layers to local models, but the effectiveness of this method in handling more complex heterogeneity, such as quantity skew, remains uncertain. Yoon et al. [36] proposed a global Mixup method that allows clients to access averaged data from other clients, optimizing their data distribution to mitigate the non-i.i.d issue. Wang et al. [37] proposed a general framework for analyzing the convergence of federated heterogeneous optimization algorithms. They provided a first-principle understanding of solution bias and the slowdown of convergence due to objective inconsistencies, which can eliminate discrepancies in the presence of local quantities. Chen et al. [38] proposed a dynamically weighted federated aggregation strategy that uses the Manhattan distance to measure the discrepancy between local and global models as the basis for weight aggregation, effectively reducing the impact of poor local data. However, the Manhattan distance may fail to converge when a client’s data distribution is complex under multiple working conditions. To address the non-i.i.d. data heterogeneity problem, Chen et al. [39] proposed a synthetic data generation method. This approach employs knowledge distillation to learn from heterogeneous models of participating clients, facilitating model aggregation at the central server. Wang et al. [40] proposed a dynamic model aggregation algorithm to mitigate the impact of low-quality clients by evaluating each client’s weight based on its contribution to the total consensus diagnostic knowledge and then aggregating local models using adaptive weights. However, this approach performs poorly when confronted with unknown and complex operational tasks due to lacking more informative data.
Current strategies for mitigating federated heterogeneity under multiple working conditions primarily focus on two approaches: regularization terms and transfer learning. Regularization terms address heterogeneity solely from a model perspective, leading to significantly compromised diagnostic performance when confronted with data from newly emerging fault classes. Conversely, transfer learning methods aim to align feature distributions across different working conditions at the feature level; similarity clustering of client performance under a certain dataset can only be used for client performance discrimination and cannot be personalized for a specific low-quality client. However, their diagnostic performance exhibits considerable variability. These existing methods show limitations when dealing with complex multi-condition FL scenarios.

3. Federated Adversarial Fault Diagnosis Methodology Driven by Information Discrepancy

  • Step 1: Federated adversarial fault diagnosis framework driven by information discrepancy
The accuracy of high-dimensional fault features and probability density distributions extracted by various clients directly influences the performance of their diagnostic models. However, due to the heterogeneity of feature distributions across clients, this fault information can undergo significant drift, making the global model misaligned with the optimization direction of local training. This misalignment ultimately reduces fault diagnosis accuracy for each client. Existing methods primarily address the issue of global model bias caused by inconsistent optimization directions among clients from the perspective of model drift. However, these approaches fail to fully exploit the fault information shared among clients and overlook the impact of fault information drift on the generalization capability of the global model.
This section establishes the FedAdv_ID framework between clients and the federated center. The framework includes metrics for measuring global fault information and fault information discrepancies, a federated adversarial training mechanism to address statistical distribution inconsistencies between clients, and a federated aggregation strategy driven by fault information discrepancy. This design ensures that the federated center can utilize client information both fairly and efficiently.
Specifically, the federated center integrates client fault information to derive the global average fault information. A consistency evaluation metric is designed based on the distance between the client and global average fault information. This metric guides the federated adversarial training and is the updated direction for the client models to reduce discrepancies between local and federated fault information. After each round of federated adversarial training, the discrepancies in the consistency evaluation metric between local and global average fault information are used to assign aggregation weights to different client models. This approach ensures that rare fault information from lower-quality clients operating under various conditions is effectively incorporated into the federated diagnosis model. The schematic diagram of the FedAdv_ID method is shown in Figure 1, and implementation details are provided in the following steps.
  • Step 2: Federated global fault information metric
During the t 1 round of communication, the federated center initializes the federated discriminator model parameter θ d t 1 and the global model θ g t 1 , which are then distributed to each client. Upon downloading the global model, each client uses locally available bearing data with inconsistent statistical distributions ( X i , i = 1,2 , n ) to train their local model parameters θ i t 1 . The feature extraction process for the clients is described in Equation (1).
F T k t 1 = F ( θ g t 1 , X i ) , i 1 , 2 , , n
where k denotes the client index, and F T k t 1 denotes the high-dimensional fault features extracted by the k-th client after the t 1 round of communication, F ( · ) denotes Feature extraction networks. Subsequently, these high-dimensional features are input into the classifier layer to obtain the probability distribution of different fault classes, as shown in Equation (2).
Y k t 1 = exp ( w j F T k t 1 ) j = 1 c exp ( w j F T k t 1 ) , k 1 , 2 , , K
where Y k t 1 denotes the probability distribution of the fault samples belonging to different fault categories for client k; c denotes fault classes.
The clients then upload the fault information and the corresponding classes probability distribution ( F T k t 1 , Y k t 1 ) to the federated center. The federated center integrates fault information from different clients and computes the average to form a global fault information set, as shown in Equations (3) and (4).
F T g t 1 = 1 K k = 1 K F T k t 1
Y g t 1 = 1 K k = 1 K Y k t 1
where K denotes the total number of clients participating in the federation. F T g t 1 and Y g t 1 denote the average fault feature information and average classification probability information at the federation center, respectively. The new fault information set encompasses the average fault information from all clients, effectively addressing the issues of insufficient richness in local fault features and low probability density entropy. Furthermore, the average fault information serves as the optimal direction for updating fault information in the local models toward the global model, guiding client models to update in a direction that reduces information discrepancies.
This approach results in an optimal global fault diagnosis model that integrates fault information from various working conditions. It is important to note that when discrepancies exist in the number of fault samples across clients, batch gradient descent optimization or the expansion of output features and their corresponding probability distributions is necessary to ensure the accurate calculation of the federated center’s average fault information. The process of expanding the fault information from clients is detailed in Equations (5) and (6).
F T k t 1 = B r o a d c a s t [ F T k t 1 , 0 ]
Y k t 1 = B r o a d c a s t [ Y k t 1 , 0 ]
  • Step 3: Consistency evaluation metric for fault information discrepancy
The participating federated clients possess only limited local working conditions and fault information, reflecting the localized characteristics of fault samples. In contrast, the federated center’s global fault information integrates fault data from diverse working conditions. Therefore, a consistency evaluation metric based on Step 2 is used to measure the discrepancies between client-specific fault information and the federated global fault information. This metric guides the adversarial federated training process, aiming to reduce inconsistencies between local and global fault information. Consequently, the optimization directions of the client model parameters become more aligned, ensuring the effective aggregation of the federated model. The construction process of the consistency evaluation metric is detailed in Equations (7) and (8).
L f c ( F T k t 1 , F T g t 1 ) = 1 j n = 1 j φ ( f t k , n t 1 ) 1 j m = 1 j φ ( f t g , m t 1 ) H
L c l f ( Y g t 1 , n , Y k t 1 , n ) = i = 1 n y g n log ( y g t 1 , n y k t 1 , n )
where L f c and L c l f denote the feature consistency evaluation metric and the probability consistency evaluation metric, respectively. φ ( · ) denotes the mapping to the Reproducing Kernel Hilbert Space (RKHS) of the features [41]. f t k , n t 1 denotes the high-dimensional features of the n-th sample from client k in the t−1 communication round, and y k t 1 denotes the probability distribution of fault categories for the n-th sample from client k in the (t − 1) communication round.
The metrics L f c and L c l f quantify the discrepancies between local fault information and the average fault information. These metrics guide the optimization directions of feature extractors and classifiers across clients, thereby mitigating the issue of fault information drift caused by statistical distribution heterogeneity.
  • Step 4: Federated adversarial optimization for feature inconsistencies between clients
The statistical distribution heterogeneity of local client fault data leads to a drift in the fault information of clients. To mitigate the negative impact of this drift on model optimization, it is crucial to eliminate the discrepancies between the clients’ fault information and the federated center’s average fault information. To this end, a federated adversarial training mechanism is designed to address statistical distribution inconsistencies among clients, thereby reducing the discrepancies between their fault information and that of the federated center. The schematic diagram of the algorithm designed to eliminate these discrepancies is presented in Figure 2.
The purpose of federated adversarial training is to eliminate discrepancies in features and probability distributions across different working conditions through a max-min game between local fault information and global average fault information. This process enhances the generalization capability of the global model. By incorporating the consistency evaluation metric as a regularization term in the adversarial loss function, the model penalizes fault information drift caused by the statistical distribution heterogeneity of local clients. This approach reduces the extent of fault information drift and improves the accuracy of high-dimensional fault features extracted by the feature extractor, as well as the probability density distribution estimated by the classifier. The value function for adversarial federated training is provided in Equations (9) and (10).
Loss adv = min c k max D V ( C k , D ) + L f c + L c l f
min c k max D V ( C k , D ) = E F T g t 1 ~ P F T g t 1 [ log D ( F T g t 1 ) ] + E F T k t 1 ~ P F T k t 1 [ log ( 1 D ( F T k t 1 ) ) ]
where C k denotes the client generator, and D denotes the federated center discriminator. L f c denotes fault feature consistency regularization term; L c l f denotes fault information probability distribution consistency regularization term. F T g t 1 ~ P F T g t 1 denotes a sample feature that follows the distribution of the average fault features, while F T k t 1 ~ P F T k t 1 denotes the fault feature distribution of a sample from client k. During adversarial training, the clients minimize the loss function in Equation (9) to optimize the generator parameters, continuously reducing the discrepancies between fault features. The federated center’s discriminator maximizes the adversarial loss function to optimize its parameters. The optimization directions of the two parts are entirely opposite, forming a minimax adversarial game between the “two players,” ultimately reaching a steady-state Nash equilibrium [42].
It is important to note that the adversarial loss function in Equation (9) can experience significant instability during the optimization of generator and discriminator parameters due to differences in fault information among clients. Therefore, adding consistency metrics to the adversarial loss function can guide the fault features extracted by each client to align with the federated center’s average fault features, ensuring the consistency of the optimization direction of the client generators and thereby reducing the instability in the optimization of generator parameters. Additionally, the optimized generator-extracted features can better deceive the discriminator, ensuring the stability of the discriminator’s optimization process. However, when there are significant differences in fault features among clients, the quality of the federated center’s average fault feature information may be affected. To ensure the stability of the generative adversarial process, new regularization terms need to be designed to penalize the imbalance in the optimization process. Furthermore, for clients with poor fault information quality, it is necessary to set a threshold for their adversarial loss to reduce fluctuations in training loss. By introducing new gradient penalty terms into Equation (9), potential instability in adversarial training can be addressed.
Loss a d v k = min c k max D V ( C k , D ) + L f c + L c l f + γ E F T k t 1 ~ P F T k t 1 ( F T k t 1 D ( F T k t 1 ) 2 1 ) 2
where E F T k t 1 ~ P F T k t 1 F T k t 1 D F T k t 1 2 1 2 denotes the gradient penalty term; γ denotes the gradient penalty term coefficient.
Due to the significant statistical distribution differences in the fault data used by clients for training, minimizing V C k , D would increase the confidence of the federated center’s discriminator, thereby rejecting new samples generated by clients through local data and adversarial training. This outcome is counterproductive to the goal of enabling local clients to learn the global average fault features and reduce distribution discrepancies. Therefore, this study opts to have clients maximize the loss function while the central server’s discriminator minimizes the loss function, creating a game between the local clients and the central server’s discriminator. This approach provides stronger gradients for early adversarial training. The process of backpropagating the adversarial training loss function L o s s a d v based on the consistency evaluation metric to optimize the parameters of the federated discriminator model and the client generator models is detailed in Equation (12).
{ θ d t = θ d t 1 l r L o s s a d v d θ d t 1 θ 1 t = θ 1 t 1 l r L o s s a d v 1 θ 1 t 1 θ 2 t = θ 2 t 1 l r L o s s a d v 2 θ 2 t 1 θ k t = θ k t 1 l r L o s s a d v k θ k t 1
where θ d t denotes the updated parameters of the federated discriminator; θ k t 1 denotes the backpropagation-optimized model parameters of the k-th client, and lr is the learning rate. Equation (12) demonstrates that when each client’s generator engages in adversarial training with the federated discriminator, both sets of parameters are updated simultaneously. Through multiple rounds of adversarial training, the federated discriminator optimizes its parameters and enhances its discriminative ability. This process encourages clients to actively improve their local models’ learning of average fault information.
It is important to note that because the optimization directions of the discriminator and the client’s generator are opposite, the adversarial training error will decrease but eventually converge to a Nash equilibrium point. Let L o s s c k = min c k V ( C k , D ) denote the loss function of the client and L o s s d = max d V ( C k , D ) denote the loss function of the discriminator. The equilibrium point obtained through adversarial training is given by Equation (13).
lim t | Loss c k Loss d | = N , N R +
where N denotes the Nash equilibrium point at which the client generator and the federated discriminator can no longer update through adversarial training.
When L o s s a d v continually decreases and reaches a predefined threshold, it indicates that the clients have been effectively optimized through adversarial training, and the discrepancies between local fault information and aggregate fault information are minimized. This process mitigates the fault information drift caused by statistical distribution heterogeneity. Through federated adversarial training among clients, local fault information becomes initially adaptable to the global fault diagnosis model. However, the performance of this global model in local fault diagnosis tasks is contingent upon the soundness and reliability of the global model aggregation strategy. To address this, an information discrepancy-driven federated aggregation strategy has been developed to ensure that the federated center achieves an optimal global diagnosis model.
  • Step 5 Effective recognition of fault features by clients
In step 4, by measuring the discrepancy between each client’s fault information and the federated center’s average fault information, adversarial optimization effectively resolves the inconsistency in model optimization directions caused by differences in operating conditions between clients. This approach ensures that the federated diagnosis model can generalize fault information across different client distributions. However, the diagnostic accuracy of the federated model relies heavily on the training effectiveness of the client models. Therefore, while eliminating the discrepancies in fault information among clients, it is also essential to further analyze the fault information inconsistency metrics and the guidance provided by adversarial loss during the client model training process. This ensures that the local models can effectively recognize fault features.
The local client model training process should not only include a fixed cross-entropy loss to achieve fault classification accuracy on private datasets but also incorporate an adversarial loss function to address fault information discrepancies. This optimizes the feature extraction network to progressively reduce fault feature differences among clients. Additionally, to better measure the discrepancy between client fault features and the federated center’s average features, a feature discrepancy regularization term needs to be added during the client training process to ensure consistent optimization direction of the local model. The collaborative optimization of these three loss functions ensures that the local model can effectively extract well-represented fault features while further reducing fault information discrepancies among clients. It should be noted that the fault information discrepancy regularization term is already included in the adversarial loss. Therefore, the training loss of the local model can be expressed as
L o s s c l i e n t k = L o s s a d v k + ( c i C y c i log ( y ^ c i ) )
where C denotes fault type; y c i denotes the true label of the sample, and y ^ c i denotes the predictive label of the sample.
As described in Equation (14), the parameter optimization process of the local model in federated learning follows the formulation outlined in Equation (15).
{ θ 1 t , n = θ 1 t , n 1 l r L o s s c l i e n t 1 θ 1 t , n 1 θ 2 t , n = θ 2 t , n 1 l r L o s s c l i e n t 2 θ 2 t , n 1 θ k t , n = θ k t , n 1 l r L o s s c l i e n t k θ k t , n 1
where θ k t , n denotes the parameters of the k-th client after the n-th iteration of optimization in the t-th round of the federated process, and l r denotes the learning rate for local model optimization.
In this process, the adversarial loss is generated at the federated center and then transmitted to the client, where it works in conjunction with the cross-entropy loss function to optimize the model parameters. This approach does not impede the real-time updating of local model parameters, as the adversarial loss does not require multiple iterations at the federated center. Instead, once the global model is received, the local model undergoes multiple iterative updates on the client’s private data. Consequently, both the global model and the adversarial loss can be delivered simultaneously to the client, facilitating the optimization of the local model.
  • Step 6: Federated aggregation strategy driven by fault information discrepancy
The information discrepancy between different clients varies, and the federated center should account for this when aggregating local models by assigning weights accordingly. Clients whose local fault information significantly deviates from the average fault information tend to perform poorly in adversarial training and should be assigned less weight. Conversely, clients with smaller discrepancies perform better and should be assigned greater weight. In this section, we propose a federated aggregation strategy driven by fault information discrepancies, where the weights are linked to the federated adversarial loss function. This approach ensures that the allocation of aggregation weights is determined solely by the discrepancy between the client and global fault information, as illustrated in Equation (16).
Loss sever = k = 1 K P k t 1 Loss a d V k
where P k t 1 denotes the aggregation weight of the k-th client in the t−1 round of the federation. The adversarial loss function adaptively learns the aggregation weights based on the client discrepancies, reducing the overall inconsistency in fault information to obtain the optimal global model. The optimization process for the client generator and aggregation weights is detailed in Equation (17).
{ P k t 1 = P k t 1 l r L o s s s e r v e r p k t 1 θ k t 1 = θ k t 1 l r L o s s s e r v e r θ k t 1
The global fault model aggregation process for the new federated round is shown in Equation (18) derived from Equation (17).
θ g t = k = 1 K P k t 1 θ k t 1
In the new aggregation round, the federated center’s loss function is constructed based on this discrepancy, and it adaptively learns and updates the optimal aggregation weights and model parameters according to this loss function. As a result, it is distributed to each local client for a new round of information discrepancy-driven adversarial federated fault model training.
The dynamic interaction between clients and the federated center regarding fault information and model parameters implies that a client’s fault feature distribution evolves with changing operating conditions. In response, the client extracts new fault features using the generator and uploads them to the federated center, thereby updating the federated center’s average fault information. As a result, the consistency evaluation metric between the client’s fault features and the average fault features is adjusted, leading to a new adversarial loss that optimizes the parameters of both the federated center’s discriminator and the client’s generator. Changes in the adversarial loss value clearly reflect the quality of fault features across clients and the effectiveness of their local models. Therefore, it is essential to fully consider the data quality attributes among clients during the federated model aggregation process. At this stage, we introduce a learnable parameter P k that adapts based on changes in the adversarial loss value between the client’s fault features and those of the federated center. This mechanism ensures that clients with varying data quality are assigned weight parameters that correspond to their data quality during model aggregation. Consequently, the federated model can dynamically adapt to the continuously evolving data distributions across clients, thereby enhancing its generalization capability.

Fault Diagnosis Based on FedAdv_ID

The implementation process of the FedAdv_ID method comprises two parts: offline training and online diagnosis. The proposed algorithm trains the fault diagnosis model using multiple working condition data within the client during the offline training stage. The training process culminates when the loss function at the federated center achieves a predefined threshold, at which point both the global and local model parameters are saved. Subsequently, each client utilizes its local model to extract fault features from online monitor data, thereby diagnosing the motor’s fault state at the current moment, as illustrated in Figure 3.
The objective of adversarial training in FedAdv_ID is to identify consistent fault features by reducing discrepancies in fault characteristics among clients rather than merely suppressing the differences in fault features. The underlying principle of this adversarial mechanism is to guide the updating of client generator parameters by addressing inconsistencies between features, thereby extracting new and relevant feature information. Essentially, the overall amount of fault feature information is not diminished; instead, the critical information is preserved while the heterogeneity of fault features across clients is progressively reduced. This process indirectly enhances the representation of useful fault features. However, in this context, useful fault features are not necessarily given greater weight; rather, the degree of suppression of useful features is significantly lessened compared to that of heterogeneous features during the adversarial training process. Furthermore, the federated aggregation strategy, driven by fault information discrepancies, fully accounts for the data quality among clients, effectively mitigating the negative transfer effects caused by low-quality clients within the federated model. As a result, through the synergistic collaboration of these steps, FedAdv_ID primarily focuses on suppressing irrelevant fault information across different clients. While some useful information may inevitably be affected during this process, the impact is minimal compared to the suppression of irrelevant fault information.

4. Experiments and Analysis

The Case Western Reserve University (CWRU) bearing dataset [43] and the Shanghai Maritime University (SMU) electric motor fault simulation experimental platform are utilized as fault data sources to evaluate the performance of the proposed method under varying statistical distributions within FL. The experimental design encompasses a range of scenarios and employs accuracy metrics to assess the method’s performance across diverse conditions. Furthermore, a comparative analysis with several state-of-the-art methods is conducted to validate the superiority of the proposed approach.

4.1. Experimental Design and Analysis of the Motor Bearing Benchmark Dataset

The CWRU motor bearing platform is shown in Figure 4. Vibration acceleration signals are collected by sensors installed at both the drive and fan ends of the motor. The sampling frequencies are 12 kHz and 48 kHz at the drive end and 12 kHz at the fan end. The motor shaft, supported by the test bearing, operates under four load conditions: 0 HP; 1 HP; 2 HP; and 3 HP, corresponding to approximate motor speeds of 1797 rpm, 1772 rpm, 1750 rpm, and 1730 rpm, respectively.
The fault data in this section are derived from vibration signals collected by sensors at a sampling frequency of 48 kHz at the motor drive end. The dataset comprises four health states: one normal state and three fault states. The fault states correspond to rolling bearing damage at three different locations: the ball; inner race; and outer race, each with a defect size of 0.007 inches. A sliding window approach, with a window length of 100 and a step size of 50, is applied to form the final dataset. The time–domain waveform diagrams for each fault type at the drive end are illustrated in Figure 5, with the corresponding labels provided in Table 1.
Four experimental scenarios were designed to validate the effectiveness of the proposed FedAdv_ID fault diagnosis method under various working conditions. Each scenario includes three client participants within the federation. Although multiple clients may be involved in real-world applications, using a limited number here demonstrates the method’s efficacy. The specific experimental details are shown in Table 2.
To minimize the impact of random variability in the experimental results, each experiment was repeated five times with different random seeds, and the average value was calculated. To verify the superiority of the proposed method, advanced FL approaches that address statistical heterogeneity were employed as comparative baselines, and these comparative experiments were conducted under identical conditions. Table 3 provides a detailed description of the comparative methods used, while Table 4 outlines the architecture of the client generator models and the associated hyperparameters for each method.

4.2. Experimental Results Analysis for the CWRU Dataset

To explore the sample distribution under different working conditions, we visualized the probability density distributions of fault samples from Experiments 1 and 2, as shown in Figure 6. In Experiment 1, all clients operated under two stable conditions, with typical fault features within each client exhibiting varying degrees of drift. In Experiment 2, all clients operated under three conditions, where the intrinsic features of the same fault type displayed diverse drift patterns. This visualization analysis provides deeper insight into how different working conditions influence the distribution of fault features.
Figure 6 displays the probability density distribution curves (solid lines) of the samples, along with their corresponding means and variances (dashed lines). In Figure 6a, Client 1’s mean converges to approximately zero, although the distribution remains uneven; Client 2 shows a distinct bimodal distribution, while Client 3 exhibits a relatively uniform distribution. Figure 6b reveals that all clients demonstrate some degree of left skewness and bimodality, reflecting the increasing statistical distribution heterogeneity. Table 5 presents the fault diagnosis results of the proposed method compared with those of the other methods under the above experimental scenarios.
The results of Experiment 1 reveal significant differences in fault diagnosis accuracy across the various methods. The DNN approach, which was unable to detect feature drift, achieved an average accuracy of only 53.80%. FedAvg improved accuracy by 7.74% through joint optimization but failed to account for the differing optimization directions among clients. FedRL_F1 introduced F1 regularization, enhancing adaptability under different statistical distributions, and improved accuracy by 13.66% compared to FedAvg. However, it did not address the impact of statistical distribution heterogeneity on error information. Fed_BN tackled feature scale disparities but did not fully resolve feature drift issues. The experiment demonstrated that larger loads and more significant drifts had a pronounced impact on fault diagnosis accuracy. WAFedL employed an adaptive weighting strategy, which improved accuracy by approximately 8% over Fed_BN but potentially neglected fault information from low-weight clients. Building on FedRL_F1, FedAdv_ID addressed information drift by incorporating fault information consistency evaluation and an adversarial federated framework. This approach effectively resolved feature drift issues while ensuring alignment between individual client optimization directions and the global model. As a result, it mitigated the potential risks associated with WAFedL’s handling of low-weight clients and addressed Fed_BN’s limitations in managing feature drift.
Figure 7 presents the confusion matrices for the different methods used in Experiment 1. Figure 7a illustrates that while the DNN method effectively distinguishes between normal and faulty samples, it tends to misclassify ball flaws as normal conditions. In addition, the DNN method demonstrates lower accuracy in identifying inner-race and outer-race defects compared to its performance with ball defects and normal conditions, likely due to the similarities in the defect characteristics of the inner and outer races. Figure 7b shows that FedAvg performs well in detecting outer-race faults and normal conditions, significantly improving diagnostic accuracy for ball faults; however, its performance in identifying inner-race faults remains suboptimal.
The methods depicted in Figure 7c–e demonstrate varying degrees of improvement in fault diagnostic accuracy. Notably, the ball faults and normal conditions consistently achieve the highest detection rates across most diagnostic models, although they are prone to being misclassified as each other. In contrast, inner-race and outer-race faults exhibit lower diagnostic rates and are more susceptible to cross-misclassification.
Experiment 2 introduced more complex working conditions, leading to various feature shifts within the same fault type, which posed significant challenges to the model’s feature extraction capabilities. Under these conditions, the DNN method experienced a notable 8.73% decrease in diagnostic accuracy compared to Experiment 1, as it struggled to accurately identify fault categories. FedAvg’s performance declined by 7.54% due to the increased complexity of the statistical distributions, which diminished the effectiveness of its aggregation strategy. The limitations of FedRL_F1, which primarily addresses model drift, became more pronounced under these conditions. Although it mitigated discrepancies between the model and the federation center, it failed to address the heterogeneity of data distribution, resulting in the largest accuracy drop of 13.17%. Fed_BN experienced an 11.99% decrease in accuracy, indicating that significantly shifted features continued to affect decision boundaries even when feature scaling was applied. While WAFedL experienced a relatively smaller drop in accuracy, its overall diagnostic accuracy of 75.20% suggests that its improved aggregation strategy still has limitations in optimizing specific clients. This is further evidenced by the significantly lower diagnostic accuracy of client 1 in Experiment 2 compared to Experiment 1.
In contrast, FedAdv_ID maintained the highest fault diagnostic accuracy in Experiment 2, exhibiting the smallest decrease in performance. This outcome underscores the method’s effectiveness in enhancing the generalization capability of the global model by eliminating fault information disparities and maintaining stable diagnostic performance across different experimental scenarios. Moreover, the federated aggregation strategy based on information disparity ensured alignment between the global model and local client optimization directions, facilitating improvements in local models and boosting fault diagnostic accuracy. Even under complex working conditions, FedAdv_ID achieved an impressive diagnostic accuracy of 86.06%, demonstrating its superior performance.
Experiments 3 and 4 further investigate the effectiveness of different methods in scenarios with missing fault types, building on the foundations of Experiments 1 and 2. When the failure data types differ between clients, the statistical distribution differences increase significantly (Table 6).
In Experiment 3, the DNN method that relies only on its local data for training not only suffers from low diagnostic accuracy due to multiple working conditions but also fails to learn missing fault features from other clients, resulting in further performance degradation compared to Experiment 1. Although FedAvg can learn fault features from other clients through a global model, its diagnostic effectiveness for missing faults remains suboptimal due to feature shift effects. FedRL_F1 and Fed_BN specifically address the problem of statistical distribution heterogeneity, but their ability to learn missing defect types remains limited to the global model level, yielding limited improvements. WAFedL improves the diagnostic accuracy of missing defect types by optimizing the global model, improving about 20.67% over FedAvg. However, its learning is limited to the model level. FedAdv_ID, on the other hand, generates comprehensive fault information that includes missing fault data from different clients through the federation center’s information fusion module. Using a federated adversarial framework helps eliminate inconsistencies in fault information. This approach enables learning of missing fault characteristics and overcomes the problems of insufficient fault characteristics and low probability density entropy in individual clients. As a result, it provides richer fault information for model diagnostic decision-making, thereby improving the accuracy of identifying missing fault types.
Experiment 4 builds on the complex multi-condition scenario of Experiment 2 to validate the effectiveness of the approach in extreme cases where fault types are missing. In this scenario, not only are certain fault types missing across the three conditions, but the statistical distribution heterogeneity is further exacerbated, making it impossible to effectively diagnose the missing fault types. While most methods become nearly ineffective in this context, FedAdv_ID can adaptively retrieve the missing fault information from other clients through fault feature discrepancy measurement, assigning greater weight parameters to ensure that the aggregated model can diagnose the missing faults and maintain a certain level of diagnostic accuracy. This demonstrates that FedAdv_ID not only eliminates fault information discrepancies among clients but, more importantly, captures useful fault information from clients. It also effectively leverages unique fault information within imbalanced fault categories, thereby enhancing the fault diagnosis capability of the federated model. In addition, compared to direct aggregation by FedAvg, the adversarial loss function with regularization ensures consistency in the optimization direction of client parameters, reducing distribution differences between fault features. The aggregation strategy based on fault information differences fully considers the quality of client data, ensuring that useful information from low-quality clients is effectively integrated into the federated model. This improves the utilization of useful information, allowing FedAdv_ID to achieve better fault diagnosis results in more complex data scenarios compared to other methods.

4.3. Description of SMU Test Bench and Dataset

This section aims to evaluate the effectiveness of the proposed method on a multi-sensor, multi-condition dataset. This experiment utilizes data collected from the motor fault test bench at SMU. In multi-sensor signal fault diagnosis tasks, each data channel captures different characteristic manifestations of the same fault, providing rich and diverse data. This multidimensional information allows for a more comprehensive assessment of the motor’s operational state, offering a more accurate basis for fault diagnosis. However, the richness of this information also exacerbates the disparity in fault information across clients, potentially impacting the diagnostic performance of the federated model. To address this challenge, this study employs the SMU motor dataset to validate the proposed method’s effectiveness in eliminating fault information disparities. Special attention is given to its diagnostic accuracy in complex scenarios, such as multiple operational conditions and missing fault categories. Through this experiment, we aim to demonstrate the robustness and adaptability of the proposed method in handling high-dimensional, heterogeneous data, thereby providing stronger support for its application in real industrial environments.
The SMU motor fault test bench is shown in Figure 8. This platform comprises a 2.2 kW three-phase asynchronous motor manufactured by ABB Group (The motor model is M2QA), a motor controller, a torque sensor, a magnetic powder brake, and data acquisition software. Various fault conditions can be simulated by replacing motors with different defects.
This platform simultaneously monitors physical quantities across seven channels, including vibration signals, three-phase currents (U, V, W), torque, and rotational speed. Vibration signals are captured via accelerometers installed at both the fan and drive ends; three-phase currents are monitored by integrating current clamps with the input wiring; and torque and rotational speed are measured using dedicated sensors. The loading device can provide a load range of 0–10 Nm.
As shown in Table 7, the platform can simulate a diverse array of fault types. Figure 9 presents the time–domain waveforms of various monitoring signals under normal working conditions. This comprehensive data acquisition and fault simulation capability provides an ideal experimental environment for in-depth research on motor fault diagnosis.

4.4. Complex Experimental Scenario Design for SMU Dataset

In this section, we focus on vibration signals sampled at 25.6 kHz. A sliding window technique segments the original vibration signals and facilitates practical data analysis. The window length is set to 100 data points with a step size of 50 points. This segmentation approach ensures data continuity while providing sufficient overlap to capture the dynamic characteristics of the signals.
Table 8 provides a detailed list of the fault types considered in this study, along with their corresponding labels. The selection of these fault types aims to encompass a wide range of anomalous conditions commonly encountered in motor operations, thereby offering a comprehensive testing basis for fault diagnosis algorithms. Through this systematic approach to data processing and fault classification, we aim to establish a robust experimental framework to thoroughly validate the effectiveness and generalization capability of the proposed method.
To thoroughly evaluate the effectiveness and robustness of the proposed method, particularly its performance under extreme scenarios, we designed two sets of experiments that encompass a broader range of working conditions. These experiments are intended to simulate the complex and variable operating conditions encountered in real industrial environments. Table 9 provides a detailed overview of the parameters and settings for these two experimental sets.
The model structure and hyperparameters of the comparison methods involved and the proposed method are shown in Table 10.

4.5. Experimental Results Analysis for the SMU Dataset

The SMU motor fault tester uses a multi-sensor system that collects data from multiple channels, including vibration signals, current signals, and rotational speed, as the basis for fault diagnosis. Compared to the single-channel sensor data from the CWRU dataset, this multidimensional data collection approach provides richer and more diverse fault information. The results of Experiment 1 show that even when processing such complex multi-channel fault data, the FedADV_ID method exhibits superior diagnostic performance. FedADV_ID shows excellent fault diagnosis accuracy compared to other methods, improving by 37.78% over the DNN method and reducing the error rate by 46.55% over the second-best method. This significant performance improvement can be attributed to the rich features of multi-channel fault data, which enables the model to learn fault characteristics from various physical quantities. Even if the vibration signals do not show any anomalies at a particular time, changes in other physical quantities can be detected promptly, leading to accurate fault diagnosis. Compared to Experiment 1, the increased complexity of working conditions in Experiment 2 led to a decline in fault diagnosis accuracy across all comparison models. However, FedAdv_ID exhibited the smallest decline in accuracy, indicating that its adversarial training mechanism based on fault feature consistency measurement effectively mitigates data distribution discrepancies between clients caused by changes in working conditions. Furthermore, a comparison between FedRL_F1 and WAFedL demonstrates that merely adding a single model parameter regularization term can only eliminate inconsistencies between models, but it fails to fundamentally address the differences between fault features. This results in the continuous amplification of model discrepancies in subsequent federated processes. Moreover, while dynamic weighting can reduce the impact of low-quality clients on the global model when the data quality of clients varies, the differences between features still interfere with the generalization capability of the federated model (Table 11).
In addition, the advantage of the FedADV_ID method is further demonstrated by its unique consistency evaluation mechanism. By constructing evaluation metrics, this method can quantify the degree of deviation between local client-specific fault information and the average fault information at the federated center. Using the federated adversarial framework, FedADV_ID eliminates this information discrepancy through interactive adversarial training, ensuring that the global model remains consistent with the optimization directions of different clients, further improving the accuracy of fault diagnosis.
Experiments 3 and 4 investigated the performance of different comparative models under multiple working conditions where certain fault types were missing. In Experiment 3, compared to Experiment 1, the absence of specific fault types under the same conditions further exacerbated the statistical distribution heterogeneity. Due to the inherent feature drift in fault data, the DNN model exhibited significant negative transfer under these conditions, rendering it almost ineffective in both experimental scenarios. Additionally, the DNN model lacks the information fusion capabilities of federated learning, resulting in a trained model that fails to capture missing fault features. These two factors together significantly compromised the diagnostic accuracy of the DNN model (Table 12).
In contrast, FedAvg addressed the issue of missing faults through collaborative information complementation, leveraging the feature extraction capabilities of the global model to compensate for its shortcomings in diagnosing unobserved fault types. Compared to DNN, this method improved fault diagnosis accuracy by 12.25%, but the fault feature differences caused by varying working conditions still limited its diagnostic performance. FedRL_F1, compared to FedAvg, further enhanced the collaborative training of the global model by incorporating federated learning and mitigated model drift through parameter regularization, minimizing the bias between local and global models and reducing the adverse effects of model heterogeneity. Similarly, Fed_BN reduced feature scale differences across different clients by normalizing the features, aiming to mitigate the impact of heterogeneity. Therefore, although the strategies of FedRL_F1 and Fed_BN for diagnosing missing fault types still focus on optimizing the global model, they outperform FedAvg in terms of accuracy.
On the other hand, WAFedL constructed a globally optimal model through the federated center’s loss function to support local optimization across clients. However, its dynamic weighted aggregation approach only assigns weights to local models based on validation set performance without fully considering the impact of client data distribution differences on the global model, leading to a significant drop in diagnostic accuracy in Experiment 4 compared to Experiment 3.
The proposed FedADV_ID method eliminates client discrepancies caused by changes in working conditions at the fault information level by minimizing the difference between local fault information and average fault information. It can learn features of all known fault types in the federated center’s average fault information, compensating for the shortcomings in diagnosing missing fault types. This ensures that each client has the capability to diagnose fault types that have not been observed locally. Although the diagnostic results of the proposed method in Experiment 4 declined compared to other scenarios; its decline was the smallest, and its diagnostic accuracy was the highest among all methods, further demonstrating the superiority of the proposed method. Figure 10 visually showcases the accuracy advantage of the proposed method through experiments on a multi-channel dataset. By improving these methods, the proposed method not only ensures comprehensive diagnostic capability under different client environments but also significantly enhances overall fault diagnosis accuracy.
To further validate the superiority of the proposed algorithm across various experimental scenarios, the STAC statistical platform was employed to assess the experimental results under different experimental designs [44]. Under normal distribution testing, the results obtained by all algorithms within the same experimental scheme followed a normal distribution, aligning with actual fault diagnosis outcomes. Diverse fault diagnosis algorithms consistently yielded results displaying the classification label with the highest probability across different experimental designs, thus conforming to the normality of the results. Among numerous non-parametric test outcomes, FedAdv_ID exhibited the best overall performance and highest score compared to all other algorithms across various experimental settings. The consistent ranking of the remaining comparison algorithms with the actual experimental results further underscores the superiority of the proposed method relative to other approaches.

5. Conclusions and Future Work

FL has been widely adopted to address the challenges of insufficient fault samples within individual clients and the limited generalizability of fault models. However, due to the dynamic changes in motor operating conditions across different clients, significant disparities in the statistical distribution of fault data arise. These disparities limit the generalization capability of the aggregated global model and induce negative transfer phenomena in local clients, ultimately reducing fault diagnosis accuracy. This study proposes a novel method, FedAdv_ID, which begins by constructing a consistency evaluation metric to assess the degree of deviation of local fault information relative to the average fault information at the federated center. Subsequently, within an adversarial federated framework, interactive adversarial training between local clients and the federated center is employed to reduce information discrepancies and mitigate the drift in local fault information caused by statistical distribution heterogeneity. This process enhances the generalization capability of the global model. Finally, an information discrepancy-driven federated aggregation strategy is designed at the federated center to construct an optimal global model by taking into account the degree of information deviation across different clients, thereby maximizing the support for local model optimization. Experimental validation using both the CWRU single-channel dataset and the SMU multi-channel dataset demonstrates that the proposed method significantly outperforms existing approaches in terms of diagnostic accuracy.
The proposed method effectively addresses the challenge of ensuring fault diagnosis effectiveness under conditions of multiple operational modes and insufficient fault types. By utilizing benchmark and private datasets, four experimental scenarios with different operational conditions and insufficient fault types were designed to validate the effectiveness of the FedAdv_ID method. Although the proposed method demonstrates excellent performance in handling fault diagnosis across clients with varying fault information distributions due to differences in operational conditions or fault types, there are still some unresolved limitations. Specifically, when fault data collected by online monitoring equipment is subjected to network attacks, resulting in high noise levels, the inconsistency in fault information among clients increases significantly. In such cases, the proposed method lacks an effective mechanism to filter useful information from high-noise data, thereby impacting the model aggregation effect and the personalization strategy for updating the federated model. Future research could build on the current method by designing a personalized information enhancement mechanism to improve the model aggregation effect within the federated framework and enhance the processing of useful fault information.

Author Contributions

Methodology, F.Z., J.S., C.W. and X.H.; writing—original draft, J.S.; writing—review and editing, F.Z., J.S. and J.C.; investigation, T.W. All authors have read and agreed to the published version of this manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (62073213), the Shanghai Maritime University Graduate Student Training Program for Top Innovative Talents (2022YBR019), and the National Natural Science Foundation Youth Science Foundation Project (52205111).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data involved in this article have been presented in this article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kwiatoń, P.; Cekus, D.; Waryś, P. Computer-aided rotary crane stability assessment. Autom. Constr. 2024, 162, 105370. [Google Scholar] [CrossRef]
  2. Zhang, J.; Zhang, Q.; Qin, X.; Sun, Y. Robust fault diagnosis of quayside container crane gearbox based on 2D image representation in frequency domain and CNN. Struct. Health Monit. 2024, 23, 324–342. [Google Scholar] [CrossRef]
  3. Xie, F.; Li, G.; Hu, W.; Fan, Q.; Zhou, S. Intelligent Fault Diagnosis of Variable-Condition Motors Using a Dual-Mode Fusion Attention Residual. J. Mar. Sci. Eng. 2023, 11, 1385. [Google Scholar] [CrossRef]
  4. Liu, X.; Fan, Z.; Cao, Z.; Liu, Y.; Kang, Z.; Hu, Y. Fault signal simulation and service condition monitoring of cracked gear system. IEEE Trans. Instrum. Meas. 2023, 72, 1–14. [Google Scholar] [CrossRef]
  5. Xue, Y.; Wen, C.; Wang, Z.; Liu, W.; Chen, G. A novel framework for motor bearing fault diagnosis based on multi-transformation domain and multi-source data. Knowl.-Based Syst. 2024, 283, 111205. [Google Scholar] [CrossRef]
  6. Luo, P.; Yin, Z.; Yuan, D.; Gao, F.; Liu, J. An intelligent method for early motor bearing fault diagnosis based on Wasserstein distance generative adversarial networks meta learning. IEEE Trans. Instrum. Meas. 2023, 72, 1–11. [Google Scholar] [CrossRef]
  7. Qiao, Z.; Chen, S.; Lai, Z.; Zhou, S.; Sanjuán, M.A.F. Harmonic-Gaussian double-well potential stochastic resonance with its application to enhance weak fault characteristics of machinery. Nonlinear Dyn. 2023, 111, 7293–7307. [Google Scholar] [CrossRef]
  8. Qiao, Z.; Lei, Y.; Li, N. Applications of stochastic resonance to machinery fault detection: A review and tutorial. Mech. Syst. Signal Process. 2019, 122, 502–536. [Google Scholar] [CrossRef]
  9. Qiao, Z.; Shu, X. Coupled neurons with multi-objective optimization benefit incipient fault identification of machinery. Chaos Solitons Fractals 2021, 145, 110813. [Google Scholar] [CrossRef]
  10. Kumar, A.; Kumar, R.; Xiang, J.; Qiao, Z.; Zhou, Y.; Shao, H. Digital twin-assisted AI framework based on domain adaptation for bearing defect diagnosis in the centrifugal pump. Measurement 2024, 235, 115013. [Google Scholar] [CrossRef]
  11. Chen, J.; Li, J.; Huang, R.; Yue, K.; Chen, Z.; Li, W. Federated transfer learning for bearing fault diagnosis with discrepancy-based weighted federated averaging. IEEE Trans. Instrum. Meas. 2022, 71, 1–11. [Google Scholar] [CrossRef]
  12. Ma, J.; Jiang, X.; Han, B.; Wang, J.; Zhang, Z.; Bao, H. Dynamic simulation model-driven fault diagnosis method for bearing under missing fault-type samples. Appl. Sci. 2023, 13, 2857. [Google Scholar] [CrossRef]
  13. Zhou, F.; Liu, S.; Fujita, H.; Hu, X.; Zhang, Y.; Wang, B.; Wang, K. Fault diagnosis based on federated learning driven by dynamic expansion for model layers of imbalanced client. Expert Syst. Appl. 2024, 238, 121982. [Google Scholar] [CrossRef]
  14. Zhou, Y.; Fu, Z.; Zhang, J.; Li, W.; Gao, C. A digital twin-based operation status monitoring system for port cranes. Sensors 2022, 22, 3216. [Google Scholar] [CrossRef] [PubMed]
  15. Lin, S.; Zhen, L.; Wang, W. Planning low carbon oriented retrofit of diesel-driven cranes to electric-driven cranes in container yards. Comput. Ind. Eng. 2022, 173, 108681. [Google Scholar] [CrossRef]
  16. Li, Y.; Chen, Y.; Zhu, K.; Zhang, J. An effective federated learning verification strategy and its applications for fault diagnosis in industrial IOT systems. IEEE Internet Things J. 2022, 9, 16835–16849. [Google Scholar] [CrossRef]
  17. Geng, D.Q.; He, H.W.; Lan, X.C.; Liu, C. Bearing fault diagnosis based on improved federated learning algorithm. Computing 2022, 104, 1–19. [Google Scholar] [CrossRef]
  18. Yu, Y.; Guo, L.; Gao, H.; He, Y.; You, Z.; Duan, A. FedCAE: A new federated learning framework for edge-cloud collaboration based machine fault diagnosis. IEEE Trans. Ind. Electron. 2023, 71, 4108–4119. [Google Scholar] [CrossRef]
  19. Liu, Q.; Yang, B.; Wang, Z.; Zhu, D.; Wang, X.; Ma, K.; Guan, X. Asynchronous decentralized federated learning for collaborative fault diagnosis of PV stations. IEEE Trans. Netw. Sci. Eng. 2022, 9, 1680–1696. [Google Scholar] [CrossRef]
  20. Lu, S.; Gao, Z.; Xu, Q.; Jiang, C.; Zhang, A.; Wang, X. Class-imbalance privacy-preserving federated learning for decentralized fault diagnosis with biometric authentication. IEEE Trans. Ind. Inform. 2022, 18, 9101–9111. [Google Scholar] [CrossRef]
  21. Liu, J.; Wang, J.H.; Rong, C.; Xu, Y.; Yu, T.; Wang, J. Fedpa: An adaptively partial model aggregation strategy in federated learning. Comput. Netw. 2021, 199, 108468. [Google Scholar] [CrossRef]
  22. Wang, R.; Huang, W.; Zhang, X.; Wang, J.; Ding, C.; Shen, C. Federated contrastive prototype learning: An efficient collaborative fault diagnosis method with data privacy. Knowl.-Based Syst. 2023, 281, 111093. [Google Scholar] [CrossRef]
  23. Mehta, M.; Chen, S.; Tang, H.; Shao, C. A federated learning approach to mixed fault diagnosis in rotating machinery. J. Manuf. Syst. 2023, 68, 687–694. [Google Scholar] [CrossRef]
  24. Guo, W.; Wang, Y.; Chen, X.; Jiang, P. Federated transfer learning for auxiliary classifier generative adversarial networks: Framework and industrial application. J. Intell. Manuf. 2024, 35, 1439–1454. [Google Scholar] [CrossRef]
  25. Feng, S.; Li, B.; Yu, H.; Liu, Y.; Yang, Q. Semi-supervised federated heterogeneous transfer learning. Knowl.-Based Syst. 2022, 252, 109384. [Google Scholar] [CrossRef]
  26. Ye, M.; Fang, X.; Du, B.; Yuen, P.C.; Tao, D. Heterogeneous federated learning: State-of-the-art and research challenges. ACM Comput. Surv. 2023, 56, 1–44. [Google Scholar] [CrossRef]
  27. Yao, Z.; Song, P.; Zhao, C. Finding trustworthy neighbors: Graph aided federated learning for few-shot industrial fault diagnosis with data heterogeneity. J. Process Control 2023, 129, 103038. [Google Scholar] [CrossRef]
  28. Li, L.; Fan, Y.; Tse, M.; Lin, K.-Y. A review of applications in federated learning. Comput. Ind. Eng. 2020, 149, 106854. [Google Scholar] [CrossRef]
  29. Lu, H.; Thelen, A.; Fink, O.; Hu, C.; Laflamme, S. Federated learning with uncertainty-based client clustering for fleet-wide fault diagnosis. Mech. Syst. Signal Process. 2024, 210, 111068. [Google Scholar] [CrossRef]
  30. Zhang, Z.; Zhou, F.; Wang, C.; Wen, C.; Hu, X.; Wang, T. A Multiscale Recursive Attention Gate Federation Method for Multiple Working Conditions Fault Diagnosis. Entropy 2023, 25, 1165. [Google Scholar] [CrossRef] [PubMed]
  31. Chen, H.; Wang, X.B.; Yang, Z.X.; Li, J. Privacy-preserving intelligent fault diagnostics for wind turbine clusters using federated stacked capsule autoencoder. Expert Syst. Appl. 2024, 254, 124256. [Google Scholar] [CrossRef]
  32. Qin, N.; Du, J.; Zhang, Y.; Huang, D.; Wu, B. Fault diagnosis of multi-railway high-speed train bogies by improved federated learning. IEEE Trans. Veh. Technol. 2023, 72, 7184–7194. [Google Scholar] [CrossRef]
  33. Du, J.; Qin, N.; Huang, D.; Zhang, Y.; Jia, X. An efficient federated learning framework for machinery fault diagnosis with improved model aggregation and local model training. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 10086–10097. [Google Scholar] [CrossRef] [PubMed]
  34. Shi, Y.; Zhang, Y.; Xiao, Y.; Niu, L. Federated learning with L1 regularization. Pattern Recognit. Lett. 2023, 172, 15–21. [Google Scholar] [CrossRef]
  35. Li, X.; Jiang, M.; Zhang, X.; Kamp, M.; Dou, Q. Fedbn: Federated learning on non-iid features via local batch normalization. arXiv 2021, arXiv:2102.07623. [Google Scholar]
  36. Yoon, T.; Shin, S.; Hwang, S.J.; Yang, E. Fedmix: Approximation of mixup under mean augmented federated learning. arXiv 2021, arXiv:2107.0023. [Google Scholar]
  37. Wang, J.; Liu, Q.; Liang, H.; Joshi, G.; Poor, H.V. Tackling the objective inconsistency problem in heterogeneous federated optimization. Adv. Neural Inf. Process. Syst. 2020, 33, 7611–7623. [Google Scholar]
  38. Chen, J.; Li, J.; Huang, R.; Yue, K.; Chen, Z.; Li, W. Federated learning for bearing fault diagnosis with dynamic weighted averaging. In Proceedings of the 2021 International Conference on Sensing, Measurement & Data Analytics in the era of Artificial Intelligence, IEEE, Nanjing, China, 21–23 October 2021; pp. 1–6. [Google Scholar]
  39. Chen, K.; Zhang, X.; Zhou, X.; Mi, B.; Xiao, Y.; Zhou, L.; Wu, Z.; Wu, L.; Wang, X. Privacy preserving federated learning for full heterogeneity. ISA Trans. 2023, 141, 73–83. [Google Scholar] [CrossRef]
  40. Wang, R.; Yan, F.; Yu, L.; Shen, C.; Hu, X.; Chen, J. A federated transfer learning method with low-quality knowledge filtering and dynamic model aggregation for rolling bearing fault diagnosis. Mech. Syst. Signal Process. 2023, 198, 110413. [Google Scholar] [CrossRef]
  41. Bisiacco, M.; Pillonetto, G. A refinement of the stability test for reproducing kernel Hilbert spaces. Automatica 2024, 163, 111606. [Google Scholar] [CrossRef]
  42. Ye, M.; Han, Q.-L.; Ding, L.; Xu, S. Distributed Nash equilibrium seeking in games with partial decision information: A survey. Proc. IEEE 2023, 111, 140–157. [Google Scholar] [CrossRef]
  43. Case Western Reserve University Bearing Data Center Website. Available online: https://engineering.case.edu/bearingdatacenter (accessed on 10 July 2024).
  44. Rodríguez-Fdez, I.; Canosa, A.; Mucientes, M.; Bugarín, A. STAC: A web platform for the comparison of algorithms using statistical tests. In Proceedings of the 2015 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), Istanbul, Turkey, 2–5 August 2015; pp. 1–8. [Google Scholar]
Figure 1. Schematic diagram of the federated adversarial method driven by information discrepancy.
Figure 1. Schematic diagram of the federated adversarial method driven by information discrepancy.
Entropy 26 00718 g001
Figure 2. Eliminate discrepancy between fault information.
Figure 2. Eliminate discrepancy between fault information.
Entropy 26 00718 g002
Figure 3. Flowchart of the FedAdv_ID method.
Figure 3. Flowchart of the FedAdv_ID method.
Entropy 26 00718 g003
Figure 4. The CWRU bearing experiment bench [43].
Figure 4. The CWRU bearing experiment bench [43].
Entropy 26 00718 g004
Figure 5. Vibration signal waveforms of ten types of bearing data.
Figure 5. Vibration signal waveforms of ten types of bearing data.
Entropy 26 00718 g005
Figure 6. Statistical distribution of failure data under different working conditions. (a) Statistical distribution of fault data in experiment 1; (b) Statistical distribution of fault data in experiment 2.
Figure 6. Statistical distribution of failure data under different working conditions. (a) Statistical distribution of fault data in experiment 1; (b) Statistical distribution of fault data in experiment 2.
Entropy 26 00718 g006
Figure 7. Confusion matrix of each method for Experiment 1 multiple working scenarios.
Figure 7. Confusion matrix of each method for Experiment 1 multiple working scenarios.
Entropy 26 00718 g007aEntropy 26 00718 g007b
Figure 8. SMU motor fault test bench.
Figure 8. SMU motor fault test bench.
Entropy 26 00718 g008
Figure 9. Time–domain waveform of the normal signal of SMU motor fault experiment platform.
Figure 9. Time–domain waveform of the normal signal of SMU motor fault experiment platform.
Entropy 26 00718 g009
Figure 10. Experimental accuracy of SMU motor fault diagnosis.
Figure 10. Experimental accuracy of SMU motor fault diagnosis.
Entropy 26 00718 g010
Table 1. Fault types and corresponding labels.
Table 1. Fault types and corresponding labels.
Fault TypeFault Size (Inch)Label
Ball fault0.0070
Inner race fault0.0071
Outer race fault0.0072
Normal03
Table 2. Experimental scenario design for the CWRU dataset.
Table 2. Experimental scenario design for the CWRU dataset.
Experimental ScenariosClientsMotor Working ConditionsFault TypeQuantity of Label Data
Experiment 1client 10 HP/1 HP0/1/2/34 × 120
client 21 HP/2 HP0/1/2/34 × 120
client 32 HP/3 HP0/1/2/34 × 120
Experiment 2client 10 HP/1 HP/2 HP0/1/2/34 × 120
client 21 HP/2 HP/3 HP0/1/2/34 × 120
client 30 HP/2 HP/3 HP0/1/2/34 × 120
Experiment 3client 10 HP/1 HP0/1/24 × 120
client 21 HP/2 HP0/2/34 × 120
client 32 HP/3 HP1/2/34 × 120
Experiment 4client 10 HP/1 HP/2 HP0/1/24 × 120
client 21 HP/2 HP/3 HP0/2/34 × 120
client 30 HP/2 HP/3 HP1/2/34 × 120
Table 3. Comparison methods and their descriptions.
Table 3. Comparison methods and their descriptions.
Comparison MethodsRelated Description
DNNTraditional DL fault diagnosis methods
FedAvg [14]Traditional FL approach
FedRL_F1 [29]A FL approach based on F1 regularization
Fed_BN [27]FL approach to incorporate BN layers in fault diagnosis model
WAFedL [32]A FL approach to adaptive weight aggregation
FedAdv_IDAn information discrepancy-driven federated adversarial learning approach
Table 4. Model structure and hyperparameters.
Table 4. Model structure and hyperparameters.
Hidden Layers and NeuronsLearning RateBatch SizeOptimizerTraining Stopping Threshold
100/600/400/100/40.001128Adam 10 - 4
Table 5. Comparative fault diagnosis results for experiments 1 and 2.
Table 5. Comparative fault diagnosis results for experiments 1 and 2.
ExperimentsClientsDNNFedAvgFedRL_F1Fed_BNWAFedLFedAdv_ID
Experiment 1Client 155.20%61.66%73.94%76.71%83.53%93.45%
Client 253.59%61.74%75.79%74.88%81.35%93.14%
Client 352.61%61.22%75.87%73.11%83.82%92.69%
Average53.80%61.54%75.20%74.90%82.90%93.09%
Experiment 2Client 143.61%52.62%62.67%63.01%73.94%86.76%
Client 245.27%55.67%62.42%63.95%75.79%85.93%
Client 346.35%53.72%61.02%61.78%75.87%85.51%
Average45.07%54.00%62.03%62.91%75.20%86.06%
Table 6. Fault diagnosis results for Experiments 3 and 4.
Table 6. Fault diagnosis results for Experiments 3 and 4.
ExperimentsClientsDNNFedAvgFedRL_F1Fed_BNWAFedLFedAdv_ID
Experiment 3Client 139.51%46.44%56.84%59.22%67.29%82.51%
Client 239.65%46.44%56.67%59.24%66.99%81.77%
Client 339.52%45.88%56.08%58.35%66.49%81.56%
Average39.56%46.25%56.53%58.93%66.92%81.94%
Experiment 4Client 133.99%40.80%55.74%54.08%62.66%77.17%
Client 233.88%40.80%55.71%54.14%62.62%77.22%
Client 333.63%40.53%54.81%53.29%62.19%76.09%
Average33.83%40.71%55.42%53.83%62.49%76.82%
Table 7. SMMU Motor Data Fault Type Description.
Table 7. SMMU Motor Data Fault Type Description.
Fault TypeLoad (N/m)Sampling FrequencyFault Degree
Normal0/4/6/1025,6000
Inner race fault0.5 mm
Outer race fault0.3 mm
Phase faultLack of V-phase
Winding short circuit faultWinding short circuit 10%
Winding short circuit faultWinding short circuit 5%
Shaft bending fault0.3 mm
Rotor unbalance fault4 g
Parallel misalignment fault0.25 mm
Broken bar faultBroken two bars
Table 8. Experimental fault types and corresponding labels.
Table 8. Experimental fault types and corresponding labels.
Fault TypeFault DegreeLabel
Fan end bearing inner race fault0.05 mm0
Drive end bearing outer race fault0.05 mm1
Bent shaft0.1 mm2
Broken bar faulttwo3
Table 9. Design of complex experimental scenarios.
Table 9. Design of complex experimental scenarios.
ExperimentsClientsWorking Conditions (N·m)Fault TypeQuantity of Labeled Data
Experiment 1Client 10/60/1/2/34 × 100
Client 20/100/1/2/34 × 100
Client 36/100/1/2/34 × 100
Experiment 2Client 10/4/60/1/2/34 × 100
Client 24/6/100/1/2/34 × 100
Client 30/6/100/1/2/34 × 100
Experiment 3Client 10/60/1/24 × 100
Client 20/100/1/34 × 100
Client 36/101/2/34 × 100
Experiment 4Client 10/4/60/1/24 × 100
Client 24/6/100/1/34 × 100
Client 30/6/101/2/34 × 100
Table 10. Structure and hyperparameters of all models.
Table 10. Structure and hyperparameters of all models.
Hidden Layers and NeuronsLearning RateBatch SizeOptimizerTraining Stopping Thresholds
100/1200/600/400/40.001128Adam 10 4
Table 11. Client diagnostic accuracy of different methods in Experiments 1 and 2.
Table 11. Client diagnostic accuracy of different methods in Experiments 1 and 2.
ExperimentsClientsDNNFedAvgFedRL_F1Fed_BNWAFedLFedAdv_ID
Experiment 1Client 154.21%61.50%71.20%73.66%84.79%92.13%
Client 254.12%61.58%71.20%72.95%84.87%92.14%
Client 353.23%60.62%70.26%72.15%83.37%90.62%
Average53.85%61.54%70.88%72.92%84.34%91.63%
Experiment 2Client 141.43%49.15%61.53%65.37%73.97%85.83%
Client 241.87%48.93%60.86%66.05%74.62%86.47%
Client 342.56%51.24%62.31%65.79%75.08%85.12%
Average41.95%49.77%61.56%65.74%74.55%85.81%
Table 12. Client diagnostic accuracy of different methods in Experiments 3 and 4.
Table 12. Client diagnostic accuracy of different methods in Experiments 3 and 4.
ExperimentsClientsDNNFedAvgFedRL_F1Fed_BNWAFedLFedAdv_ID
Experiment 3Client 143.29%55.53%64.90%62.78%75.45%83.39%
Client 243.28%55.70%65.03%62.70%75.41%83.49%
Client 342.75%54.83%64.01%61.58%74.09%82.09%
Average43.10%55.35%64.64%62.35%74.98%82.99%
Experiment 4Client 132.47%42.61%51.55%55.81%63.76%79.49%
Client 233.36%41.85%50.91%56.27%64.25%78.22%
Client 332.24%42.17%52.74%56.38%63.71%77.65%
Average32.69%42.21%51.76%56.15%63.91%78.45%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, J.; Zhou, F.; Chen, J.; Wang, C.; Hu, X.; Wang, T. A Federated Adversarial Fault Diagnosis Method Driven by Fault Information Discrepancy. Entropy 2024, 26, 718. https://doi.org/10.3390/e26090718

AMA Style

Sun J, Zhou F, Chen J, Wang C, Hu X, Wang T. A Federated Adversarial Fault Diagnosis Method Driven by Fault Information Discrepancy. Entropy. 2024; 26(9):718. https://doi.org/10.3390/e26090718

Chicago/Turabian Style

Sun, Jiechen, Funa Zhou, Jie Chen, Chaoge Wang, Xiong Hu, and Tianzhen Wang. 2024. "A Federated Adversarial Fault Diagnosis Method Driven by Fault Information Discrepancy" Entropy 26, no. 9: 718. https://doi.org/10.3390/e26090718

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop