Next Article in Journal
STID-Net: Optimizing Intrusion Detection in IoT with Gradient Descent
Previous Article in Journal
A Rapid Deployment Method for Real-Time Water Surface Elevation Measurement
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Domain Adversarial Transfer Learning Bearing Fault Diagnosis Model Incorporating Structural Adjustment Modules

School of Mechanical and Electrical Engineering, Henan University of Science and Technology, Luoyang 471023, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(6), 1851; https://doi.org/10.3390/s25061851
Submission received: 19 February 2025 / Revised: 9 March 2025 / Accepted: 12 March 2025 / Published: 17 March 2025
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

:
With the improvement in industrial equipment intelligence and reliability requirements, bearing fault diagnosis has become a key technology to ensure the stable operation of mechanical equipment. Traditional bearing fault diagnosis methods are ineffective in diagnosing complex faults and mostly rely on the manual adjustment of hyperparameters. To this end, this paper proposes a domain adversarial migratory learning bearing fault diagnosis model incorporating structural adjustment modules. First, the pre-trained model of the source domain is applied to the target domain dataset through an adversarial domain adaptation technique. Then, the network depth and width are dynamically adjusted in the Optuna optimization framework to accommodate more complex fault types in the target domain. Finally, the performance of the model is further improved by automatically optimizing the hyperparameters. The experimental results show that the model exhibits high accuracy in the diagnosis of different fault types, especially in the face of complex and variable industrial environments, demonstrating strong adaptability and robustness. The method provides an effective solution for fault diagnosis of intelligent devices.

1. Introduction

Under the development trend of intelligent and highly reliable industrial equipment, bearing fault diagnosis technology has become a key link to ensure the stable operation of modern manufacturing systems. As the core supporting components of rotating machinery, minor defects in bearings are very likely to cause chain failures or even major safety accidents under high-speed and heavy-duty working conditions, and according to statistics, about 45% of mechanical failures in industrial scenarios can be traced back to bearing anomalies, so bearing fault diagnosis has always been a hotspot for research.
With the increasing maturity of machine learning technology, technicians have proposed a series of intelligent diagnostic methods [1,2,3], which significantly improve the rolling bearing fault diagnosis capability. Lee and Sabater proposed the conversion of time series data into visibility graphs from which persistent homology is applied to extract topological features based on topological data analysis (TDA) [4]. Shinde et al. proposed the basic bearing frequencies withdrawn from the vibration response as novel extracted features and provided the data to a K nearest neighbor network (KNN) for fault classification [5]. Phan et al. proposed to apply EMD and EEMD decomposition on the signals and then combine the features extracted from the preprocessed and decomposed signals to select the information function through BGWO and input the selected function into the classifier, and experiments proved that this method can lead to a significant improvement in ML classification accuracy [6]. However, ML still faces some problems when used for bearing fault diagnosis: the performance of the model relies on a large amount of labeled data, and existing methods may suffer from insufficient generalization ability when dealing with complex conditions or new fault types.
In recent years, deep learning has been introduced into the field of bearing fault diagnosis, and a series of results have been achieved [7,8,9]. Shan et al. proposed a Mel-CNN model [10] for the fault diagnosis of noisy motor bearings and achieved more accurate and reliable diagnostic results compared with the previous model. The Mel-CNN model utilizes variational mode decomposition (VMD) to remove the high-frequency components of the motor noise, extracts Mel spectral acoustic features, and re-extracts the Mel acoustic features with the help of convolutional neural network (CNN). Li et al. proposed to find the optimal hyperparameter combination of long-short-term memory (LSTM) network based on SOA [11]. Zhou et al. proposed a fault diagnosis framework based on evidence theory and the improved visual set group (VGG) neural network (EVGG) to obtain accurate and reliable diagnostic results with additional estimation of the prediction uncertainty to obtain accurate and reliable diagnostic results [12]. Chen and Zou [13] trained separate convolutional neural network (CNN) and long short-term memory network (LSTM) models and implemented three different fusion strategies. However, the application of deep learning in the field of bearing fault diagnosis still faces many challenges [14,15,16]: the performance of deep learning models is highly dependent on hyperparameters, which requires professional technicians to spend a lot of time on debugging; the generalization ability of deep learning models is limited; and the deep learning models require that the test data be distributed in the same way as the training data.
In order to lower the threshold of machine learning [17] and shorten the model development cycle, automated machine learning (AutoML) has become a hot research topic in recent years [18,19,20]. There are already many mature AutoML platforms and tools, like Google AutoML, Auto-sklearn, H2O.ai, TPOT, etc. A number of new tools are being developed as research progresses. Feurer et al. proposed Auto-sklearn 2.0 [21], an automatic learning system based on Auto-sklearn 1.0, which enables AutoML to handle large datasets well by using a new, meta-featureless meta-learning technique with a successful bandit strategy for budget allocation. Neutatz et al. proposed Caml [22], a model that uses meta-learning to automatically adjust its AutoML parameters to accomplish the appropriate task. However, AutoML faces many challenges: AutoML for Neural Architecture Search (NAS) and hyperparameter optimization usually requires significant computational resources [23], especially when dealing with complex deep learning tasks [24]; when the dataset is not large enough or the features are not well chosen [25], it is difficult for AutoML to guarantee the model’s ability to generalize to new datasets [26]; and AutoML is susceptible to local optimal solutions in the optimization process.
With the development of transfer learning [27,28,29,30], there is hope that the problems associated with deep learning can be solved. Transfer learning shortens the model development cycle and improves the learning efficiency and performance of new tasks by transferring relevant knowledge from previously trained tasks to related tasks. Wu et al. proposed the use of joint distribution adaptation to reduce the difference in probability distributions between the auxiliary dataset and the target domain dataset and introduced the GWO algorithm for adaptive learning of key parameters of the model [31]. Huo et al. proposed an enhanced transfer learning method based on a linear superposition network [32] for rolling bearing fault diagnosis, which improves the structure of a one-dimensional convolutional neural network (1D-CNN) by constructing linear superposition of convolutional blocks, which enhances the ability of the model to extract fault features. Thuan proposed a robust transfer learning based on a multi-layer maximum mean difference loss function [33], and the model achieved better results in cross-machine scenarios. Zhang et al. proposed a fault diagnosis method based on an image information fusion and visual transform (ViT) transfer learning model [34]. Despite the good results of transfer learning in the field of bearing fault diagnosis, there are still some issues to be aware of: existing public bearing fault datasets are not as large as datasets such as ImageNet, The Pile, etc., which are small and medium-sized datasets; however, transfer learning cannot provide ideal diagnosis with existing models and learning strategies in cases where the source domain dataset is not very large or the source domain dataset is smaller than the target domain dataset.
The above issues are summarized below:
(1)
There are no publicly available large bearing failure datasets that can be used to train neural networks, leading to poor performance of transfer learning in solving multi-failure complex problems.
(2)
Deep learning to find the optimal hyperparameters for a model takes a lot of developer time and effort.
(3)
Existing AutoML platforms and tools generally require significant computational resources when performing Neural Architecture Search (NAS).
In this paper, we propose an automated transfer learning bearing fault diagnosis model for multi-fault complexity. First, transfer learning is utilized to apply a well-trained model in the source domain ( D s ) to the task in the target domain ( D t ), and the distribution difference between the source and target domains is reduced by domain adaptation. Then, a neural network depth and width dynamic adjustment module is embedded in Optuna [35], which not only dynamically freezes and unfreezes the layers of the neural network but also chooses whether to increase and how to increase the depth and width of the neural network within a certain range to adapt to larger data sizes and more complex fault types in the target domain through Optuna’s built-in optimization algorithm. Finally, model hyperparameters including learning rate, number of training rounds, task loss weights, and domain loss weights are optimized by Optuna to further improve the model performance. The innovation of the proposed method is that Optuna automatically optimizes the depth and width of the neural network to adapt to the complexity of the target domain while using ADA to reduce the difference in the distribution of the source and target domains. This approach not only improves the generalization ability of the model across different fault types and complex environments but also reduces the need for manual parameter tuning and improves the accuracy and efficiency of fault diagnosis. The contributions of this paper are summarized as follows:
(1)
Applying pre-trained models to the target domain using transfer learning avoids the use of AutoML for Neural Architecture Search (NAS), reduces the demand for computational resources during model development, and significantly shortens the development cycle relative to building models from scratch.
(2)
Compared with the fine-tuning [36] strategy commonly used in transfer learning, this paper proposes the method of embedding the neural network depth and width dynamic adjustment module, which largely improves the model’s ability to classify complex faults and effectively copes with the current lack of large public bearing fault datasets.
(3)
The introduction of Optuna to optimize the hyperparameters of the model no longer requires professionals to spend a lot of time on manual debugging, which lowers the threshold of model development and use and is conducive to the promotion of “intelligence” in various industries.

2. Theoretical Foundation

2.1. Adversarial Domain Adaptation

Adversarial domain adaptation (ADA) [37] and generative adversarial networks (GANs) [38] share similar adversarial ideas. The main goal of ADA is to reduce the distributional differences between the source and target domains, allowing models trained in the source domain to migrate to the target domain. When ADA is applied to transfer learning and domain adaptation tasks, the aim is for the model to achieve better adaptation between the source and target domains. The main goal of GANs is to generate realistic data samples, enabling the generator to produce samples that are similar to the real data distribution, while the discriminator’s task is to determine whether the input samples are real or not.

2.1.1. Basic Idea

Adversarial domain adaptation is inspired by generative adversarial networks (GANs), and its core idea is to introduce an “adversarial” process into the training process. Specifically, it includes the following:
Feature extractor: The role of the feature extractor is to extract useful features from the input data. In ADA, the feature extractor is usually a deep neural network that is used to extract generic features of the source and target domain data.
The discriminator aims to determine whether the input features come from the source domain or the target domain by training the discriminator to help the generator learn a more generalized feature representation.
The goal of adversarial training is to obtain a domain-invariant feature representation for the generator by optimizing the game process of the generator and the discriminator, allowing the model to generalize better to the target domain.

2.1.2. Processes for Adversarial Domain Adaptation

Adversarial domain adaptation typically involves the following major steps:
  • The feature extractor extracts features from the source and target domain data, usually implemented using deep neural networks (e.g., convolutional neural networks, fully connected networks, etc.).
  • Discriminator: the task of the discriminator is to determine whether the input features are from the source or target domain. Ideally, after training, the discriminator will not be able to distinguish between features from the source and target domains.
  • The classifier is responsible for classification tasks based on extracted features (e.g., image classification, text classification, etc.).
  • Adversarial training aims to optimize the generator and discriminator by backpropagation. The goal of the discriminator is to differentiate as much as possible between data in the source and target domains, while the goal of the feature extractor and classifier is to try to make the discrimination as ineffective as possible by generating an adversarial process.

2.1.3. Principles and Formulas: An Example of Domain Adversarial Neural Network (DANN)

  • Basic Settings and Symbol Definitions
D s = x i s , y i s i = 1 n s is the source domain dataset, where x i s is the input data from which it originated and y i s is the label of the source domain.
D t = x i t i = 1 n t is the target domain dataset, where x i t is the input data for the target domain, which is unlabeled.
F is the feature extractor that maps the input x to the shared feature space H : F x .
C is the classifier for a feature-based F x for a source domain classification task.
D is an adversarial discriminator that is used to determine whether the features of a sample come from the source or target domain.
2.
Categorized losses
In the source domain, the goal is to train a classifier C , which maps the inputs from the source domain to the category labels y . The categorization loss is usually measured using Cross-Entropy Loss to measure the performance of the source domain classification.
L c = 1 n s i = 1 n s log P y i s | F x i s
where P y i s | F x i s is the predicted probability of classification of the classifier C on the source domain sample x i s .
In conjunction with this paper, an increase in the depth of the neural network improves the feature extraction capability and allows the model to learn more complex features, thereby improving classification accuracy and reducing classification loss. An increase in the width of the neural network improves the feature representation per layer, allowing the model to handle more features, improving classification performance, and reducing classification loss. Considering the effect of depth and width adjustments on the classification loss, a regularization term is introduced into the loss function to penalize overly complex model structures and thus avoid overfitting, and the new classification loss function is
L c = L c + λ r e g · R ( L n e w , L l n e w )
where λ r e g is the regularization factor and R ( L n e w , L l n e w ) is a regularization term resulting from depth and width adjustments to limit the network complexity. This can improve the classification accuracy while avoiding the overfitting problem caused by an overly complex network.
3.
Adversarial loss
The core of adversarial domain adaptation is to make features from the source and target domains as indistinguishable as possible in the shared feature space. To this end, we introduce an adversarial discriminator D , whose task is to determine whether a given feature comes from the source domain. Adversarial loss is often used to measure the performance of the discriminator D using the logarithmic loss of the discriminator:
L a D = 1 n s n = 1 n s log D F x i s 1 n t n = 1 n t log 1 D F x i t
where D F x i s is the probability that the discriminator D predicts that the source domain sample feature F x i s comes from the source domain and D F x i t is the probability that the discriminator D predicts that the target domain sample feature F x i t comes from the target domain. The discriminator D tries to maximize this loss so that it accurately distinguishes between samples from the target and source domains.
For the method proposed in this paper, the increase in the depth of the neural network helps the generator to learn richer domain invariant features, which makes the data features in the source and target domains more indistinguishable and thus effectively reduces the adversarial loss. The increase in the width of the neural network enhances the feature representation of the generator and discriminator, allowing the generator to learn more domain-invariant features, thus reducing adversarial loss. At the same time, the increase in width also enhances the discriminator’s discriminative ability, making it more difficult to distinguish between features in the source and target domains and helping to improve the model’s ability to migrate across domains. The adjusted adversarial loss function can be expressed as
L a = i = 1 N s log ( D ( x i ; L n e w , N l n e w ) ) i = 1 N t log ( x j ; L n e w , N l n e w )
where D ( x i ; L n e w , N l n e w ) is the output of the discriminator for either the source domain sample or the target domain, dependent on the adjusted depths L n e w and N l n e w .
4.
Objectives of Confrontation Training
In adversarial training, the feature extractor and the discriminator are trained in a game. The goal is for the feature extractor to learn a domain-invariant feature representation that makes it impossible for the discriminator to distinguish between samples in the source and target domains. The loss function of the feature extractor is the negative of the adversarial loss; we want to maximize the discriminator’s error. Therefore, the function of the feature extractor is
L a F = L a D
5.
Joint optimization
The training objective of DANN [39] is to optimize both classification loss and adversarial loss with a total loss function:
L t o t a l = L c + λ L a F
where λ is the hyperparameter that balances the classification loss and the adversarial loss. Classification loss ( L c ) is used to ensure that the classification task on the source domain is learned correctly. Adversarial loss ( L a ) optimizes the feature extractor and discriminator by backpropagation, making it difficult to distinguish between features of the source and target domains in the shared space. The basic DANN architecture is shown in Figure 1.

3. Proposed Methodology

3.1. Adversarial Domain Adaptive Transfer Learning

Adversarial domain adaptation (ADA) is a transfer learning method based on the idea of adversarial training in the field of deep learning. Its main purpose is to solve the problem of distributional differences between the source domain and the target domain, thus enabling the model to migrate the knowledge learned from the source domain to the target domain.

3.1.1. Source Domain Model Selection and Training

Convolutional neural networks (CNNs) [40] are well suited for processing time-series signals because they are effective in extracting useful features from localized regions. The basic CNN architecture is shown in Figure 2. In bearing fault classification, a CNN can automatically extract time-frequency features in the signal without manual feature engineering. The steps for building and training the pre-trained model are as follows.
  • Data preprocessing
Vibration signals from bearings are usually time series data. These data need to be preprocessed appropriately before using CNNs, The data preprocessing flowchart is shown in Figure 3:
Signal denoising: the vibration signal may contain noise from the environment, so it is necessary to use a filter (low-pass filter) to remove high-frequency noise, assuming that the original signal is x t and the filtered signal x t can be calculated by convolution with the filter h t :
x t = x t h t
Signal segmentation: the original vibration signal is segmented, and each segment is used as a sample for model training.
Feature scaling: normalization of the signal is required to ensure consistent scaling of each input feature.
x i = x i μ σ
where x i is a sample from the input signal matrix, μ is the sample mean, σ is the standard deviation, and x i is the normalized sample.
2.
Building CNN models
Convolutional layer: a convolutional layer is used to extract localized features from the input data. When dealing with one-dimensional signals, the size of the convolution kernel is taken to be a smaller value, which helps in capturing the local features of the signal. In the case of one-dimensional convolution, the formula for convolution operation is
y t = k = 0 k 1 x t k · ω k
where x t is the input signal, ω t is the convolution kernel, and y t is the convolved output.
Activation layer: the activation function is a ReLU (Rectified Linear Unit), which increases the nonlinear capability of the model, thus improving the classification performance, and is defined as
Re L U x = max 0 , x
Pooling layer: the pooling layer is usually used to reduce the size of the feature map while retaining important information. When dealing with 1D vibration signals, this paper uses 1D Max Pooling:
y t = max x t , x t + 1 , , x t + k 1
where k is the size of the pooling window, x t is the input signal, and y t is the pooled input.
Fully connected layer: in the last layers of the network, the features extracted through the convolution and pooling layers are spread into a one-dimensional vector and passed into the fully connected layer for classification decisions:
y = W x + b
where x is the input vector, W is the weight matrix, b is the bias term, and y is the output.
Output layer: the output layer uses the Softmax activation function, which, for a multicategorization problem, is capable of outputting the probability of each category:
P y i = e z i j e z j
where z i is the score of the first i category and P y i is the probability of that category.
3.
Model training
In order to train the CNN model, we need to use a labeled vibration signal dataset. The training process includes the following steps:
Define loss function: the loss function used in this paper is Cross-Entropy Loss, which is used to calculate the difference between the predicted and actual categories:
L y , y = i y i log y i
where y i is the probability distribution of the actual categories and y i is the probability distribution of the predicted categories.
Optimizer: the optimizer used in this paper is Adam, which can adaptively adjust the learning rate to improve the training efficiency. the updated formula of the Adam optimization algorithm is
θ t = θ t 1 η · m t v t + ε
where θ t is the parameter at step t ; η is the learning rate; m t and v t are the first-order and second-order moment estimates of the gradient, respectively; and ε is a constant that prevents division by zero.
Training process: the training data are input into the CNN, and the weights of the model are optimized after several rounds of iterations until the loss function converges.
4.
Evaluation and Testing of Models
After training the model, it needs to be evaluated. The evaluation metrics chosen in this paper are accuracy and F1 score.

3.1.2. Domain Adaptive Migration of Pre-Trained Models

Developing a new model from scratch often requires a long lead time, but there already exist many models that have been designed and debugged with a lot of effort by professionals, and these models are able to perform well on the corresponding datasets. Transfer learning is able to apply the mature models to the relevant tasks. The adversarial training process is as follows:
  • Pre-training dissemination
Input data: each batch of data in the source domain training set is fed into the model for forward propagation.
Feature extraction: the data are passed through the feature extraction part of the model to generate a shared feature representation f , and useful features are extracted from the input data:
f = ϕ x
where ϕ x is the feature extractor function.
Task classification: the output of the feature extractor is passed to the task classifier for classification prediction. The task classifier calculates the classification loss and is trained on the target label.
Domain classification: the features inverted by the GradientReverseLayer are fed into the DomainClassifier for domain discrimination. The DomainClassifier tries to determine whether the features come from the source or target domain, and the task is to identify the source (source or target) of these features.
2.
Calculating the loss function
Task classification loss: task classification loss is computed using CrossEntropyLoss, which measures the difference between the classifier output and the true label.
Adversarial loss: adversarial loss is computed through DomainClassifier with the goal of minimizing the distinction between source and target domain features. With the GradientReverseLayer (GRL), the feature extractor adjusts the parameters so that the domain discriminator cannot distinguish between source and target domain data.
Total loss function: the final total loss function is the weighted sum of the task classification loss and the adversarial loss.
3.
Backpropagation and Optimization
Backpropagation: by calculating the gradient of task loss and adversarial loss and updating the weights of the model using the optimizer, the gradient update formula is
θ t + 1 = θ t η · θ L t o t a l
where θ t is the parameter, η is the learning rate, and θ L t o t a l is the gradient of the total loss function.
Optimizer: the optimizer Adam adjusts the model parameters according to the gradient of the loss function. At the end of each training round, the optimizer updates the weights of the feature extractor, classifier, and domain classifier based on the backpropagated gradient.
4.
Training and validation
Training phase: in each epoch, the training data from the source domain are trained by the model and the training accuracy is calculated.
Validation phase: at the end of each epoch, the migration effectiveness of the model is evaluated using validation data from the target domain. Evaluation metrics such as accuracy, precision, recall, and F1 score are calculated to measure the performance of the model on the target domain.
5.
Updating the best indicators
After each epoch, the code calculates the accuracy, precision, recall, and F1 score on the validation set and updates the best results. Eventually, the optimal accuracy, precision, recall, and F1 score are recorded to choose the best hyperparameters.

3.2. Structural Conditioning Module Incorporating Optuna Hyperparametric Optimization Algorithm

The module is able to dynamically adjust the depth and width of the neural network. At the same time, the module, in combination with Optuna, chooses whether to increase the depth or width of the network or to adjust these parameters within a given range by means of a built-in optimization algorithm. This dynamic tuning not only improves the performance of the model on the target domain but also optimizes computational resources to avoid overfitting or underfitting.

3.2.1. Defining the Network Structure

We use the pre-trained model as a base and allow Optuna to adjust the depth, width, and freeze layer of the network during training.

3.2.2. Setting up Optuna to Optimize Space

Optuna [33] selects the appropriate hyperparameter optimization space to search for the optimal depth and width. The Optuna optimization problem can be defined as
O b j e c t i v e D , W = L o s s D , W + α · C o m p l e x i t y D , W
where L o s s D , W is the loss of the model on the validation set, α is the weight that weighs the loss against the computational complexity, and C o m p l e x i t y D , W is the computational complexity associated with the network depth D and width W .

3.2.3. Depth and Width Adjustment

1. Depth tuning: the model increases the depth of the network by adding convolutional layers, fully connected layers, etc. At the same time, the model will decide whether to add layers and how many layers. Assuming that the initial network has L 0 layers and Optuna has the option to add and subtract layers, the new depth of the network is
D = L 0 + L a d d , L a d d 0 , 1 , 2 , , L a d d
where L a d d is the number of layers added and L max is the maximum amount of additions Optuna can select.
2. Width adjustment: for each layer, an increase in width usually means adding more neurons or a larger size convolution kernel. Optuna will determine the width of each layer based on the complexity of the data in the target domain. Assuming that the width of each layer can be expressed as ω i , the width is adjusted for the i th layer:
W i = ω i × Δ W ,   Δ W W min , W max
where Δ W is the percentage increase in width and W min , W max is the range of width adjustment.

3.2.4. Freezing and Thawing Layers

1. Freezing layers: to avoid overfitting, pre-trained layers can be frozen so that their parameters are not involved in backpropagation and optimization. Assuming that the layer i is frozen, the parameters of this layer θ i remain unchanged during the training process:
L θ = 0
2. Unfreeze layers: newly added layers can start training. When the network is trained to a certain point, there is an option to unfreeze the pre-trained layers in order to fine-tune them on the target domain. For unfrozen layers, weights are trained. At some stage, the parameters of the unfrozen layers can be updated:
θ i n e w = θ i o l d η · L θ i
where η is the learning rate.

3.2.5. Training and Assessment

After each trial, the loss and accuracy are computed through the training set, and the performance of the model on the target domain is evaluated through the validation set, which ultimately returns the evaluation metrics for the objective function, which is
F = L o s s D , W + A c c u r a c y D , W

3.3. Hyperparameter Optimization of the Model

In this paper, we use Optuna to optimize the hyperparameters of the model. With Bayesian optimization, dynamic spatial tuning of hyperparameters, and multi-objective optimization, Optuna is able to efficiently balance exploration and exploitation in the hyperparameter optimization process. New areas are explored to discover potentially optimal hyperparameters, while known good areas are exploited to progressively optimize the performance of the model. The hyperparameters being optimized include learning rate, task weight, and domain weight, and the workflow of Optuna can be divided into the following steps.

3.3.1. Defining the Objective Function

The objective function is the core of the Optuna optimization process and represents the task we want to optimize. Inside the objective function, the training process of the model uses these hyperparameters and returns evaluation metrics (accuracy and F1 score), expressed in mathematical formulas as follows:
max f ( θ ) = arg ( Pr e c i s i o n ( θ ) , Re c a l l ( θ ) , F 1 ( θ ) , A c c u r a c y ( θ ) )
where θ is the model hyperparameters, Pr e c i s i o n ( θ ) is the precision, Re c a l l ( θ ) is the recall probability, F 1 ( θ ) is the F1 score, and A c c u r a c y ( θ ) is the accuracy.

3.3.2. Defining a Hyperparametric Search Space

In the objective function, we define the search space of hyperparameters such that these hyperparameters will be explored by Optuna during each trial. The search space of the proposed method is detailed in Table 1.

3.3.3. Creating an Optuna Study

Optuna uses the study object to manage the optimization process. A study is a container for Optuna that contains multiple trials, each representing a different combination of hyperparameters at a time. A study automatically manages the optimization process and selects the next set of hyperparameters based on the results of the trials.

3.3.4. Implementing the Optimization Process

During the execution of the optimization, Optuna automatically selects hyperparameters based on the settings of the objective function, runs the objective function, and records the results of each trial. Each trial generates a new set of hyperparameter configurations and performs model training using these hyperparameters.

3.3.5. Recording and Selecting Optimal Hyperparameters

At the end of all trials, the optimal hyperparameters and corresponding results can be viewed. The flowchart of the proposed method is shown in Figure 4.

4. Experimental Verification

4.1. Description of the Dataset

In order to verify the effectiveness of the proposed method for classifying complex faults, the CWRU bearing dataset was selected as the source domain data and the Paderborn University bearing dataset [41] was selected for the target domain data. The feature pairs of the two datasets are shown in Table 2. The advantages of this are as follows.
(1)
Adaptation of cross-domain learning: the CWRU dataset as the source domain has better standardization and consistency and can be used to train a base model, whereas the Paderborn dataset as the target domain is more challenging and complex, which is relevant to the point of this paper.
(2)
Adaptability to real-world environments: through cross-domain learning, the model is able to learn some generalized feature extraction methods from the concise and standard CWRU dataset and then migrate to the noisy and data-complex Paderborn dataset, which is able to better cope with the variations and disturbances of real-world industrial environments.
(3)
Improve robustness: By training the base model on the source domain and then migrating it to the target domain for tuning, the robustness of the model in the face of complex environments, noise, and disturbances in real applications can be enhanced.
Table 2. Comparison of CWRU dataset and Paderborn dataset characteristics.
Table 2. Comparison of CWRU dataset and Paderborn dataset characteristics.
FeatureCWRU DatasetPaderborn Dataset
Fault TypeSingle faultsCompound faults
Fault CreationArtificial EDM machiningArtificial machining + natural wear
Operating ConditionsFixed loadDynamic loads, variable speed operation
Signal TypesVibration signalsVibration + AE + current signals
Data ScaleSmallLarge
AuthenticityLab environmentIndustrial-like scenarios
ApplicationBasic algorithm validationTesting in complex scenarios

4.1.1. Introduction to CWRU Bearing Dataset

The CWRU [42] dataset was provided by Case Western Reserve University and is widely used for fault diagnosis, machine learning, and health monitoring studies. The dataset contains bearing vibration data under several different operating conditions, which is used to study different failure modes of rolling bearings, and the signal acquisition device is shown in Figure 5. The dataset is characterized as follows.
(1)
The failure types include inner ring failure, outer ring failure, rolling body failure, and mixed failure (inner ring and outer ring failure at the same time). This paper selects the inner ring failure and outer ring failure data for research.
(2)
Experimental setup: Datasets were acquired at a fixed load and multiple rotational speeds with four load settings (loaded to 0, 1/3, 2/3, and 1× the rated load) and different rotational speeds (1797 rpm, 1750 rpm, 2100 rpm, etc.).
(3)
The signal type mainly uses vibration signals; the acquisition directions for the X-axis, Y-axis, and Z-axis; and a sampling frequency of 12 kHz. The data acquisition environment is relatively simple, with less interference.
Figure 5. CWRU bearing dataset test bed.
Figure 5. CWRU bearing dataset test bed.
Sensors 25 01851 g005

4.1.2. Introduction to Paderborn University Bearing Dataset

The Paderborn dataset was provided by the University of Paderborn, Germany, and its data acquisition device is shown in Figure 6. The dataset focuses on the simulation of bearing failure data in industrial environments, aiming to study how to diagnose failures in more complex and variable environments, and only the vibration signals in this dataset are studied in this paper. The dataset is characterized as follows.
(1)
The failure types include inner ring failures, outer ring failures, and mixed inner and outer ring failures, and the failure types in the dataset are a mixture of manual processing and natural aging failures. These fault types are selected in the experiment.
(2)
Experimental setting: This dataset uses dynamic load and variable speed operating conditions to simulate load variations closer to those found in industrial environments. The vibration signals in the experiments are collected by multiple sensors, including acoustic emission (AE) and current signals in addition to accelerometers.
(3)
The signal type contains vibration signals (X- and Y-axis), acoustic emission signals (AE), and current signals, with sampling frequencies typically 25.6 kHz or higher. In this paper, only vibration signals are studied.
Figure 6. Paderborn University bearing dataset test bed.
Figure 6. Paderborn University bearing dataset test bed.
Sensors 25 01851 g006

4.2. Experimental Setup

In order to demonstrate the effectiveness of the proposed method, this paper sets up seven sets of comparison experiments:
(1)
In order to show that the proposed method occupies less computational resources than the AutoML tool alone, the proposed method is trained with AutoGluon and Auto-Keras under the same conditions, and its running time on the GPU is compared;
(2)
To illustrate the need for model migration and hyperparameter optimization, the proposed method is compared with RNN and CNN;
(3)
In order to demonstrate that the addition of a neural network depth and width dynamic adjustment module can facilitate the positive migration of pre-trained models and the ability to classify complex faults, the proposed method is compared with DANN, MMD, and DAN.
In order to make the experimental results more convincing, 10 repetitions of each model were conducted, and the main parameters of TL and DL were set consistently: the learning rate was 0.001, the batch size was 128, and the epoch was 20, as detailed in Table 3.

4.3. Analysis of Experimental Results

In order to clearly demonstrate the model performance, accuracy (Table 4), F1 score (Figure 7), confusion matrix (Figure 8), and GPU time (Figure 9) are selected to compare the performance of each method on the target domain dataset.
In the confusion matrix, artificial damage in the outer ring corresponds to label 0, true damage in the outer ring corresponds to label 1, artificial damage in the inner ring corresponds to label 2, true damage in the inner ring corresponds to label 3, damage in the outer and inner rings corresponds to label 4, and no damage corresponds to label 5. The horizontal coordinate is the predicted label, and the vertical coordinate is the true label.
Analyzing the experimental data, the following conclusions can be obtained:
(1)
Comparing method 2 and method 3, the accuracy of the proposed method is better than that of AutoGluon and Auto-Keras, which indicates that the proposed method is able to automatically optimize the hyperparameters of the model while guaranteeing higher fault diagnosis accuracy than the more advanced AutoML tools. This is due to the fact that the proposed method uses TL combined with the structural adjustment module to construct the model in the case where the target domain task is related to the source domain task, which is more advantageous than the model construction methods of AutoGluon and Auto-Keras.
(2)
According to Figure 9, it can be seen that the GPU time of the proposed method is less than those of AutoGluon and Auto-Keras. This in part reflects the fact that the proposed method requires less computational resources than AutoGluon and Auto-Keras. This is due to the fact that TL applies pre-trained models to the target domain, which accurately reduces the search space and thus accelerates the NAS process.
(3)
Comparing method 4 and method 7, it can be seen that the diagnostic accuracy of DANN is slightly higher than that of CNN, indicating that adversarial domain adaptive migration learning can effectively improve the model’s ability to diagnose across domains, and the use of this strategy can effectively promote positive migration.
(4)
Comparing method 1, method 4, method 5. and method 6 shows that the classification accuracy and F1 score of the proposed method are better than those of TL. This indicates that the structural adjustment module can effectively improve the adaptive ability of the pre-trained model when facing a complex fault dataset.

5. Conclusions

In this paper, a domain adversarial transfer learning bearing fault diagnosis model incorporating structural adjustment modules is proposed. First, based on the transfer learning framework, the model that has been trained and matured in the source domain is migrated to the target domain task, and the data distribution difference between the source and target domains is narrowed by the adversarial domain adaptation technique to enhance the cross-domain generalization ability of the model. Then, the Optuna optimization framework is embedded with a module to dynamically adjust the depth and width of the neural network, which achieves the dual function of flexibly freezing (locking parameters) or unfreezing (enabling training) specific network layers according to the requirements of the task in the target domain, and at the same time, utilizing the built-in optimization algorithms (e.g., Bayesian optimization) in Optuna to automatically decide whether to expand the depth of the network (increasing the number of layers) or the width of the network (increasing the number of neurons), determine the expansion range (increase the number of neurons), and determine the scope of expansion to accommodate larger data sizes and more complex failure modes in the target domain. Finally, key hyperparameters of the model are globally optimized by Optuna, including the learning rate, the number of training iterations, the balance coefficient between the task classification loss weight and the domain adaptation loss weight, etc., so as to further improve the diagnostic accuracy and robustness of the model on the target domain.
However, the approach proposed in this paper may have challenges when facing situations where the source and target domains differ significantly, especially when the target domain contains new fault types that are not present in the source domain. Although the distributional differences between domains can be somewhat reduced by adversarial domain adaptation (ADA), it may be difficult for this approach to fully cope with new failure modes in the target domain. For such completely new fault types, the model may need more target domain data or additional policies to enhance its adaptability to the new fault types. In addition, although the dynamic adjustment of depth and width can improve the performance of the model, the adjusted model may still face the problem of insufficient generalization ability if the difference between the target domain and the source domain is too large. Therefore, a combination of more domain-specific knowledge or self-supervised learning methods will be explored in the future to cope with this limitation.

Author Contributions

Writing—original draft, Z.Z. (Zhidan Zhong) and H.X.; Writing—review and editing, Z.Z. (Zhidan Zhong); Visualization, Z.W.; Project administration, Z.Z. (Zhihui Zhang); Data curation, H.X.; Funding acquisition, Z.Z. (Zhidan Zhong). All authors have read and agreed to the published version of the manuscript.

Funding

This project received funding from the Henan Provincial Science and Technology Research and Development Joint Fund (Grant No. 225101610001) and the Henan Provincial Key Research and Development Project (Grant No. 231111222900).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The experimental data can be downloaded from https://engineering.case.edu/bearingdatacenter/apparatus-and-procedures (accessed on 13 March 2025) and https://mb.uni-paderborn.de/kat/forschung/kat-datacenter/bearing-datacenter (accessed on 13 March 2025).

Conflicts of Interest

All authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Liu, R.; Yang, B.; Zio, E.; Chen, X. Artificial intelligence for fault diagnosis of rotating machinery: A review. Mech. Syst. Signal Process. 2018, 108, 33–47. [Google Scholar] [CrossRef]
  2. Zhang, S.; Zhang, S.; Wang, B.; Habetler, T.G. Deep Learning Algorithms for Bearing Fault Diagnostics—A Review. In Proceedings of the 2019 IEEE 12th International Symposium on Diagnostics for Electrical Machines, Power Electronics and Drives (SDEMPED), Toulouse, France, 27–30 August 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 257–263. [Google Scholar] [CrossRef]
  3. Lei, Y.; Yang, B.; Jiang, X.; Jia, F.; Li, N.; Nandi, A.K. Applications of machine learning to machine fault diagnosis: A review and roadmap. Mech. Syst. Signal Process. 2020, 138, 106587. [Google Scholar] [CrossRef]
  4. Lee, D.; Sabater, A. Visibility Graphs, Persistent Homology, and Rolling Element Bearing Fault Detection. In Proceedings of the 2022 IEEE International Conference on Prognostics and Health Management (ICPHM), Detroit (Romulus), MI, USA, 6–8 June 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–5. [Google Scholar] [CrossRef]
  5. Shinde, P.V.; Desavale, R.G.; Jadhav, P.M.; Sawant, S.H. A multi fault classification in a rotor-bearing system using machine learning approach. J. Braz. Soc. Mech. Sci. Eng. 2023, 45, 121. [Google Scholar] [CrossRef]
  6. Phan, Q.N.X.; Le, T.M.; Tran, H.M.; Van Tran, L.; Dao, S.V.T. Novel Machine Learning Techniques for Classification of Rolling Bearings. IEEE Access 2024, 12, 176863–176879. [Google Scholar] [CrossRef]
  7. He, M. Development of Deep Learning Based Methodology on Rotating Machine Fault Diagnosis. Ph.D. Thesis, University of Illinois Chicago, Chicago, IL, USA, 2018. [Google Scholar]
  8. Hoang, D.-T.; Kang, H.-J. A survey on Deep Learning based bearing fault diagnosis. Neurocomputing 2019, 335, 327–335. [Google Scholar] [CrossRef]
  9. Saufi, S.R.; Ahmad, Z.A.B.; Leong, M.S.; Lim, M.H. Challenges and Opportunities of Deep Learning Models for Machinery Fault Detection and Diagnosis: A Review. IEEE Access 2019, 7, 122644–122662. [Google Scholar] [CrossRef]
  10. Shan, S.; Liu, J.; Wu, S.; Shao, Y.; Li, H. A motor bearing fault voiceprint recognition method based on Mel-CNN model. Measurement 2023, 207, 112408. [Google Scholar] [CrossRef]
  11. Li, R.; Yu, P.; Cao, J. Rolling bearing fault diagnosis method based on SOA-BiLSTM. In Proceedings of the 2023 7th International Conference on Electronic Information Technology and Computer Engineering, Xiamen, China, 20–22 October 2023; ACM: New York, NY, USA, 2023; pp. 41–45. [Google Scholar] [CrossRef]
  12. Zhou, H.; Chen, W.; Cheng, L.; Williams, D.; De Silva, C.W.; Xia, M. Reliable and Intelligent Fault Diagnosis with Evidential VGG Neural Networks. IEEE Trans. Instrum. Meas. 2023, 72, 1–12. [Google Scholar] [CrossRef]
  13. Chen, S.; Zou, S. Enhancing Bearing Fault Diagnosis with Deep Learning Model Fusion and Semantic Web Technologies. Int. J. Semantic Web Inf. Syst. 2024, 20, 1–20. [Google Scholar] [CrossRef]
  14. Fan, J.; Ma, C.; Zhong, Y. A Selective Overview of Deep Learning. Stat. Sci. 2021, 36, 264–290. [Google Scholar] [CrossRef]
  15. Hu, X.; Chu, L.; Pei, J.; Liu, W.; Bian, J. Model complexity of deep learning: A survey. Knowl. Inf. Syst. 2021, 63, 2585–2619. [Google Scholar] [CrossRef]
  16. Fristiana, A.H.; Alfarozi, S.A.I.; Permanasari, A.E.; Pratama, M.; Wibirama, S. A Survey on Hyperparameters Optimization of Deep Learning for Time Series Classification. IEEE Access 2024, 12, 191162–191198. [Google Scholar] [CrossRef]
  17. Majidi, F.; Openja, M.; Khomh, F.; Li, H. An Empirical Study on the Usage of Automated Machine Learning Tools. In Proceedings of the 2022 IEEE International Conference on Software Maintenance and Evolution (ICSME), Limassol, Cyprus, 3–7 October 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 59–70. [Google Scholar] [CrossRef]
  18. Sun, Y.; Song, Q.; Gui, X.; Ma, F.; Wang, T. AutoML in The Wild: Obstacles, Workarounds, and Expectations. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, Hamburg, Germany, 23–28 April 2023; ACM: New York, NY, USA, 2023; pp. 1–15. [Google Scholar] [CrossRef]
  19. Hanussek, M.; Blohm, M.; Kintz, M. Can AutoML outperform humans? An evaluation on popular OpenML datasets using AutoML Benchmark. In Proceedings of the 2020 2nd International Conference on Artificial Intelligence, Robotics and Control, Cairo, Egypt, 12–14 December 2020; ACM: New York, NY, USA, 2020; pp. 29–32. [Google Scholar] [CrossRef]
  20. Kim, D.; Koo, J.; Kim, U.-M. A Survey on Automated Machine Learning: Problems, Methods and Frameworks. In Human-Computer Interaction. Theoretical Approaches and Design Methods; Kurosu, M., Ed.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2022; Volume 13302, pp. 57–70. [Google Scholar] [CrossRef]
  21. Feurer, M.; Eggensperger, K.; Falkner, S.; Falkner, S.; Lindauer, M.; Hutter, F. Auto-Sklearn 2.0: Hands-free AutoML via Meta-Learning. arXiv 2020, arXiv:2007.04074. [Google Scholar]
  22. Neutatz, F.; Lindauer, M.; Abedjan, Z. AutoML in heavily constrained applications. VLDB J. 2024, 33, 957–979. [Google Scholar] [CrossRef]
  23. Pham, H.; Guan, M.Y.; Zoph, B.; Le, Q.V.; Dean, J. Efficient Neural Architecture Search via Parameter Sharing. arXiv 2018, arXiv:1802.03268. [Google Scholar]
  24. He, X.; Zhao, K.; Chu, X. AutoML: A survey of the state-of-the-art. Knowl.-Based Syst. 2021, 212, 106622. [Google Scholar] [CrossRef]
  25. Luo, Y.; Wang, M.; Zhou, H.; Yao, Q.; Tu, W.; Chen, Y.; Chen, Y.; Dai, W.; Yang, Q. AutoCross: Automatic Feature Crossing for Tabular Data in Real-World Applications. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; ACM: New York, NY, USA, 2019; pp. 1936–1945. [Google Scholar] [CrossRef]
  26. Elsken, T.; Metzen, J.H.; Metzen, J.; Hutter, F. Neural Architecture Search: A Survey. arXiv 2019, arXiv:1808.05377. [Google Scholar]
  27. Lu, J.; Behbood, V.; Hao, P.; Zuo, H.; Xue, S.; Zhang, G. Transfer learning using computational intelligence: A survey. Knowl.-Based Syst. 2015, 80, 14–23. [Google Scholar] [CrossRef]
  28. Zhuang, F.; Qi, Z.; Duan, K.; Xi, D.; Zhu, Y.; Zhu, H.; Xiong, H.; He, Q. A Comprehensive Survey on Transfer Learning. Proc. IEEE 2021, 109, 43–76. [Google Scholar] [CrossRef]
  29. Chen, X.; Yang, R.; Xue, Y.; Huang, M.; Ferrero, R.; Wang, Z. Deep Transfer Learning for Bearing Fault Diagnosis: A Systematic Review Since 2016. IEEE Trans. Instrum. Meas. 2023, 72, 1–21. [Google Scholar] [CrossRef]
  30. Tang, S.; Ma, J.; Yan, Z.; Zhu, Y.; Khoo, B.C. Deep transfer learning strategy in intelligent fault diagnosis of rotating machinery. Eng. Appl. Artif. Intell. 2024, 134, 108678. [Google Scholar] [CrossRef]
  31. Wu, Z.; Jiang, H.; Zhao, K.; Li, X. An adaptive deep transfer learning method for bearing fault diagnosis. Measurement 2020, 151, 107227. [Google Scholar] [CrossRef]
  32. Huo, C.; Jiang, Q.; Shen, Y.; Zhu, Q.; Zhang, Q. Enhanced transfer learning method for rolling bearing fault diagnosis based on linear superposition network. Eng. Appl. Artif. Intell. 2023, 121, 105970. [Google Scholar] [CrossRef]
  33. Thuan, N.D. Robust knowledge transfer for bearing diagnosis in neural network models using multilayer maximum mean discrepancy loss function. Meas. Sci. Technol. 2024, 35, 126129. [Google Scholar] [CrossRef]
  34. Zhang, Z.; Li, J.; Cai, C.; Ren, J.; Xue, Y. Bearing Fault Diagnosis Based on Image Information Fusion and Vision Transformer Transfer Learning Model. Appl. Sci. 2024, 14, 2706. [Google Scholar] [CrossRef]
  35. Akiba, T.; Sano, S.; Yanase, T.; Ohta, T.; Koyama, M. Optuna: A Next-generation Hyperparameter Optimization Framework. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; ACM: New York, NY, USA, 2019; pp. 2623–2631. [Google Scholar] [CrossRef]
  36. Pan, S.J.; Yang, Q. A Survey on Transfer Learning. IEEE Trans. Knowl. Data Eng. 2010, 22, 1345–1359. [Google Scholar] [CrossRef]
  37. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M.; Lempitsky, V. Domain-Adversarial Training of Neural Networks. In Domain Adaptation in Computer Vision Applications; Csurka, G., Ed.; Advances in Computer Vision and Pattern Recognition; Springer International Publishing: Cham, Switzerland, 2017; pp. 189–209. [Google Scholar] [CrossRef]
  38. Gao, H.; Zhang, X.; Gao, X.; Li, F.; Han, H. ICoT-GAN: Integrated Convolutional Transformer GAN for Rolling Bearings Fault Diagnosis Under Limited Data Condition. IEEE Trans. Instrum. Meas. 2023, 72, 1–14. [Google Scholar] [CrossRef]
  39. Li, Y.; Yang, R.; Wang, H. Unsupervised Method Based on Adversarial Domain Adaptation for Bearing Fault Diagnosis. Appl. Sci. 2023, 13, 7157. [Google Scholar] [CrossRef]
  40. Jeong, S.; Kim, B.; Cha, S.; Seo, K.; Chang, H.; Lee, J.; Kim, Y.; Noh, J. Real-Time CNN Training and Compression for Neural-Enhanced Adaptive Live Streaming. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 6023–6039. [Google Scholar] [CrossRef]
  41. Lessmeier, C.; Kimotho, J.K.; Zimmer, D.; Sextro, W. Condition Monitoring of Bearing Damage in Electromechanical Drive Systems by Using Motor Current Signals of Electric Motors: A Benchmark Data Set for Data-Driven Classification. PHM Soc. Eur. Conf. 2016, 3, 1. [Google Scholar] [CrossRef]
  42. Smith, W.A.; Randall, R.B. Rolling element bearing diagnostics using the Case Western Reserve University data: A benchmark study. Mech. Syst. Signal Process. 2015, 64–65, 100–131. [Google Scholar] [CrossRef]
Figure 1. DANN basic architecture.
Figure 1. DANN basic architecture.
Sensors 25 01851 g001
Figure 2. CNN architecture diagram.
Figure 2. CNN architecture diagram.
Sensors 25 01851 g002
Figure 3. Data preprocessing.
Figure 3. Data preprocessing.
Sensors 25 01851 g003
Figure 4. Flowchart of the proposed method.
Figure 4. Flowchart of the proposed method.
Sensors 25 01851 g004
Figure 7. F1 scores for the eight models.
Figure 7. F1 scores for the eight models.
Sensors 25 01851 g007
Figure 8. Confusion matrix for eight models. (a) Method 1. (b) Method 2. (c) Method 3. (d) Method 4. (e) Method 5. (f) Method 6. (g) Method 7. (h) Method 8.
Figure 8. Confusion matrix for eight models. (a) Method 1. (b) Method 2. (c) Method 3. (d) Method 4. (e) Method 5. (f) Method 6. (g) Method 7. (h) Method 8.
Sensors 25 01851 g008
Figure 9. GPU time.
Figure 9. GPU time.
Sensors 25 01851 g009
Table 1. Hyperparametric search space.
Table 1. Hyperparametric search space.
Convolutional
Layer Count
Convolutional Layer
Output Channel Number
Convolution
Kernel Size
Convolutional StrideFull Connection
Layer Quantity
Fully Connected Layer Neuron CountLearning RateTask WeightDomain Weight
2–1016–2562–641–161–316–2561 × e−5–1 × e−20–20–1
Table 3. Main parameters of the eight models.
Table 3. Main parameters of the eight models.
No.MethodLearning RateBatch SizeEpochTrialHard Ware
1Proposed Method 12820150GPU
2AutoGluon 12820150GPU
3Auto-keras 12820150GPU
4DANN0.00112820 GPU
5MMD0.00112820 GPU
6DAN0.00112820 GPU
7CNN0.00112820 GPU
8RNN0.00112820 GPU
Table 4. Accuracy of the eight models.
Table 4. Accuracy of the eight models.
No.MethodAccuracy (%)
1Proposed Method98.53 ± 0.35
2AutoGluon88.75 ± 0.22
3Auto-keras81.25 ± 0.13
4DANN92.47 ± 5.10
5MMD85.46 ± 4.05
6DAN86.24 ± 1.60
7CNN89.56 ± 1.45
8RNN89.23 ± 1.32
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhong, Z.; Xie, H.; Wang, Z.; Zhang, Z. Domain Adversarial Transfer Learning Bearing Fault Diagnosis Model Incorporating Structural Adjustment Modules. Sensors 2025, 25, 1851. https://doi.org/10.3390/s25061851

AMA Style

Zhong Z, Xie H, Wang Z, Zhang Z. Domain Adversarial Transfer Learning Bearing Fault Diagnosis Model Incorporating Structural Adjustment Modules. Sensors. 2025; 25(6):1851. https://doi.org/10.3390/s25061851

Chicago/Turabian Style

Zhong, Zhidan, Hao Xie, Zhenxin Wang, and Zhihui Zhang. 2025. "Domain Adversarial Transfer Learning Bearing Fault Diagnosis Model Incorporating Structural Adjustment Modules" Sensors 25, no. 6: 1851. https://doi.org/10.3390/s25061851

APA Style

Zhong, Z., Xie, H., Wang, Z., & Zhang, Z. (2025). Domain Adversarial Transfer Learning Bearing Fault Diagnosis Model Incorporating Structural Adjustment Modules. Sensors, 25(6), 1851. https://doi.org/10.3390/s25061851

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop