Next Article in Journal
Design of Amorphous Carbon Coatings Using Gaussian Processes and Advanced Data Visualization
Next Article in Special Issue
Analysis of Grooves Used for Bearing Lubrication Efficiency Enhancement under Multiple Parameter Coupling
Previous Article in Journal
Tilting-Pad Journal Bearings—Frequency-Dependent Dynamic Coefficients and Pivot Flexibility Effects
Previous Article in Special Issue
A Study on Bearing Dynamic Features under the Condition of Multiball–Cage Collision
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SACGNet: A Remaining Useful Life Prediction of Bearing with Self-Attention Augmented Convolution GRU Network

1
Key Laboratory of Knowledge Engineering with Big Data, School of Computer Science and Information Engineering, Hefei University of Technology, Hefei 230009, China
2
Shanghai Aerospace Control Technology Institute, Shanghai 201109, China
3
Luoyang Bearing Research Institute Co., Ltd., Luoyang 471033, China
4
School of Computer and Information, Hefei University of Technology, Hefei 230009, China
*
Author to whom correspondence should be addressed.
Lubricants 2022, 10(2), 21; https://doi.org/10.3390/lubricants10020021
Submission received: 11 January 2022 / Revised: 24 January 2022 / Accepted: 1 February 2022 / Published: 3 February 2022
(This article belongs to the Special Issue Advances in Bearing Lubrication and Thermal Sciences)

Abstract

:
In recent years, the development of deep learning-based remaining useful life (RUL) prediction methods of bearings has flourished because of their high accuracy, easy implementation, and lack of reliance on a priori knowledge. However, there are two challenging issues concerning the prediction accuracy of existing methods. The run-to-failure sequential data and its RUL labels are almost inaccessible in real-world scenarios. Meanwhile, the existing models usually capture the general degradation trend of bearings while ignoring the local information, which restricts the model performance. To tackle the aforementioned problems, we propose a novel health indicator derived from the original vibration signals by combining principal components analysis with Euclidean distance metric, which was motivated by the desire to resolve the dependency on RUL labels. Then, we design a novel self-attention augmented convolution GRU network (SACGNet) to predict the RUL. Combining a self-attention mechanism with a convolution framework can both adaptively assign greater weights to more important information and focus on local information. Furthermore, Gated Recurrent Units are used to parse the long-term dependencies in weighted features such that SACGNet can utilize the important weighted features and focus on local features to improve the prognostic accuracy. The experimental results on the PHM 2012 Challenge dataset and the XJTU-SY bearing dataset have demonstrated that our proposed method is superior to the state of the art.

1. Introduction

Bearings are one of the key components in a rotating machinery system. The remaining useful life (RUL) of a bearing is often defined as the length of a bearing from the current time to failure [1]. If the damage time or the trend of the vibration signal can be predicted from the collected vibration signal of a bearing, it is beneficial for identifying the adverse running condition in time to avoid the sudden danger of bearings. Thus, the RUL of a bearing is essential for the maintenance and management of mechanical systems [1,2].
In general, the RUL prediction of bearings can be sorted into two different directions: physics-based methods and data-driven methods. Physics-based methods focus on physical and mathematical models, e.g., partial differential equations and state-space models, which require extensive prior knowledge [3,4,5,6].
Data-driven RUL methods directly use historical data to model the degradation process of bearings without any prior knowledge.
Deep learning is a popular approach among data-driven methods, which can directly build a deep neural network to model the degradation process as a functional relationship between health states and original sensory data [7].
Deep learning-based RUL approaches typically include the steps of data acquisition, health indicator (HI) construction, and remaining useful life prediction [7].
Data acquisition is to collect the run-to-failure signals from different sensors that can reflect the degradation process of bearings. The complete lifecycle data of a bearing are usually high-dimension and nonlinear. Therefore, a suitable processing method to retain the degradation features (i.e., HI) is necessary.
Health indicators such as the RUL labels are constructed by selecting the appropriate characteristics from the original sensory signal. Since the damage extent of bearings cannot be directly observed, the RUL labels are almost inaccessible in real-world scenarios. Thereby, the critical information from the original data are extracted as HI to train the prediction model, which is a crucial issue to the model [8].
Afterwards, several deep neural networks are designed to extract deep features from the original sensory data and then predict the RUL. The prevalent models include RNN, LSTM, CNN, etc. RNN are often used for predicting the time-series data. The prevalent RNNs are mainly LSTM, GRU, and their variants, which can learn the general degradation trend of the input data. However, they often overlook the local features in input data. With respect to the sequence vibration signal of the bearing, these vibration data have merely small fluctuations in long time series. Until the end of the bearing’s life, the vibration fluctuates dramatically, which is often difficult to predict; hence, the existing models cannot obtain satisfactory prediction results [9].
To tackle the aforementioned issues, we propose a novel health indicator-based remaining useful life prediction approach of bearings. The main contributions of this paper are summarized as follows:
  • We combine the PCA with Euclidean distance metric methods to construct a health indicator to tackle the problem of lack of RUL labels. Facing the high-dimensional and long-term series data, PCA can reduce the data dimensionality while retaining sufficient useful features. The Euclidean distance is to measure the similarity between data to distinguish the different degradation stages. Compared with the existing linear RUL labels, our HI is not only capable of representing the general degradation trend of bearings, but it also can retain more local features from the original vibration signal, which benefit the corresponding model’s learning and calculations.
  • We design a novel self-attention augmented convolution GRU network (SACGNet) to predict the RUL. Combining the self-attention mechanism with a convolution framework can both adaptively assign greater weights to more important information and focus on local information. Furthermore, Gated Recurrent Units (GRU) are used to parse the long-term dependencies in weighted features so that SACGNet can utilize the important weighted features and focus on local features to improve the prognostic accuracy.
  • Based on the designed HI and SACGNet, a novel remaining useful life prediction approach is proposed. We conduct ablation experiments and different comparison experiments on the PHM 2012 Challenge dataset and XJTU-SY bearing dataset. The experimental results prove the superiority of our proposed method.
The remaining part of this paper is organized as follows: In Section 2, we introduce related works in the field of RUL prediction. We describe our proposed method in detail in Section 3. The experimental results are discussed in Section 4. Finally, we conclude the paper in Section 5.

2. Related Works

2.1. Health Indicator Construction

In deep learning-based RUL methods, HI construction currently has two branches in general. One approach extracts simple physical fault characterization from the original vibration signals as HI, using statistical methods or signal processing methods. For instance, the root mean square (RMS) of the original vibration signal [10] or the percentage of useful life (the current life divided by the total useful life) [11,12,13,14]. However, such HIs cannot represent enough useful degradation information of the original data. Thereby, using such HIs as model input makes the model fail to accurately capture the degradation trend for RUL prediction.
The other branch constructs the virtual HI by fusing multiple physical characteristics or multi-sensor signals. These HI can filter out abnormal trends in the early degradation stages, which is more suitable for model learning [15,16,17,18]. Guo et al. selected six related-similarity features and combined eight time-frequency features so as to form an original feature set that contains rich degradation signatures of bearings. Then, the selected features are fused into an HI through an RNN [19]. Li et al. used KPCA to integrate multiple features and introduced the EWMA to reduce the fluctuations for the constructed HI [20]. Li et al. designed the generative adversarial network to learn the data distribution in the health states of machine, using the output of the discriminator as HI [21]. Liang et al. proposed a novel index by calculating offset distance and offset angle between the current state and normal state of devices [22].
In summary, the existing HI can only represent the global degradation process of the vibration signal, but it fails to retain more local features. In order to extract more representative features from the vibration signal and facilitate for the model learning, a more effective HI construction method is proposed in this paper.

2.2. Prediction Model

With respect to regression model design, LSTM, GRU, CNN, and the attention mechanism have been successively introduced into the field of RUL prediction.
LSTM uses input gates, forgetting gates, and output gates to regulate the information of the input sequence, which enables the network to learn the long-term dependence of the data and gain favorable results. Hinchi et al. used convolutional layers to directly extract local features from sensor data, combined them with LSTM layers to capture the degradation process of the bearing, and finally output the prediction values [23]. Whereas LSTM solves the problem of gradient disappearance of traditional RNN to some extent, the deliberate design of LSTM for RUL prediction is very time consuming [24,25].
GRU is a variant of LSTM, whose structure is further simplified to show better performance than LSTM in smaller datasets [26,27]. Cao et al. use the BiGRU model to solve the problem of distribution discrepancy [28]. However, with regard to LSTM and GRU, they only use the features learned in the previous time step for regression prediction and often do not pay attention to local features in the long time series [29].
CNN can extract features with less computational effort because of the sparsity of parameter sharing of the convolutional kernel and inter-layer connectivity. More importantly, CNN focuses on the local features in the original vibration signal, which is suitable for RUL prediction [30,31,32,33]. Wang et al. proposed a multi-scale convolutional network to improve the domain adaptation capability of the RUL prediction model [34].
It is noted that the original vibration signal often contains different features with different levels of importance. The features that contain more important information should be paid more attention. Hence, an attention mechanism is introduced to RUL prediction of bearing to adaptively extract input features [35]. The self-attention mechanism aims to correlate different states of sequences, which reduces the dependence on external information and is more suitable for capturing the internal relevance of data or features [36,37]. Chen et al. constructed an encoder–decoder model based on the attention mechanism to mine useful degradation information from a long historical vibration signal [38]. Chen et al. proposed an attention-based deep learning framework for RUL prediction, which adopted LSTM to extract features, and then combined with the attention layer to fusion the features, LSTM extracted and manually extracted features [39].

3. Proposed Method

Without loss of generality, given a bearing, vibration signals V = { v 1 , v 2 , v m } , input V to the health indicator construction module and get the H I = { h 1 , h 2 , h m } . We expect the model to predict the h t + 1 value after input h 1 , h 2 , h t .
Then, our proposed SACGNet learns the deep features in HI:
F ( h 1 , h 2 , , h t , θ ) : h h t + 1
where θ is the parameter on the model.
There is also a testing dataset of the vibration signal V = { v 1 , v 2 , v m } , which after HI construction obtains H = { h 1 , h 2 , h n } .
Finally, inputing the H to the trained SACGNet, the model will predict the correct value: y t ^ = F ( h 1 , h 2 , , h t ) , where y t ^ is the model’s prediction value.
The whole structure of SACGNet is shown in Figure 1, including the health indicator construction module and remaining useful life prediction module. First, input the original signal of the bearings to the health indicator construction module to obtain the HI. After data normalization and sliding window processing, input it to SACGNet for training. In the testing stage, the predicted values are output by autoregression.

3.1. Health Indicator Construction Module

If the dimension of the original signal is set to d, the matrix form of the original vibration signal V = { v 1 , v 2 , v m } , where v i denoted an acquired vibration data that can be written as:
V = v 11 . . . v 1 d v m 1 v m d .
Principal components analysis (PCA) linearly transforms the data into a new coordinate system such that the first major variance of any data projection is at the first coordinate (called the first principal component), the second major variance is at the second coordinate, and so on.
V T is the de-averaged data. The singular value decomposition of V is:
V = W H T
where the matrix W is the eigenvector matrix of V V T , ∑ is a non-negative rectangular diagonal matrix, and H is the eigenvector matrix of V T V .
Assuming zero empirical means, the principal component w ( 1 ) of the dataset V can be defined as:
w ( 1 ) = a r g m a x w = 1 V a r { W T V } .
To obtain the k-th principal component, the previous k−1 principal components must first be subtracted from V.
V k 1 ^ = V i 1 k 1 w i w i T V
Then, the k-th principal component is obtained to update a new dataset and continue to search for principal components.
w k = a r g m a x w = 1 E { ( w T V k 1 ^ ) 2 }
Through PCA, we can reduce the original vibration signal’s dimensionality from d to k. We retain the k principal components of the original signals; thus, the dimension of V p c a is k, which can be abbreviated as { v p c a 1 , v p c a 2 , v p c a 3 , , v p c a n } .
V p c a = p c a ( V ) = w 11 w 1 k w m 1 w m k = { v p c a 1 , v p c a 2 , v p c a 3 , v p c a m }
Using standard PCA to reduce the dimensionality of the original vibration data, we can only retain the principal components of the data. In this paper, based on PCA, to reduce the dimensionality of vibration data, we use Euclidean distance to calculate the distance between the low-dimensional data to construct HI. The metric Euclidean distance can obtain the similarity between one data in the time series and the neighboring points, which can better reflect the trend of the neighboring data in the original vibration signal, which means “capturing local features” we mentioned.
By calculating the average of the Euclidean distance from each point in V p c a to the sequential neighboring points, we can obtain the HI corresponding to each point. HI = { h 1 , h 2 , h m } . The calculation process of h i is as follows:
h i = 1 2 ( j = 1 k ( v p c a i j v p c a ( i + 1 ) j ) 2 + j = 1 k ( v p c a i j v p c a ( i 1 ) j ) 2 ) .
HI will be input to the constructed SACGNet to make the model learn the relationship between them. In order to make HI meet the dimensionality requirements of the model input, a sliding window is set to process the data into the shape required by the model, with a sliding window size of 20. Then, we obtain the X = { x 1 , x 2 , x n } .

3.2. Remaining Useful Life Prediction Module

In this section, we describe our SACGNet in detail, as shown in the Table 1. We combine a 1D convolution (Conv1d) block with self-attention mechanisms to extract deep features from the input data. The Conv1d block focuses more on local features, and the self-attention mechanism can extract global features of the data. GRU can identify long-term features in the input data, which is beneficial to adapt the bearings under different operating conditions, thereby improving the prediction accuracy of our model [40,41].
For convenience, we use C, P, FC, D, MHA, and GRU to denote the Conv1d layer, the pooling layer, the fully connected layer, the dropout layer, Multi-Head attention layer, and the GRU layer, respectively.
In the convolution layer, the calculation of the input data can be written as follows:
X c = R e L U ( X f i + b i )
where ⊙ represents convolution operation. f i represents the ith convolution filter, and b i is the bias. The convolution layers used R e L U as the activation function. Compared to images, the vibration signal is time-series data; hence, the one-dimensional convolution (Conv1d) neural network can be used to perform convolutional operations. The filters of the Conv1d layer are set to 80, the kernel size is set to 4, the stride is set to 1. In our paper, we select the average pooling layer.
The specific calculation process of the self-attention mechanism can be summarized into two processes: calculation the weight coefficients based on the Query and Key, and summation of the weight values based on the weight coefficients. The first process can further include the following: first, calculate the similarity or relevance between Query and Key, and then normalize the found relevance. Its attention function can be described as mapping a Query and a pair of key-value pairs to an output, where Queries, Keys, and values are vectors and the output is computed as a weighted sum of values, where the weight assigned to each value is computed by the compatibility function of the Query with the corresponding Key.
In order to learn the expression of multiple meanings, the input data will be transformed; W Q , W K , W V is the matrix of assigned weights. Self-attention represents a focus on itself, so the equation can be denoted as follows:
Q = X c W Q = L i n e a r ( X c ) K = X c W K = L i n e a r ( X c ) V = X c W V = L i n e a r ( X c ) .
The output matrix of self-attention is expressed as:
A t t e n t i o n ( Q , K , V ) = s o f t m a x ( Q K T d k ) V
where d k is the dimension of K, and the use of d k is to change the attention matrix into a standard normal distribution.
Multi-head attention can make the model pay attention to the information from different representational subspaces; the output of the self-attention mechanism layer is three-dimensional vectors, which are written as X a :
X a = M u l t i H e a d ( Q , K , V ) = C o n c a t ( h e a d 1 , h e a d h ) W .
In this paper, we choose h = 8, d k = d q = d v = 80 , W R h d v × d m o d e l , which are the empirical values selected in the experiment.
Gated Recurrent Units (GRUs) are a gating mechanism in recurrent neural networks. The calculation of the GRU can be written as follows:
X a = { a 1 , a 2 , a n } z t = σ g ( W z a t + U z h t 1 + b z ) r t = σ g ( W r a t + U r h t 1 + b r ) h t ^ = ϕ h ( W h x t + U h ( r t h t 1 ) + b h ) h t = ( 1 z t ) h t 1 + z t h t ^ .
Among them, a t is input vector X a at time t, h t is output vector, h t ^ is the candidate activation vector, z t is the update gate vector, r t is reset gate vector, W, U, and b are the parameter matrices, vector σ g is a sigmoid function, and ϕ h is a hyperbolic tangent. The GRU layer receives the features extracted from the Conv1d layer and the Multi-Head attention layer and then outputs the prediction value. The units of GRU are set to 80.
Finally, after the fully connected layer, the final output is obtained:
y t ^ = F C N ( h t ) .
SACGNet is trained using the error back-propagation algorithm and gradient descent method. The loss function of the training process is the mean square error function:
MSE = 1 n i = 1 n ( y i y i ^ ) 2
where y i is the true value, y i ^ is the prediction value, and n is the total number of samples.
In addition, Adam is chosen as the optimizer of this paper, and the learning rate is set to 10 3 [42].
Dropout is added to our SACGNet with the parameter set to 0.5 in order to reduce overfitting by preventing complex co-adaptations on training data. The algorithm pseudo-code is shown in as follow Algorithm 1.
Algorithm 1: Proposed SACGNet.
1. The SACGNet algorithm for training is defined
as follows:
Input: Hyper-parameters of model (batch size, epoch,
dropout rate, learning rate, etc.), original signal V = { v 1 , v 2 , v 3 , v m }
2. H I = { h 1 , h 2 , h m }
3. By sliding window processing:
   H I = { h 1 , h 2 , h m } X = { x 1 , x 2 , x n }
  Each x represents a batch h, the number of a batch is i
4. Y = { y 1 , y 2 , y n } = { h i + 1 , h i + 2 , h m }
5. For i = 1 , 2 , n do:
   X = x i M i n ( X ) M a x ( X ) M i n ( X ) , Y = y i M i n ( Y ) M a x ( Y ) M i n ( Y )
  end
6. Build SACGNet model
7. w (parameters of the SACGNet) and b (biases) are initialized to zeros
8. Input X and Y to train SACGNet
   X c = C o n v ( X )
   Q = X c W Q = L i n e a r ( X c )             K = X c W K = L i n e a r ( X c )             V = X c W V = L i n e a r ( X c )
   A t t e n t i o n ( Q , K , V ) = s o f t m a x ( Q K T d k ) V
   X a = M u l t i H e a d ( Q , K , V )
   O u t p u t = G R U ( X a )
   C o m p u t e M S E b y ( 15 )
   w A d a m ( M S E , w )
   b A d a m ( M S E , b )
   e n d
Output: Trained SACGNet model for prediction
END

4. Experiments and Results

In this section, we use the IEEE PHM Challenge 2012 bearing dataset and the XJTU-SY Bearing dataset to validate the effectiveness of our method.

4.1. Dataset Description

The IEEE PHM 2012 Challenge dataset was collected from the PRONOSTIA testbed, as shown in Figure 2.
The PRONOSTIA test platform contains a rotating part, load part, and data collection part. The motor power of the rotating part is 250 W. The power is transferred to the bearing by the axis of rotation. The load part provides a load of 4000 N to make the bearing degrade quickly. The acceleration sensor is placed on a bearing seat in horizontal and vertical directions to select the vibration signals. The sampling frequency of the acceleration sensor is 25.6 kHz. When the test platform starts to work, the vibration signal is recorded every 10 s, and the sampling time is 0.1 s [43].
The data provided by IEEE PHM Challenge 2012 include three different operating conditions. Seven bearings (bearings 1-1 to 1-7) work in the first condition, the motor speed is 1800 rpm, and the load is 4000 N. Seven bearings (bearings 2-1 to 2-7) work in the second condition, the motor speed is 1650 rpm, and the load is 4200 N. Three bearings (bearings 3-1 to 3-3) work in the third condition, the motor speed is 1500 rpm, and the load is 5000 N. Table 2 illustrates the details of the PHM 2012 dataset.
In this paper, the vibration dada of bearings 1-1, 1-2, 2-1, 2-2, and 3-1 are selected, respectively, as the training set, while the rest of the bearings are selected as the testing set. Table 2 illustrates the details of the PHM dataset.
The XJTU-SY bearing dataset is provided by the Institute of Design Science and Fundamental Research of Xi’an Jiaotong University and contains the run-to-failure vibration data from 15 rolling bearings [44].
As shown in Figure 3, the bearing testbed is composed of an alternating current (AC) induction motor, a motor speed controller, a support shaft, two support bearings (heavy duty roller bearings), and a hydraulic loading system. This testbed is designed to conduct the accelerated degradation tests of the testing bearings under different operating conditions (i.e., different radial force and rotating speed). The radial force is generated by the hydraulic loading system and applied to the housing of tested bearings, and the rotating speed is set and kept by the speed controller of the AC induction motor [44].
Three different operating conditions are set in the accelerated degradation experiments, and five bearings are used under each operating condition. The sampling frequency is 25.6 kHz, and the sampling period is 1 min. Table 3 illustrates the details of the XJTU-SY dataset.
In this paper, the mean square error (MSE), root mean square error (RMSE), mean absolute error (MAE), and mean absolute percentage error (MAPE) are used to evaluate the prediction accuracy. They are respectively computed as follows:
MSE = 1 n i = 1 n ( y i y i ^ ) 2
RMSE = MSE = 1 n i = 1 n ( y i y i ^ ) 2
MAE = 1 n i = 1 n | y i y i ^ |
MAPE = 1 n i = 1 n | y i y i ^ y i | .
In Equations (18)–(21), y i is the label, y i ^ is the model’s prediction value, and n is the total number of samples.

4.2. Different HIs Results

In this section, we compare different HI construction methods to validate the superiority of our proposed HI construction method.
Figure 4 shows the results of different HI construction methods on the bearing 1-3 of the PHM dataset. It can be clearly observed from Figure 4a that the original vibration signal of the bearing 1-3 is in a very smooth state with little fluctuation when the bearing has just begun to work. In the degradation state, the vibration signal usually fluctuates slightly, whereas the overall trend is upward. The signal fluctuation will increase sharply when the bearing finally completely degrades.
As shown in Figure 4b–h, the HIs constructed by the combination of TSNE and other distance metrics basically do not have any regular change trend. Meanwhile, the methods of SAE and Euclidean distance metric can retain the change trend of the original vibration signal, but the early degradation and complete degradation stage of the bearing cannot be completely distinguished. For the vibration signal of the bearing, PCA is a linear transformation method for each of its principal components. Specifically, the linearity of each point is calculated to obtain the principal components and then downscaled; thus, the global trend of the original signal can be retained. Meanwhile, SAE is a nonlinear learning model that requires a lot of training data to get a satisfactory performance. In contrast, our proposed method, as shown in Figure 4i, is more suitable to reflect the change trend of the original signal and can distinguish the early degradation and complete degradation stage of the bearing, which is beneficial for SACGNet to improve its prediction accuracy.
In order to further illustrate the superiority of our HI construct method, we use the percentage of the use life as the HI; then, we compare the RUL prediction results with our proposed methods on the PHM dataset. The red lines are true values of HI, and the blue lines are the prediction values of the model. Figure 5a–c is the RUL results of comparison HI, while Figure 5d–f is the RUL results of our methods.
It can be seen that when the bearing true remaining useful life percentage is used as the HI label, the model fails to learn the degradation trend of the bearing vibration signal, and the final prediction results are not well-fitted. In contrast, using our proposed HI for prediction, the degradation trend of the vibrtion data is depicted more accurately, and the prediction accuracy of the model is improved. Note that our model can properly predict RUL in the stages of rapid degradation of bearing operation, which is of great value for the actual industrial scenarios.

4.3. Ablation Experiments

In order to observe the effects of the different layers in the proposed model, we conduct ablation experiments on the PHM dataset and the XJTU-SY dataset. We let the model constructs remain and respectively remove the Conv1d layer and Multi-Head Attention layer for comparison models. We call them the NoAttention model and NoConv1d model, respectively.
As seen in Table 4, with respect to the PHM dataset, our method achieved the best results in 10 of the 11 bearing data for MSE, RMSE, MAE, and MAPE metrics. Using the MSE metric, our model did not achieve the best result for bearing 2-5; the discrepancy with respect to the best results (i.e., NoAttention model) is 0.002. For the RMSE, MAE, and MAPE metrics, our model does not achieve the optimal results for bearing 2-7, the discrepancy with respect to the best results (i.e., the NoAttention model) is 0.148, 0.124, and 177.228, respectively.
Furthermore, we also conducted the ablation experiment on the XJTU-SY dataset, and the results are shown in Table 5. Using the MAE and MAPE metrics, our model did not achieve the best results for bearings 1-5, 2-3, and 2-4 but only a discrepancy of 1.9% from the best results. Using the MSE and RMSE metrics, our model did not achieve the best results for bearings 2-3 and 2-4, but the discrepancy with the best results (i.e., NoAttention model) was only 3.03%. Except for the afore-mentioned results, the performance of our model is superior to the comparison models.
The results of the ablation experiments conducted on both bearing datasets prove that our proposed model achieves the best results on the largest number of testing bearings. In a comprehensive analysis, the degradation features of different bearings with different operating conditions are different, Conv1d can extract the local features of the original vibration signals, and the Self-Attention mechanism focuses on the global features, which can be integrated to achieve more excellent results.

4.4. Results of Different Models

In this section, we compare the overall prediction accuracy of our proposed model with state-of-the-art methods on the PHM dataset and XJTU-SY dataset. The compared models include CNN, RNN, LSTM, and GRU.
As shown in Table 6, with respect to the PHM dataset, our model achieved the best results in nine out of 11 bearings for MSE and RMSE metrics, and in nine out of 11 bearings for MAE and MAPE metrics. Using the MSE and RMSE metrics, our model did not achieve the best results for bearings 2-5 and 2-7. Using the MAE and MAPE metrics, our model did not achieve the best values for bearings 2-7.
In order to indicate the superior performance of our model, we add the prediction results of all the comparison models in the PHM dataset. Without loss of generality, we visualize the prediction results for bearing 1-4, 1-5, and 1-6 in order to compare them with the previous prediction results, as shown in Figure 6.
From Figure 6a,d,g,j,m, it can be seen that CNN has the worst fitting results on the testing data. The difference between the CNN predicted values and the original signal is very obvious, because the single CNN models are unsuitable for processing the time-series data. The prediction results of RNN and GRU are slightly superior to those of CNN, but there is still a visible difference from the original vibration signal. Furthermore, the RNN performance is significantly inferior to LSTM on long-term series.
The prediction results of SACGNet and LSTM are similar on bearing 1-4, and the fitting results of SACGNet are more favorable when combined with the four evaluation metrics in Table 6. From Figure 6b,c,e,f,h,i,k,l,n,o, it can be seen that on bearing 1-5 and 1-6, SACGNet has superior prediction results, which is consistent with the comparison results of the four evaluation metrics in Table 6.
Upon observation of the original vibration signal of bearing 2-7, the early degradation states of bearing 2-7 show very sharp fluctuations, and the amplitude difference between the early degradation and the complete degradation stage is very small. The vibration fluctuation of the early degradation is even more severe than that of the complete degradation stage. That may be the reason why our model did not achieve optimal results on bearing 2-7.
As shown in Table 7, with respect to the XJTU-SY dataset, it can be seen that our model has achieved the best resluts on most of the testing bearings, expect for the MSE of bearings 2-3 and 3-5, RMSE of bearings 2-3, 2-5 and 3-5, MAE of bearings 1-5, 2-3, 2-5, and 3-5, and MAPE of bearings 1-5, 2-3, and 2-5, respectively.
The reason for such results may be due to the fact that the experimental environment of 2-3, 2-5, and 3-5 is slightly different from the conditions of the bearing dataset we chose as the training set, and these datasets produce fluctuations in the degradation stage that are equal to or even higher than the final complete damage. Specifically, the CNN model achieved the best results on the bearing 2-3 dataset for four metrics, which may be because the degradation process of bearing 2-3 is filled with many small-scale local fluctuations, allowing the CNN to fit this process more directly. As for bearings 2-5 and 3-5, they do not show much sharp fluctuations and also due to the small amount of data compared to the other datasets, RNN and LSTM are able to parse their long-term serial relationships on these two datasets.
In summary, from the comparison experiments on the two datasets, in most of the cases, our model achieves the best prediction accuracy compared to the other existing models, which demonstrates that our proposed model is applicable for RUL prediction.

5. Conclusions

In this paper, we explored a health indicator-based remaining useful life prediction method. First, we combine principal component analysis (PCA) with Euclidean distance measurement to construct the health indicator for tackling the dependency on RUL labels. Then, we design a self-attention augmented convolution GRU network (SACGNet) to predict the RUL task, which utilizes the globe features as well as local important features to improve the prognostic accuracy. To verify the effectiveness of the model, we conducted extensive experiments on two bearing datasets, respectively, and the results demonstrate that SACGNet is superior to these existing models under several evaluation criteria. Meanwhile, the model has excellent generalization performance for multiple bearings.

Author Contributions

J.X. contributed to the conception of the study; S.D. performed the experiment and wrote the manuscript; W.C. contributed significantly to analysis; D.W. performed the data analyses; Y.F. helped perform the analysis with constructive discussions. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Key Research And Development Plan under Grant 2018YFB2000505, in part by the Key Research and Development Plan of Anhui Province under Grant 202104a04020003, and in part by the Fundamental Research Funds for the Central Universities under Grant PA2021KCPY0045.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Public datasets used in our paper: https://github.com/wkzs111/phm-ieee-2012-data-challenge-dataset (accessed on 10 December 2021), https://biaowang.tech/xjtu-sy-bearing-datasets/ (accessed on 10 December 2021).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Uckun, S.; Goebel, K.; Lucas, P.J. Standardizing research methods for prognostics. In Proceedings of the International Conference on Prognostics and Health Management, Denver, CO, USA, 6–9 October 2008; pp. 1–10. [Google Scholar]
  2. Glowacz, A. Acoustic fault analysis of three commutator motors. Mech. Syst. Signal Process. 2019, 133, 106226. [Google Scholar] [CrossRef]
  3. Zarei, J.; Tajeddini, M.A.; Karimi, H.R. Vibration analysis for bearing fault detection and classification using an intelligent filter. Mechatronics 2014, 24, 151–157. [Google Scholar] [CrossRef]
  4. Glowacz, A.; Glowacz, W.; Kozik, J.; Piech, K.; Gutten, M.; Caesarendra, W.; Liu, H.; Brumercik, F.; Irfan, M.; Khan, Z.F. Detection of deterioration of three-phase induction motor using vibration signals. Meas. Sci. Rev. 2019, 19, 241–249. [Google Scholar] [CrossRef] [Green Version]
  5. Ordóñez, C.; Lasheras, F.S.; Roca-Pardinas, J.; de Cos Juez, F.J. A hybrid ARIMA–SVM model for the study of the remaining useful life of aircraft engines. J. Comput. Appl. Math. 2019, 346, 184–191. [Google Scholar] [CrossRef]
  6. Ma, S.; Zhang, X.; Yan, K.; Zhu, Y.; Hong, J. A Study on Bearing Dynamic Features under the Condition of Multiball—Cage Collision. Lubricants 2022, 10, 9. [Google Scholar] [CrossRef]
  7. Lei, Y.; Li, N.; Guo, L.; Li, N.; Yan, T.; Lin, J. Machinery health prognostics: A systematic review from data acquisition to RUL prediction. Mech. Syst. Signal Process. 2018, 104, 799–834. [Google Scholar] [CrossRef]
  8. Singleton, R.K.; Strangas, E.G.; Aviyente, S. Extended Kalman filtering for remaining-useful-life estimation of bearings. IEEE Trans. Ind. Electron. 2014, 62, 1781–1790. [Google Scholar] [CrossRef]
  9. Saidi, L.; Benbouzid, M. Prognostics and Health Management of Renewable Energy Systems: State of the Art Review, Challenges, and Trends. Electronics 2021, 10, 2732. [Google Scholar] [CrossRef]
  10. Zhang, N.; Wu, L.; Wang, Z.; Guan, Y. Bearing remaining useful life prediction based on Naive Bayes and Weibull distributions. Entropy 2018, 20, 944. [Google Scholar] [CrossRef] [Green Version]
  11. Malhi, A.; Yan, R.; Gao, R.X. Prognosis of defect propagation based on recurrent neural networks. IEEE Trans. Instrum. Meas. 2011, 60, 703–711. [Google Scholar] [CrossRef]
  12. Liao, H.; Tian, Z. A framework for predicting the remaining useful life of a single unit under time-varying operating conditions. IEEE Trans. 2013, 45, 964–980. [Google Scholar] [CrossRef]
  13. Hu, L.; Hu, N.Q.; Fan, B.; Gu, F.S.; Zhang, X.Y. Modeling the relationship between vibration features and condition parameters using relevance vector machines for health monitoring of rolling element bearings under varying operation conditions. Math. Probl. Eng. 2015, 2015, 123730. [Google Scholar] [CrossRef]
  14. Zhang, Z.X.; Si, X.S.; Hu, C.H. An age-and state-dependent nonlinear prognostic model for degrading systems. IEEE Trans. Reliab. 2015, 64, 1214–1228. [Google Scholar] [CrossRef]
  15. Hu, C.; Youn, B.D.; Wang, P.; Yoon, J.T. Ensemble of data-driven prognostic algorithms for robust prediction of remaining useful life. Reliab. Eng. Syst. Saf. 2012, 103, 120–135. [Google Scholar] [CrossRef] [Green Version]
  16. Wang, Y.; Peng, Y.; Zi, Y.; Jin, X.; Tsui, K.L. A two-stage data-driven-based prognostic approach for bearing degradation problem. IEEE Trans. Ind. Inform. 2016, 12, 924–932. [Google Scholar] [CrossRef]
  17. Giantomassi, A.; Ferracuti, F.; Benini, A.; Ippoliti, G.; Longhi, S.; Petrucci, A. Hidden Markov model for health estimation and prognosis of turbofan engines. In Proceedings of the International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, Washington, DC, USA, 28–31 August 2011; Volume 54808, pp. 681–689. [Google Scholar]
  18. Kumar, H.; Pai, S.P.; Sriram, N.; Vijay, G. Rolling element bearing fault diagnostics: Development of health index. Proc. Inst. Mech. Eng. Part C J. Mech. Eng. Sci. 2017, 231, 3923–3939. [Google Scholar] [CrossRef]
  19. Guo, L.; Li, N.; Jia, F.; Lei, Y.; Lin, J. A recurrent neural network based health indicator for remaining useful life prediction of bearings. Neurocomputing 2017, 240, 98–109. [Google Scholar] [CrossRef]
  20. Li, X.; Jiang, H.; Xiong, X.; Shao, H. Rolling bearing health prognosis using a modified health index based hierarchical gated recurrent unit network. Mech. Mach. Theory 2019, 133, 229–249. [Google Scholar] [CrossRef]
  21. Li, X.; Zhang, W.; Ma, H.; Luo, Z.; Li, X. Data alignments in machinery remaining useful life prediction using deep adversarial neural networks. Knowl.-Based Syst. 2020, 197, 105843. [Google Scholar] [CrossRef]
  22. Zeming, L.; Jianmin, G.; Hongquan, J.; Xu, G.; Zhiyong, G.; Rongxi, W. A similarity-based method for remaining useful life prediction based on operational reliability. Appl. Intell. 2018, 48, 2983–2995. [Google Scholar] [CrossRef]
  23. Hinchi, A.Z.; Tkiouat, M. Rolling element bearing remaining useful life estimation based on a convolutional long-short-term memory network. Procedia Comput. Sci. 2018, 127, 123–132. [Google Scholar] [CrossRef]
  24. Wang, F.; Liu, X.; Deng, G.; Yu, X.; Li, H.; Han, Q. Remaining life prediction method for rolling bearing based on the long short-term memory network. Neural Process. Lett. 2019, 50, 2437–2454. [Google Scholar] [CrossRef]
  25. Ragab, M.; Chen, Z.; Wu, M.; Foo, C.S.; Kwoh, C.K.; Yan, R.; Li, X. Contrastive adversarial domain adaptation for machine remaining useful life prediction. IEEE Trans. Ind. Inform. 2020, 17, 5239–5249. [Google Scholar] [CrossRef]
  26. Cheng, Y.; Hu, K.; Wu, J.; Zhu, H.; Lee, C.K. A deep learning-based two-stage prognostic approach for remaining useful life of rolling bearing. Appl. Intell. 2021, 1–16. [Google Scholar] [CrossRef]
  27. Erdenebayar, U.; Kim, Y.; Park, J.U.; Lee, S.; Lee, K.J. Automatic Classification of Sleep Stage from an ECG Signal Using a Gated-Recurrent Unit. Int. J. Fuzzy Log. Intell. Syst. 2020, 20, 181–187. [Google Scholar] [CrossRef]
  28. Cao, Y.; Jia, M.; Ding, P.; Ding, Y. Transfer learning for remaining useful life prediction of multi-conditions bearings based on bidirectional-GRU network. Measurement 2021, 178, 109287. [Google Scholar] [CrossRef]
  29. Liu, H.; Liu, Z.; Jia, W.; Lin, X. Remaining useful life prediction using a novel feature-attention-based end-to-end approach. IEEE Trans. Ind. Inform. 2020, 17, 1197–1207. [Google Scholar] [CrossRef]
  30. Ma, M.; Mao, Z. Deep-convolution-based LSTM network for remaining useful life prediction. IEEE Trans. Ind. Inform. 2020, 17, 1658–1667. [Google Scholar] [CrossRef]
  31. Liu, R.; Yang, B.; Hauptmann, A.G. Simultaneous bearing fault recognition and remaining useful life prediction using joint-loss convolutional neural network. IEEE Trans. Ind. Inform. 2019, 16, 87–96. [Google Scholar] [CrossRef]
  32. Zhu, J.; Chen, N.; Peng, W. Estimation of bearing remaining useful life based on multiscale convolutional neural network. IEEE Trans. Ind. Electron. 2018, 66, 3208–3216. [Google Scholar] [CrossRef]
  33. Cheng, C.; Ma, G.; Zhang, Y.; Sun, M.; Teng, F.; Ding, H.; Yuan, Y. Online bearing remaining useful life prediction based on a novel degradation indicator and convolutional neural networks. arXiv 2018, arXiv:1812.03315. [Google Scholar]
  34. Ge, Y.; Liu, J.; Ma, J. Remaining Useful Life Prediction Using Deep Multi-scale Convolution Neural Networks. IOP Conf. Ser. Mater. Sci. Eng. 2021, 1043, 032011. [Google Scholar] [CrossRef]
  35. Luo, J.; Zhang, X. Convolutional neural network based on attention mechanism and Bi-LSTM for bearing remaining life prediction. Appl. Intell. 2021, 52, 1–16. [Google Scholar] [CrossRef]
  36. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2017; pp. 5998–6008. [Google Scholar]
  37. Mo, Y.; Wu, Q.; Li, X.; Huang, B. Remaining useful life estimation via transformer encoder enhanced by a gated convolutional unit. J. Intell. Manuf. 2021, 32, 1–10. [Google Scholar] [CrossRef]
  38. Chen, Y.; Peng, G.; Zhu, Z.; Li, S. A novel deep learning method based on attention mechanism for bearing remaining useful life prediction. Appl. Soft Comput. 2020, 86, 105919. [Google Scholar] [CrossRef]
  39. Chen, Z.; Wu, M.; Zhao, R.; Guretno, F.; Yan, R.; Li, X. Machine remaining useful life prediction via an attention-based deep learning approach. IEEE Trans. Ind. Electron. 2020, 68, 2521–2531. [Google Scholar] [CrossRef]
  40. Bello, I.; Zoph, B.; Vaswani, A.; Shlens, J.; Le, Q.V. Attention augmented convolutional networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 3286–3295. [Google Scholar]
  41. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  42. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  43. Nectoux, P.; Gouriveau, R.; Medjaher, K.; Ramasso, E.; Chebel-Morello, B.; Zerhouni, N.; Varnier, C. PRONOSTIA: An experimental platform for bearings accelerated degradation tests. In Proceedings of the IEEE International Conference on Prognostics and Health Management, PHM’12, London, UK, 23–27 September 2012; pp. 1–8. [Google Scholar]
  44. Wang, B.; Lei, Y.; Li, N.; Li, N. A hybrid prognostics approach for estimating remaining useful life of rolling element bearings. IEEE Trans. Reliab. 2018, 69, 401–412. [Google Scholar] [CrossRef]
Figure 1. Proposed complete model.
Figure 1. Proposed complete model.
Lubricants 10 00021 g001
Figure 2. Pronostia bearing testbed.
Figure 2. Pronostia bearing testbed.
Lubricants 10 00021 g002
Figure 3. XJTU-SY bearing testbed.
Figure 3. XJTU-SY bearing testbed.
Lubricants 10 00021 g003
Figure 4. Different HI construction methods on the PHM dataset.
Figure 4. Different HI construction methods on the PHM dataset.
Lubricants 10 00021 g004
Figure 5. The remaining useful life prediction using the True RUL label and our method as the label.
Figure 5. The remaining useful life prediction using the True RUL label and our method as the label.
Lubricants 10 00021 g005
Figure 6. The prediction results of different models in the PHM dataset.
Figure 6. The prediction results of different models in the PHM dataset.
Lubricants 10 00021 g006
Table 1. The architecture of SACGNet.
Table 1. The architecture of SACGNet.
NoSymoblOperatorKernel SizeDimension
1InputInput signal/(None, 20, 2)
2C1Convolution4 × 4(None, 20, 80)
3P1Average pooling1 × 1(None, 20, 80)
4C2Convolution4 × 4(None, 20, 80)
5P2Average pooling1 × 1(None, 20, 80)
6FC1Fully connected80 × 1(None, 20, 80)
7D1Dropout/(None, 20, 80)
8MHAMulti-Head Attention/(None, 20, 80)
9FC2Fully connected80 × 1(None, 20, 80)
10FC3Fully connected80 × 1(None, 20, 80)
11D2Dropout/(None, 20, 80)
12GRUGated recurrent units/(None, 80)
13D3Dropout/(None, 80)
14FC4Fully connected1 × 1(None, 1)
Table 2. The detail of the PHM dataset.
Table 2. The detail of the PHM dataset.
Working ConditionRotation SpeedLoadDatasetSample NumberBearing LifetimeDivision
11800 rpm4000 NBearing 1-128037 h 47 mtraining
Bearing 1-28712 h 25 mtraining
Bearing 1-318025 h 10 stesting
Bearing 1-411393 h 9 m 40 stesting
Bearing 1-523026 h 23 m 30 stesting
Bearing 1-623026 h 23 m 29 stesting
Bearing 1-715024 h 10 m 11 stesting
21650 rpm4200 NBearing 2-19112 h 31 m 40 straining
Bearing 2-27972 h 12 m 40 straining
Bearing 2-312023 h 20 m 10 stesting
Bearing 2-46121 h 41 m 50 stesting
Bearing 2-520025 h 33 m 30 stesting
Bearing 2-65721 h 35 m 10 stesting
Bearing 2-717228 m 30 stesting
31500 rpm5000 NBearing 3-15151 h 25 m 40 straining
Bearing 3-216374 h 32 m 40 straining
Bearing 3-335258 m 30 stesting
Table 3. The details of the XJTU-SY dataset.
Table 3. The details of the XJTU-SY dataset.
Working ConditionRotation SpeedLoadDatasetSample NumberBearing LifetimeDivision
12100 rpm12,000 NBearing 1-11232 h 3 mtraining
Bearing 1-21612 h 41 mtraining
Bearing 1-31582 h 38 mtesting
Bearing 1-41222 h 2 mtesting
Bearing 1-55252 mtesting
22250 rpm11,000 NBearing 2-14918 h 11 mtraining
Bearing 2-21612 h 41 mtraining
Bearing 2-35338 h 53 mtesting
Bearing 2-44242 mtesting
Bearing 2-53395 h 39 mtesting
32400 rpm10,000 NBearing 3-1253842 h 18 mtraining
Bearing 3-2249641 h 36 mtraining
Bearing 3-33716 h 11 mtesting
Bearing 3-4151525 h 15 mtesting
Bearing 3-51141 h 54 mtesting
Table 4. Ablation experiments in PHM dataset.
Table 4. Ablation experiments in PHM dataset.
Metric Bearing
1-3
Bearing
1-4
Bearing
1-5
Bearing
1-6
Bearing
1-7
Bearing
2-3
Bearing
2-4
Bearing
2-5
Bearing
2-6
Bearing
2-7
Bearing
3-3
MSESACGNet0.0100.0530.0390.0420.0120.0170.0420.0660.0420.0610.078
NoAttention0.2030.1380.2120.1930.2230.0490.0630.0640.0610.0620.162
NoConv1d0.0550.0990.1600.0600.0690.1620.1100.1610.1290.0760.095
RMSESACGNet0.1010.2300.1970.2050.1080.1310.2040.2020.2050.3970.280
NoAttention0.4510.3720.4610.4390.4720.2200.2500.2530.2460.2490.403
NoConv1d0.2360.3140.4010.2450.2630.4030.3320.4020.3590.2760.309
MAESACGNet0.0410.1570.0770.0790.0220.0330.0810.0710.0830.2200.161
NoAttention0.3730.3040.3820.3590.3940.2010.1480.2250.1670.0960.368
NoConv1d0.2160.2150.2730.2030.2560.3870.2710.3750.3030.1780.205
MAPESACGNet1.3001.4615.8002.7072.52613.29014.12815.77848.944188.95211.879
NoAttention26.6163.54264.02138.05086.29889.24933.65447.49383.54211.72422.317
NoConv1d16.8092.62733.06019.33947.315175.06162.93277.439140.79978.59615.225
Table 5. Ablation experiments in the XJTU-SY dataset.
Table 5. Ablation experiments in the XJTU-SY dataset.
Metric Bearing 1-3Bearing 1-4Bearing 1-5Bearing 2-3Bearing 2-4Bearing 2-5Bearing 3-3Bearing 3-4Bearing 3-5
MSESACGNet0.0220.0280.1290.1020.2610.1160.1340.0370.249
NoAttention0.0710.0450.1510.0990.2380.1890.1500.0500.252
NoConv1d0.2820.1410.1500.2620.2920.1220.1470.0930.254
RMSESACGNet0.1470.1660.3600.3200.5110.3410.3690.1930.500
NoAttention0.2660.2120.3880.3150.4880.4350.3870.2230.502
NoConv1d0.5310.3760.3870.5120.5400.3500.3830.3040.504
MAESACGNet0.1170.0880.2060.3070.4280.2490.2560.0690.447
NoAttention0.2290.1370.1980.3010.4000.3330.2760.0980.450
NoConv1d0.4470.2740.1940.4790.4620.2970.2940.2110.447
MAPESACGNet12.9040.7143.21712.0900.8628.71417.10531.2401.251
NoAttention39.5831.9030.79812.4420.7508.62619.34153.3581.286
NoConv1d84.2174.4350.56819.2830.9989.18833.756172.8341.314
Table 6. Different models’ result in the PHM dataset.
Table 6. Different models’ result in the PHM dataset.
Metric Bearing
1-3
Bearing
1-4
Bearing
1-5
Bearing
1-6
Bearing
1-7
Bearing
2-3
Bearing
2-4
Bearing
2-5
Bearing
2-6
Bearing
2-7
Bearing
3-3
MSESACGNet0.0100.0530.0390.0420.0120.0170.0420.0660.0420.0610.078
CNN0.2760.2360.3610.3390.4010.0170.0440.0330.0490.0630.159
RNN0.0870.0790.1910.1670.0890.0980.1070.1020.1060.0520.080
LSTM0.0510.1540.0990.0880.1000.1340.1260.0800.1290.1530.232
GRU0.1560.1300.2200.2120.1220.0470.1730.0740.1500.6210.255
RMSESACGNet0.1010.2300.1970.2050.1080.1310.2040.2020.2050.3970.280
CNN0.5260.4860.6010.5830.6330.1320.2090.1820.2210.2500.399
RNN0.2950.2820.4370.4090.2990.3130.3270.3190.3260.2290.282
LSTM0.2270.3930.3150.2960.3170.3660.3540.2830.3600.3920.482
GRU0.3950.3600.4690.4610.3500.2160.4160.2720.3870.7880.505
MAESACGNet0.0410.1570.0770.0790.0220.0330.0810.0710.0830.2200.161
CNN0.4310.4010.4920.4730.5290.0790.1210.1270.1290.0940.361
RNN0.2720.2300.4050.3710.2770.3050.2970.3070.3020.1370.208
LSTM0.0820.2700.2940.2750.3080.3520.1850.2190.2150.3690.376
GRU0.3780.3150.4490.2820.3370.1630.2500.1670.2200.7440.433
MAPESACGNet1.3001.4615.8002.7072.52613.29014.12815.77848.944188.95211.879
CNN34.1284.70284.79468.833116.33134.08226.48226.33866.65310.53221.550
RNN20.3912.99964.36651.72258.369135.45877.72763.910132.20333.09214.056
LSTM4.0843.94146.74534.37264.918157.41143.52445.461107.898160.43825.883
GRU28.1074.76552.20652.41770.27172.18163.66835.833110.955364.58528.784
Table 7. Different models’ results in the XJTU-SY dataset.
Table 7. Different models’ results in the XJTU-SY dataset.
Metric Bearing 1-3Bearing 1-4Bearing 1-5Bearing 2-3Bearing 2-4Bearing 2-5Bearing 3-3Bearing 3-4Bearing 3-5
MSESACGNet0.0220.0280.1290.1020.2610.1160.1340.0370.249
CNN0.0240.0360.1510.0600.2760.1900.1610.0410.255
RNN0.2970.1220.1390.1270.2920.9840.1410.0840.225
LSTM0.2760.1360.1500.2760.2920.2600.6150.0970.147
GRU0.2760.1440.1500.1230.2920.1070.1700.3310.236
RMSESACGNet0.1470.1660.3600.3200.5110.3410.3690.1930.500
CNN0.1540.1910.3890.2440.5250.4360.4010.2030.505
RNN0.5450.3490.3730.3570.5400.3140.3750.2900.474
LSTM0.5250.3680.3870.5250.5400.5100.7840.3120.384
GRU0.5260.3800.3870.3510.5400.3310.4130.5750.486
MAESACGNet0.1170.0880.2060.3070.4280.2490.2560.0690.447
CNN0.1340.0930.2000.2280.4440.3330.2970.0770.454
RNN0.4690.2490.1940.3320.4620.2310.3110.2520.421
LSTM0.4420.2800.1940.5200.4620.4570.7300.1350.297
GRU0.4460.2900.1940.3340.4620.2460.3680.5630.433
MAPESACGNet12.9040.7143.21712.0900.8628.71417.10531.2401.251
CNN17.2190.7620.9228.6110.9298.88321.78234.9251.292
RNN85.0294.0891.62814.2170.9988.48029.475184.1341.257
LSTM84.9454.5990.56819.2320.99815.02290.75677.4181.546
GRU85.1154.7360.56814.0070.9988.64442.677425.4271.258
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xu, J.; Duan, S.; Chen, W.; Wang, D.; Fan, Y. SACGNet: A Remaining Useful Life Prediction of Bearing with Self-Attention Augmented Convolution GRU Network. Lubricants 2022, 10, 21. https://doi.org/10.3390/lubricants10020021

AMA Style

Xu J, Duan S, Chen W, Wang D, Fan Y. SACGNet: A Remaining Useful Life Prediction of Bearing with Self-Attention Augmented Convolution GRU Network. Lubricants. 2022; 10(2):21. https://doi.org/10.3390/lubricants10020021

Chicago/Turabian Style

Xu, Juan, Shiyu Duan, Weiwei Chen, Dongfeng Wang, and Yuqi Fan. 2022. "SACGNet: A Remaining Useful Life Prediction of Bearing with Self-Attention Augmented Convolution GRU Network" Lubricants 10, no. 2: 21. https://doi.org/10.3390/lubricants10020021

APA Style

Xu, J., Duan, S., Chen, W., Wang, D., & Fan, Y. (2022). SACGNet: A Remaining Useful Life Prediction of Bearing with Self-Attention Augmented Convolution GRU Network. Lubricants, 10(2), 21. https://doi.org/10.3390/lubricants10020021

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop