Next Article in Journal
Dynamic Modeling and Stability Analysis for a Spur Gear System Considering Gear Backlash and Bearing Clearance
Next Article in Special Issue
Semi-Supervised Transfer Learning Method for Bearing Fault Diagnosis with Imbalanced Data
Previous Article in Journal
A Unified Constitutive Model of Stress Relaxation of Ti-6Al-4V Alloy with Different Temperatures from Elastic to Plastic Loading
Previous Article in Special Issue
Morphological Component Analysis-Based Hidden Markov Model for Few-Shot Reliability Assessment of Bearing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cross-Domain Remaining Useful Life Prediction Based on Adversarial Training

1
Research Center for High-Speed Railway Network Management, School of Computer and Information Technology, Beijing Jiaotong University, Beijing 100044, China
2
Beijing Chenzhu Technology Co., Ltd., Beijing 100091, China
*
Author to whom correspondence should be addressed.
Machines 2022, 10(6), 438; https://doi.org/10.3390/machines10060438
Submission received: 27 April 2022 / Revised: 26 May 2022 / Accepted: 27 May 2022 / Published: 1 June 2022
(This article belongs to the Special Issue Fault Diagnosis and Health Management of Power Machinery)

Abstract

:
Remaining useful life prediction can assess the time to failure of degradation systems. Currently, numerous neural network-based prediction methods have been proposed by researchers. However, most of the work contains an implicit prerequisite: the network training and testing data have the same operating conditions. To solve this problem, an adversarial discriminative domain adaption prediction method based on adversarial training is proposed to improve the accuracy of cross-domain prediction under different working conditions. First, an LSTM feature extraction network is constructed to mine the source domain data and the target domain data for deep feature representation. Subsequently, the parameters of the target domain feature extraction network are adjusted based on the idea of adversarial training to achieve domain invariant feature mining. The proposed scheme is experimented on a publicly available dataset and achieves state-of-the-art prediction performance compared to recent unsupervised domain adaptation prediction methods.

1. Introduction

Remaining useful life (RUL) prediction, an important research component of prognostic and health management (PHM), can estimate the time to failure of complex systems and reduce maintenance costs and risks [1,2,3]. In general, RUL prediction schemes can be divided into two types: physics-based methods and data-driven methods [4].
Physics-based approaches usually use mathematical modeling methods and classical physical models to describe the health state of mechanical systems [5]. Although these methods achieve accurate prediction results and have good interpretability, they require complex domain expertise, which limits the generalizability of these methods [6]. In contrast, data-driven methods, especially neural network methods, are able to directly learn the relationship between monitoring data and RUL labels by designing a rational network structure. A large number of contributing neural network prediction schemes have been proposed and applied, e.g., Convolutional Neural Network (CNN) [7,8,9,10,11], Gated Recurrent Units (GRU) [12,13,14] and Long Short-Term Memory Networks (LSTM) [15,16,17].
However, in terms of data-driven method, it is challenging (intolerably labor-expensive and time-consuming) to collect sufficient run-to-failure data, which may reduce the prediction performance of the prognostic method. Even when enough historical data are available, the pre-trained model under one specific working condition cannot be generalized to another although they are similar [2]. To cope with the aforementioned limitations, models trained under a single operating condition should be adapted to testing data under other operating conditions, i.e., domain adaptation, a sub-problem of transfer learning [18].
In recently RUL prediction schemes, existing domain adaption methods can be mainly grouped into two categories: moment matching methods and adversarial training methods [19,20]. The former methods align feature distributions across domains by minimizing the distribution discrepancy. Zhu et al. [1] proposed transfer multiple layer perceptron (MLP) with Maximum Mean Discrepancy (MMD) loss [21]. The source and target domain training data feed to the MLP at the same time after feature extraction. By reducing the source domain regression loss and MMD loss, the domain-invariant features are obtained. Yu et al. [22] developed a domain adaptive CNN-LSTM (DACL) model with MMD to align the domain distributions and evaluated the model on C-MAPSS dataset. The latter methods achieve cross-domain prediction feature extraction by adversarial training [23]. Da costa et al. [18] proposed an RUL domain adaptation network named LSTM-DANN. Firstly, degraded pattern knowledge is learned in the source domain, and then the source domain regression model is adjusted to adapt to the target domain feature distribution with the gradient inversion layer to achieve cross-domain RUL prediction [24].
In this work, LSTM combined with a domain adaptive scheme of adversarial discriminative learning framework is proposed, LSTM-ADDA. The proposed scheme can search the domain invariant feature space in the training phase using an adversarial training approach [25] to adjust the network parameters of the LSTM extractor to achieve cross-domain RUL prediction. Compared with the latest adversarial method, LSTM-DANN, the proposed method is not required to repeatedly train the source domain regression model to adapt to different target domains. The extraction of cross-domain RUL prediction features is achieved by reversing the domain feature labels. The implementation process is relatively simple and converges faster. To verify the effectiveness of the proposed method, we conduct experiments on the C-MAPSS datasets and compare the RUL prediction results of aircraft engines with LSTM-DANN and other adaptation methods.
The contributions can be summarized as follows:
  • A novel unsupervised adversarial learning RUL prediction scheme is proposed, which can adapt to different fault modes and operating settings domain data.
  • Our experimental results show that the proposed LSTM-ADDA method achieves competitive performance compared with other unsupervised domain adaptation methods.
In Section 2, LSTM and unsupervised domain adaption methods are reviewed. In Section 3, the proposed LSTM-ADDA architecture and training process are detailed introduced. In Section 4, we briefly describe the dataset preprocess method and list the hyperparameter selection in experiments. In Section 5, the superiority of the proposed method is proved through two sets of comparative experimental results. The conclusion and future research are drawn in Section 6.

2. Background

In this section, we review the problem definition and generalized adversarial learning framework. Subsequently, the adversarial discriminative domain adaptation (ADDA) method and the detailed LSTM structure are introduced.

2.1. Problem Definition

In previous neural network-based prediction methods, training and testing data are usually acquired under the same operating conditions, as shown in Figure 1a. However, it is difficult to obtain data under the same operating conditions in the actual data acquisition, resulting in the limitation of the method’s generalizability. Suppose there are two datasets for remaining useful life prediction under different operating conditions, one of which has sufficient training and testing data (source domain dataset), and the other has only unlabeled training and testing data (target domain dataset). We define this application scenario as an unsupervised domain adaption scenario [26], i.e., with labeled source domain training data and unlabeled target domain training data to predict the target testing data RUL. This paper aims to adaptively implement the prediction of the target domain testing data by transferring the knowledge of the source domain regression model, as shown in Figure 1b. Suppose the source domain data is denoted as D S = { ( x S i , y S i ) } i = 1 N S , containing N S training samples, where x S i is a subset of source domain feature space χ S . T i denotes the length of time and f S denotes the feature dimensions, i.e., x S i = { x t i } t = 1 T i T i × f S . Moreover, y S i = { y t i } t = 1 T i 0 T i × 1 represents the RUL vector of x S i . x t i 1 × f S and y t i 0 1 denotes the observed feature vector and RUL labels at time t i , respectively. Besides this, we assume a target domain D T = { x T i } i = 1 N T , where x T i χ T and x T i = { x t i } t = 1 T i T i × f T . Specially, there are no available RUL labels.
We assume that there are somewhat similar or invariant parts between the deep feature representation of source and target domain mapping. In the model training stage, source domain data and RUL value are available, i.e., { ( x S i , y S i ) } i = 1 N S . The target domain data can only use the training monitoring data without RUL labels, { x T i } i = 1 N T . It is aimed to adjust the domain feature extractation network g by both source and target training data to perform the target testing sample RUL transfer regression, i.e y t r u e i y p r e d i c t i = g ( x T i ) .

2.2. Generalized Adversarial Adaptation

Tzeng et al. [25] presented a general adversarial adaptive framework, and pointed out that unsupervised domain adaptation could be achieved by reducing the source domain classification loss and alternately minimizing the domain discriminator loss and the distribution discrepancy between the source and target mapping distributions.
Adversarial adaptive methods aim to minimize the distance between the source and target mapping distributions by regularizing the learning of the source and target mappings, M S and M T . Once the adversarial training is finished, the domain-invariant feature representations are found. Source classification block network can directly receive the target domain feature representation for subsequent RUL prediction, C = C S = C T .
The source supervised classification network optimize object function is shown as follows:
m i n M S , C L c l s ( x S i , y S i ) = k = 1 K 1 ( k = y S i ) l o g C ( M S ( x S i ) )
In adversarial adaptation training part, the domain discriminator, D, is used to distinguishing the source of deep domain features. Similarly, the network parameters of the feature extractor are optimized using the common classification loss, L a d v D ( x S i , x T i , M S , M T ) .
m i n D L a d v D ( x S i , x T i , M S , M T ) = E x S i χ S [ l o g D ( M S ( x S i ) ) ] E x T i χ T [ l o g ( 1 D ( M T ( x T i ) ) ) ]
Finally, the label of the target domain feature representation is reversed to adjust the weights of the target domain mapping structure, achieving the source domain degradation knowledge transfer. The reversed label training loss is denoted as follows, L a d v M ( x S i , x T i , D ) :
m i n M T L a d v M ( x S i , x T i , D ) = E x T i χ T [ l o g D ( M T ( x T i ) ) ]

2.3. Adversarial Discriminative Domain Adaptation

Adversarial discriminative domain adaptation (ADDA), based on general adversarial adaptive framework, selects the discriminative base model, untied weights and standard GAN loss.
For the mapping optimization constraints, M T uses the M S model architecture and weights pre-trained in the source domain during the M T initialization. In adversarial training, the M S model is fixed, and the source and target data are input into the M S and M T , respectively, to obtain feature representations for domain discriminative and M T weights adjustment. The asymmetric feature representation obtained by unshared weight mappings makes the learning paradigm more flexible.
In adversarial adaptation training, the optimization process is split into two independent objectives, domain discriminator and target feature mapping, and alternately minimizes their loss functions. The L a d v D of domain discriminator remains unchanged, and the target mapping uses the standard loss function of the inverted labels [23].
Similar to the training method of GAN, the mapping feature of the source domain can be labeled as ‘1’ and the target domain as ‘0’, which are input into the discriminator for domain discrimination. During the adversarial training, the mapping feature label of the target is inverted as ‘1’, and input to the discriminator for adversarial training with an inverting the labels manner [26].
The ADDA method divides the optimization process into two stages. First, the M S and C are trained using labeled source domain data. After that, the M S is fixed and the M T is initialized, and the unlabeled target domain data is used to alternately optimize the L a d v D and L a d v M to adjust the model parameters of the M T .

2.4. Long Short-Term Memory Neural Network

The temporal feature extractor in our methodology is the LSTM network [27], which is a subclass of recurrent neural networks (RNNs). LSTM is one of the most exploited RNN versions because of the capability of detecting long- and short-term dependencies and relaxes the gradient vanishing or exploding problems. As shown in Figure 2, the LSTM cell uses input gate, forget gate and output gate to update current cell state c t and output hidden state vector h t .
Mathematically, the computation of the LSTM cell can be described as follows:
i t = σ ( W i x x t + W i h H t 1 + B i ) g t = ( W g x x t + W g h H t 1 + B g ) o t = σ ( W o x x t + W o h H t 1 + B o ) c t = g t i t + c t 1 f t f t = σ ( W f x x t + W f h H t 1 + B f ) h t = ( s t ) o t
where W g x , W i x , W f x , W o x are weight values between input layer x t and hidden layer H t at time t; W g h , W i h , W f h , W o h are hidden layer weight values between time t and t 1 ; B g , B i , B f , B o are bias of the input layer, input gate, forget gate and output gate, respectively. H t 1 is the hidden layer output at previous time. g t , i t , f t , o t are the output values. c t represents the current LSTM cell state. ⊗ is the pointwise multiplication operator. ϕ , σ represent different activation functions.

3. Methodology

In this section, we detail the proposed method architecture and training processes. Adversarial discriminative domain adaptation based on LSTM, referred as LSTM-ADDA and depicted in Figure 3, can be decomposed into three modules: feature extractor, regression and domain discriminator.
The temporal feature extractor, a combination of LSTMs, establishes a mapping between the input data and the hidden feature state h t h . Later, a dense layer transforms the LSTM hidden vector to deep feature representation f t f , which will be used as temporal feature space for subsequent regression tasks. The parameters of feature extractor is denoted as Θ f , i.e., f = g f ( x S i ; Θ f ) .
Both regression and domain discriminator modules consist of simple fully connected layers. We regard the RUL prediction as a regression task. The mean absolute error (MAE) function is selected as the loss function of the source domain regression to minimization. The parameters in the regression module are denoted as Θ y ^ , i.e., y ^ = g y ^ ( g f ( x S i ; Θ f ) ; Θ y ^ ) . The LSTM-ADDA optimization equations applied in regression scenarios are described as follows:
m i n M S , y ^ L M A E ( x S i , y S i ) = y ^ y S i = g y ^ ( g f ( x S i ; Θ f ) ; Θ y ^ ) y S i m i n D L a d v D ( x S i , x T i , M S , M T ) = E x S i χ S [ l o g D ( M S ( x S i ) ) ] E x T i χ T [ l o g ( 1 D ( M T ( x T i ) ) ) ] m i n M T L a d v M ( x S i , x T i , D ) = E x T i χ T [ l o g D ( M T ( x T i ) ) ]
The training process of the proposed method can be summarized into three phases, as shown in Figure 3 and Figure 4.
In stage I, the source domain training data and training labels are input to a source domain regression network consisting of source domain mapping structure and regressor for standard supervised learning training, as shown by the blue arrows in Figure 3. After the source domain training is completed, the network parameters of the source domain mapping structure and regressor will be fixed and will not change with subsequent training. The purpose of stage I is to learn the deep feature representation and degradation patterns of the source domain data from the source domain data.
In stage II, the target domain mapping structure first replicates the structure and parameters of the source domain mapping, which is used as the training starting point for the target domain mapping. Then, the source domain training data and the unlabeled training data of the target domain are input to the source domain mapping and target domain mapping structures to obtain the corresponding deep feature representations, respectively. Subsequently, the feature representations of the two domains are input to the discriminator for adversarial training. The adversarial training is achieved by inverting the domain labels of the two domains. By alternately inverting the domain labels, the parameters of the target domain mapping structure are adjusted to automatically adapt to the data distribution of the target domain, as shown by the orange arrows in Figure 3. Stage II aims to adjust the feature mapping of the target domain to obtain an adaptive target domain deep feature representation for subsequent prediction.
In stage III, the trained target domain mapping structure is combined with the regressor to achieve cross-domain RUL prediction for the target domain testing data, as shown by the green arrows in Figure 3.
For researchers to understand and reproduce the ADDA method, the reference code can be found in the following Github link: https://github.com/acmllearner/pytorch-adda, (accessed on 16 May 2022).

4. Experiment Pre-Settings

In this section, dataset description, data preprocessing, evaluation metric and hyperparameter selection are introduced in this section.

4.1. Dataset Description

The Commercial Modular Aero-Propulsion System Simulation (C-MAPSS) dataset contains four subsets, as shown in Table 1. Each subset has different fault modes and operating conditions. Each subset contains a training and testing set. Each engine starts with various initial wear but regarded as healthy at the beginning of the recording cycle. Every recording cycle is composed of three operational settings and 21 sensor readings. As the recording cycle increases, each engine performance will gradually degrade until it loses its function [28].

4.2. Data Preprocessing

4.2.1. Time Windows Processing

The sliding time window method is adopted to obtain the same time-window block data. Suppose the running time of an engine x i is T i , i.e., x i = { x t i } t = 1 T i . The sequential input x i divided by the time window T w is represented as { ( x t T w i , , x t 1 i ) } t = T w + 1 T i . If T i < T w , zero-padding is applied to fulfill the x i , i.e., x i = { 0 , 0 , , 0 T w T i , x 1 i , x 2 i , , x T i i T i } . This ensures that each engine unit can generate at least two samples. We use the same length of time window to preprocess the source and target domain training sets. To make full use of data, in adversarial training stage, the domain training dataset containing the smaller number of samples is oversampled.

4.2.2. Data Normalization

The min-max normalization is used to scale the subset data (including operating conditions and sensor data) and RUL value to the (0–1) range:
x ˜ t i , j = x t i , j m i n ( x j ) m a x ( x j ) m i n ( x j )
where x ˜ t i , j denotes the original i t h engine unit of the j t h feature at record cycle t.
To demonstrate the feature shift between different domains, we perform data distribution statistics on the last cycle of each engine unit in each subset of C-MAPSS. We selected two normalized sensor values for visualization. As shown in Figure 5, we can see that the same operating condition has similar feature distributions for different fault modes. On the contrary, the feature distribution of the same fault mode with the different operating conditions has a clear discrepancy.
As in the previous work [29,30], RUL can be estimated as a constant value in normal operating conditions. Therefore, a piecewise linear degradation model is used to describe the RUL curve of the observation engine. This paper selects R e = 125 cycles as the constant RUL before failure and applies it to the training and testing set.

4.3. Performance Metrics

We use two metrics to measure the LSTM-ADDA prognostic performance. The root mean squared error (RMSE), shown in (7), is selected to evaluate the error between the predicted RUL and the actual RUL. We also use the score metric [28], shown in (8), to evaluate our proposed method. The score metric penalizes positive prediction errors more than negative prediction errors because it is more meaningful on maintenance policies to predict failures as early as possible.
R M S E = 1 n i = 1 n ( R U L i ^ A c t R U L i ) 2
S c o r e = i = 1 n e y ^ i y i 13 1 if y ^ i y i < 0 i = 1 n e y ^ i y i 10 1 if y ^ i y i 0

4.4. Hyperparameter Selection

To obtain better adversarial learning performance, we minimize the regression loss in source domain. The learning rate of Adam optimizer is set as 0.0001 ( λ S ). Besides this, in the adversarial learning process, we reduce the domain classification accuracy of the domain discriminator, i.e., find the domain-invariant representation. For a fair comparison with the adversarial learning method LSTM-DANN [18], we select the same network parameters as the LSTM-DANN method, as shown in Table 2.

5. Experimental Results

In this section, two comparative experiments are conducted to illustrate the prognostic performance of the proposed method. First, the comparison with the non-adapted models proves that the proposed method can effectively improve the engine RUL prediction under different operating conditions and fault modes. Besides this, compared with other unsupervised transfer learning methods: CORrelation (CORAL) [31], Transfer Component Analysis (TCA) [32] and LSTM-DANN [18].
In the experiment, each training subset of the C-MAPSS dataset is regarded as the source data, and the remaining testing data are regarded as the target data. Therefore, 12 different experiments were performed, and the mean and stand deviations of each experiment result were recorded. All experiments are performed on an Intel Core i7 8th generation processor with 24GB RAM and CPU. We implement our method by using Python 3.6 and Pytorch 1.4.0.

5.1. Non-Adapted Models under Domain Shift

To prove the effectiveness of the proposed adversarial adaption method, we input the target testing data into the source domain model trained on the source domain training data as an experiment of the non-adapted method (source-only). Moreover, we input the target testing data into the target model trained with target domain training data, which represents the ideal situation when the target domain has sufficient training data (target-only). We use each subset of C-MAPSS as the source domain to analyze and discuss the experimental results separately. The results of cross-domain prediction experiments are shown in Figure 6.
Source FD001: FD001 has single fault mode and operating condition, so it is difficult to learn the target domain with complex operating conditions and fault modes, causing the RMSE and prediction scores of LSTM-ADDA higher than the other experiment pairs. As shown in Figure 6a,c, the prediction curve of the proposed model is similar to the source-only waveform and is closer to the target-only curve, indicating that the proposed method can improve the adaptability of the non-adapted model. Compared with the target domain FD003, which has a low distribution discrepancy between the two subsets, the prediction curve obtained in Figure 6b is closer at the end stage of the engine.
Source FD002: Compared with FD001, the operating condition of FD002 is six. When the more complex operating condition dataset as the source domain. as shown in Figure 6d, the proposed model gives better prediction results. When the target domains are FD003 and FD004, containing two fault modes, the prediction effect of FD004 with less domain shift is closer to the target-only curve in Figure 6e. Although the operating conditions and fault modes of FD003 are not the same as FD002, the proposed model can also give better prediction results than the non-adapted models.
Source FD003: Under the same operating conditions, FD003 with two fault modes achieved a lower RMSE when performing adversarial training to the target domain FD001 with one fault mode. Besides this, due to the low distribution discrepancy, Figure 6g shows that the prediction curve obtained by the proposed method also fits the RUL curve of the engine. For FD002 and FD004 with six operating conditions, LSTM-ADDA also gives better prediction results than non-adapted models.
Source FD004: FD004 is containing six operating conditions and two fault modes, which is the most complex subset of the C-MAPSS dataset. When training with the remaining target domain datasets, as shown in Table 3, lower RMSE and prediction scores are obtained. In Figure 6j,l, we can observe that the source-only curve has a large deviation from the RUL curve. For FD002 with a low domain shift, as shown in Figure 6k, the proposed method shows a better fit with the RUL curve.
In summary, the percentage changes of average RMSE and scoring performance between the proposed method and the non-adapted case (source-only) are listed in Table 3. It can be seen that all the prediction performances of the proposed LSTM-ADDA are higher than that of the non-adapted models. Although source-only models can get reasonable performance when the domain shift is small, but the proposed method can further improve the effect. Moreover, when the operating conditions and fault modes of the source domain are complex and the target domain is simple, the LSTM-ADDA model can obtain a better prediction effect. That is to say, it is easier to transfer training from complex situations to simple situations, and it is more difficult to learn from simple situations to complex situations.

5.2. Domain Adaptation Methods Comparison

Two different unsupervised domain adaption methods are used to compare with the proposed method in this section, Transfer Component Analysis (TCA) [32] and CORrelation ALignment (CORAL) [33]. Besides this, the average RMSE and Score are compared with the adversarial method LSTM-DANN.
The distribution of the two domains are mapped together to a high-dimensional Reproducing Kernel Hilbert Space (RKHS). In this space, minimize the MMD between the source and target distribution, while retaining their respective internal attributes.
Different from TCA, CORAL aligns the second-order statistics of the source and target distributions by constructing a differentiable loss function that minimizes the difference between the source and target correlations—the CORAL loss [31].
In the LSTM-DANN method, the source domain shares the network weights with the target domain. When predicting a new target domain, the source domain regression model must be repeatedly trained to extract the degradation knowledge in the source domain. Unlike the LSTM-DANN method, the proposed scheme is not required to train the source-domain regression model repeatedly, and the extraction of cross-domain features can be achieved directly by adversarial training in stage II. Moreover, the significant difference between the proposed and DANN schemes is the cross-domain degradation feature training. The LSTM-DANN method uses the gradient inversion layer, while the LSTM-ADDA method uses the reversed feature domain label. The proposed scheme is simpler to implement and converges faster.
The comparison of the evaluation metrics is shown in Table 4. We can observe that the mean of RMSE metric is close to LSTM-DANN and overall better than the TCA and CORAL models. In addition, we found that the score metric improved by nearly 45%, indicating that the proposed method can achieve better prediction results in a timely manner while maintaining a smaller RMSE metric, which is more in line with the need for advanced prediction in practical maintenance.

6. Conclusions and Future Work

In this paper, a novel domain adaptation RUL prediction scheme is proposed by combining the LSTM network and adversarial discriminative domain adaptation framework (LSTM-ADDA). The prediction effect of the proposed model is verified on the C-MAPSS dataset. The LSTM feature extractor network parameters are adjusted by the target domain training data to find the domain invariant feature representation.
The proposed method uses the source domain data to obtain source domain degradation knowledge. After that, the target domain data without RUL label is combined for adversarial learning. Finally, the target feature extractor adjusted by adversarial training combines with the source domain regression module to implement the target domain RUL adaption prediction.
To illustrate the effectiveness of the LSTM-ADDA method, we perform two comparative experiments. The experimental results show that the prediction result of the proposed method is better than other methods. In future work, we will try to remove the target domain data from the training process, and use a generalized model combined with multi-source data to directly estimate the RUL of the target domain.

Author Contributions

Conceptualization, Y.D., J.X. and H.L.; methodology, Y.D. and H.L.; validation, Y.D., H.L. and J.Z.; formal analysis, J.X.; investigation, Y.D.; resources, H.L.; data curation, Y.D.; writing—original draft preparation, Y.D.; writing—review and editing, Y.D.; supervision, H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was financially supported by the National Key Research and Development Program of China (Grant number 2019YFB2102500) and Technology Innovation Project of Shenhua Group (Grant No. SHGF-17-56).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The publicly available dataset used in this study is available at link: https://data.nasa.gov/dataset/C-MAPSS-Aircraft-Engine-Simulator-Data/xaut-bemq (accessed on 16 May 2022). The reference code for this study is available in Github: https://github.com/acmllearner/pytorch-adda (accessed on 16 May 2022).

Conflicts of Interest

The authors declare that they have no known competing financial interest or personal relationship that could have appeared to influence the work reported in this paper.

References

  1. Zhu, J.; Chen, N.; Shen, C. A new data-driven transferable remaining useful life prediction approach for bearing under different working conditions. Mech. Syst. Signal Process. 2020, 139, 106602. [Google Scholar] [CrossRef]
  2. Lei, Y.; Li, N.; Guo, L.; Li, N.; Yan, T.; Lin, J. Machinery health prognostics: A systematic review from data acquisition to rul prediction. Mech. Syst. Signal Process. 2018, 104, 799–834. [Google Scholar] [CrossRef]
  3. Zhou, T.; Han, T.; Droguett, E. Towards trustworthy machine fault diagnosis: A probabilistic Bayesian deep learning framework. Reliab. Eng. Syst. Saf. 2022, 224, 108525. [Google Scholar] [CrossRef]
  4. Fan, Y.; Nowaczyk, S.; Rognvaldsson, T. Transfer learning for remaining useful life prediction based on consensus self-organizing models. Reliab. Eng. Syst. Saf. 2020, 203, 107098. [Google Scholar] [CrossRef]
  5. Li, N.; Lei, Y.; Lin, J.; Ding, S.X. An improved exponential model for predicting remaining useful life of rolling element bearings. IEEE Trans. Ind. Electron. 2015, 62, 7762–7773. [Google Scholar] [CrossRef]
  6. Cofre-Martel, S.; Droguett, E.L.; Modarres, M. Uncovering the underlying physics of degrading system behavior through a deep neural network framework: The case of remaining useful life prognosis. arXiv 2020, arXiv:2006.09288. [Google Scholar]
  7. Zhu, J.; Chen, N.; Peng, W. Estimation of bearing remaining useful life based on multiscale convolutional neural network. IEEE Trans. Ind. Electron. 2018, 66, 3208–3216. [Google Scholar] [CrossRef]
  8. Yang, B.; Liu, R.; Zio, E. Remaining useful life prediction based on a double-convolutional neural network architecture. IEEE Trans. Ind. Electron. 2019, 66, 9521–9530. [Google Scholar] [CrossRef]
  9. Xu, X.; Wu, Q.; Li, X.; Huang, B. Dilated convolution neural network for remaining useful life prediction. J. Comput. Inf. Sci. Eng. 2020, 20, 021004. [Google Scholar] [CrossRef]
  10. Peng, W.; Ye, Z.-S.; Chen, N. Bayesian deep-learning-based health prognostics toward prognostics uncertainly. IEEE Trans. Ind. Electron. 2019, 67, 2283–2293. [Google Scholar] [CrossRef]
  11. Wang, B.; Lei, Y.; Li, N.; Wang, W. Multi-scale convolutional attention network for predicting remaining useful life of machinery. IEEE Trans. Ind. Electron. 2020, 68, 7496–7504. [Google Scholar] [CrossRef]
  12. Zhao, R.; Wang, D.; Yan, R.; Mao, K.; Shen, F.; Wang, J. Machine health monitoring using local feature-based gated recurrent unit networks. IEEE Trans. Ind. Electron. 2017, 65, 1539–1548. [Google Scholar] [CrossRef]
  13. Li, X.; Jiang, H.; Xiong, X.; Shao, H. Rolling bearing health prognosis using a modified health index based hierarchical gated recurrent unit network. Mech. Mach. Theory 2019, 133, 229–249. [Google Scholar] [CrossRef]
  14. Ren, L.; Cheng, X.; Wang, X.; Cui, J.; Zhang, L. Multi-scale dense gate recurrent unit networks for bearing remaining useful life prediction. Future Gener. Comput. Syst. 2019, 94, 601–609. [Google Scholar] [CrossRef]
  15. Zhang, Y.; Xiong, R.; He, H.; Pecht, M.G. Long short-term memory recurrent neural network for remaining useful life prediction of lithium-ion batteries. IEEE Trans. Veh. Technol. 2018, 67, 5695–5705. [Google Scholar] [CrossRef]
  16. Wu, Y.; Yuan, M.; Dong, S.; Lin, L.; Liu, Y. Remaining useful life estimation of engineered systems using vanilla lstm neural networks. Neurocomputing 2018, 275, 167–179. [Google Scholar] [CrossRef]
  17. TV, V.; Malhotra, P.; Vig, L.; Shroff, G. Data-driven prognostics with predictive uncertainty estimation using ensemble of deep ordinal regression models. arXiv 2019, arXiv:1903.09795. [Google Scholar]
  18. da Costa, P.R.D.O.; Akcay, A.; Zhang, Y.; Kaymak, U. Remaining useful lifetime prediction via deep domain adaptation. Reliab. Syst. Saf. 2020, 195, 106682. [Google Scholar] [CrossRef] [Green Version]
  19. Han, T.; Liu, C.; Yang, W.; Jiang, D. A novel adversarial learning framework in deep convolutional neural network for intelligent diagnosis of mechanical faults. Knowl.-Based Syst. 2019, 165, 474–487. [Google Scholar] [CrossRef]
  20. Wang, X.; Long, M.; Wang, J.; Jordan, M.I. Transferable calibration with lower bias and variance in domain adaptation. arXiv 2020, arXiv:2007.08259. [Google Scholar]
  21. Ghifary, M.; Kleijn, W.B.; Zhang, M. Domain adaptive neural networks for object recognition. In Pacific Rim International Conference on Artificial Intelligence; Springer: Cham, Switzerland, 2014; pp. 898–904. [Google Scholar]
  22. Yu, S.; Wu, Z.; Zhu, X.; Pecht, M. A domain adaptive convolutional lstm model for prognostic remaining useful life estimation under variant conditions. In Proceedings of the 2019 Prognostics and System Health Management Conference (PHM-Paris), Paris, France, 2–5 May 2019; pp. 130–137. [Google Scholar]
  23. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Neural Information Processing Systems, Cambridge, MA, USA, 8–13 December 2014; MIT Press: Massachusetts, MA, USA, 2014; pp. 2672–2680. [Google Scholar]
  24. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M.; Lempitsky, V. Domain-adversarial training of neural networks. J. Mach. Learn. Res. 2016, 17. [Google Scholar]
  25. Tzeng, E.; Hoffman, J.; Saenko, K.; Darrell, T. Adversarial discriminative domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7167–7176. [Google Scholar]
  26. Wilson, G.; Cook, D.J. A survey of unsupervised deep domain adaptation. ACM Trans. Intell. Syst. Technol. (TIST) 2020, 11, 1–46. [Google Scholar] [CrossRef]
  27. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  28. Saxena, A.; Goebel, K.; Simon, D.; Eklund, N. Damage propagation modeling for aircraft engine run-to-failure simulation. In Proceedings of the IEEE International Conference on Prognostics and Health Management, Denver, CO, USA, 6–9 October 2008; pp. 1–9. [Google Scholar]
  29. Cai, H.; Jia, X.; Feng, J.; Li, W.; Pahren, L.; Lee, J. A similarity based methodology for machine prognostics by using kernel two sample test. ISA Trans. 2020, 103, 112–121. [Google Scholar] [CrossRef]
  30. Yu, W.; Kim, I.Y.; Mechefske, C. An improved similarity-based prognostic algorithm for rul estimation using an rnn autoencoder scheme. Reliab. Eng. Syst. Saf. 2020, 199, 106926. [Google Scholar] [CrossRef]
  31. Sun, B.; Saenko, K. Deep coral: Correlation alignment for deep domain adaptation. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 443–450. [Google Scholar]
  32. Pan, S.J.; Tsang, I.W.; Kwok, J.T.; Yang, Q. Domain adaptation via transfer component analysis. IEEE Trans. Neural Netw. 2010, 22, 199–210. [Google Scholar] [CrossRef] [Green Version]
  33. Sun, B.; Feng, J.; Saenko, K. Return of frustratingly easy domain adaptation. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar]
Figure 1. (a) Traditional neural network prediction scenarios. (b) Unsupervised domain adaption scenarios.
Figure 1. (a) Traditional neural network prediction scenarios. (b) Unsupervised domain adaption scenarios.
Machines 10 00438 g001
Figure 2. The detail structure of LSTM cell.
Figure 2. The detail structure of LSTM cell.
Machines 10 00438 g002
Figure 3. The network structure of the proposed LSTM-ADDA method. The data mapping structure consists mainly of the LSTM network, and the output of the last time step is input to the dense layer to obtain the domain deep feature representation. The regressor and the domain discriminator consist of fully-connected network for obtaining the RUL prediction and domain adversarial training, respectively.
Figure 3. The network structure of the proposed LSTM-ADDA method. The data mapping structure consists mainly of the LSTM network, and the output of the last time step is input to the dense layer to obtain the domain deep feature representation. The regressor and the domain discriminator consist of fully-connected network for obtaining the RUL prediction and domain adversarial training, respectively.
Machines 10 00438 g003
Figure 4. An overview of the adversarial discrimination domain adaptation method based on LSTM (LSTM-ADDA). Stage I: Supervised learning of source domain training data and source domain testing data for degradation pattern extraction. Stage II: Adversarial learning of source domain training data and target domain training data for domain-invariant feature extraction. Stage III: RUL prediction and result evaluation of target domain testing data. Dashed lines indicate fixed network parameters.
Figure 4. An overview of the adversarial discrimination domain adaptation method based on LSTM (LSTM-ADDA). Stage I: Supervised learning of source domain training data and source domain testing data for degradation pattern extraction. Stage II: Adversarial learning of source domain training data and target domain training data for domain-invariant feature extraction. Stage III: RUL prediction and result evaluation of target domain testing data. Dashed lines indicate fixed network parameters.
Machines 10 00438 g004
Figure 5. Visualization of data distribution discrepancy between C-MAPSS subsets.
Figure 5. Visualization of data distribution discrepancy between C-MAPSS subsets.
Machines 10 00438 g005
Figure 6. RUL predictions of the Target-only, Source-only and LSTM-ADDA models for one target testing engine.
Figure 6. RUL predictions of the Target-only, Source-only and LSTM-ADDA models for one target testing engine.
Machines 10 00438 g006
Table 1. The C-MAPSS datasets.
Table 1. The C-MAPSS datasets.
Data SubsetFD001FD002FD003FD004
Training Engine Units (N)100260100249
Testing Engine Units100259100248
Operating Conditions (OC)1616
Fault Modes (FM)1122
Table 2. Hyperparameter settings for each group of cross-domain prediction.
Table 2. Hyperparameter settings for each group of cross-domain prediction.
SourceTargetLSTM Layers,fRegression Layers,Discriminator Layers,BatchOptimizer, λ T , λ D
DomainDomain(Units),(Dropout)(Units)(Units),(Dropout)(Units),(Dropout)Size
FD001FD0021,(128),0.5(64)1,(32),0.31,(32),0.3256SGD,0.0001,0.0001
FD001FD0031,(128),0.5(64)1,(32),0.31,(32),0.3256SGD,0.0001,0.0001
FD001FD0041,(128),0.7(64)2,(32,32),0.31,(32),0.3256SGD,0.0001,0.0001
FD002FD0011,(64),0.1(64)1,(32),0.02,(16,16),0.1512SGD,0.0001,0.0001
FD002FD0031,(64),0.1(512)2,(64,32),0.02,(64,32),0.1256Adam,0.01,0.01
FD002FD0042,(32,32),0.1(32)1,(32),0.01,(16),0.1256SGD,0.0001,0.0001
FD003FD0012,(64,32),0.3(128)2,(32,32),0.12,(32,32),0.1256SGD,0.1,0.1
FD003FD0022,(64,32),0.3(64)2,(32,32),0.12,(32,32),0.1256Adam,0.1,0.1
FD003FD0042,(64,32),0.3(64)2,(32,32),0.12,(32,32),0.1256Adam,0.1,0.1
FD004FD0011,(100),0.5(30)1,(20),0.01,(20),0.1512SGD,0.0001,0.0001
FD004FD0021,(100),0.5(30)1,(20),0.01,(20),0.1512SGD,0.0001,0.0001
FD004FD0031,(100),0.5(30)1,(20),0.01,(20),0.1512SGD,0.0001,0.0001
Table 3. Score and RMSE comparsion between Source-Only, Target-Only and LSTM-ADDA. ( Δ % denotes the percentage of evaluation metric improvement. ± denotes the standard deviation).
Table 3. Score and RMSE comparsion between Source-Only, Target-Only and LSTM-ADDA. ( Δ % denotes the percentage of evaluation metric improvement. ± denotes the standard deviation).
SourceTargetSource-OnlyLSTM-ADDASource-OnlyLSTM-ADDATarget-Only
ScoreScoreRMSERMSE(Δ%)RMSE
FD001FD002672,066.60 ± 98,239.2854,009.20 ± 9943.4879.63 ± 1.9449.60(−37.71%) ± 1.7718.2 ± 0.31
FD001FD00330,523.80 ± 6921.484209.80 ± 604.8856.76 ± 2.3935.95(−36.66%) ± 1.7715.57 ± 1.09
FD001FD004417,110.20 ± 64,229.9944,055.20 ± 8788.5575.58 ± 2.2950.76(−32.84%) ± 2.4422.5 ± 0.66
FD002FD001160,408.40 ± 84,158.7912,271.80 ± 10,575.7870.30 ± 10.9336.71(−47.79%) ± 2.3516.15 ± 0.19
FD002FD003206,608.40 ± 58,285.757374.20 ± 1940.6475.11 ± 5.2540.03(−46.70%) ± 3.1915.57 ± 1.09
FD002FD00421,064.20 ± 9648.8019,085.60 ± 7653.3436.66 ± 4.2431.68(−13.58%) ± 1.3422.5 ± 0.66
FD003FD001591,873.60 ± 133,669.194644.40 ± 2542.3054.68 ± 1.9324.96(−54.35%) ± 2.5116.15 ± 0.19
FD003FD002384,836.40 ± 72,675.4933,100.60 ± 34,567.3070.49 ± 3.1044.17(−37.34%) ± 0.6918.2 ± 0.31
FD003FD004400,365.60 ± 45,021.7225,577.00 ± 14,002.3971.11 ± 3.2644.34(−37.64%) ± 1.2422.5 ± 0.66
FD004FD001172,965.00 ± 140,697.9138,800.40 ± 23,873.1572.68 ± 14.9927.72(−61.87%) ± 1.8116.15 ± 0.19
FD004FD002104,305.40 ± 111,864.988588.20 ± 4675.5528.65 ± 3.4423.01(−19.68%) ± 1.5218.2 ± 0.31
FD004FD00393,694.00 ± 91,901.7210,711.40 ± 2832.4059.98 ± 14.5532.41(−45.96%) ± 1.9015.57 ± 1.09
Table 4. Score and RMSE comparsion of unsupervised-domain adaption methods. ( Δ % denotes the percentage of evaluation metric improvement. ± denotes the standard deviation).
Table 4. Score and RMSE comparsion of unsupervised-domain adaption methods. ( Δ % denotes the percentage of evaluation metric improvement. ± denotes the standard deviation).
SourceTargetTCA−DNNCORAL−DNNLSTM−DANNLSTM−ADDALSTM−DANNLSTM−ADDA
RMSERMSERMSERMSEScoreScore
FD001FD00290.0 ± 2.977.5 ± 4.648.6 ± 6.849.60 ± 1.7793,841 ± 55,49354,009.2 ± 9943.5
FD001FD003116.1 ± 1.069.6 ± 5.245.9 ± 3.635.95 ± 1.7727,005 ± 12,3854209.8 ± 604.9
FD001FD004113.8 ± 6.984.6 ± 7.043.8 ± 4.150.76 ± 2.4457,044 ± 60,16044,055.2 ± 8788.6
FD002FD00185.6 ± 5.580.9 ± 9.428.1 ± 5.036.71 ± 2.358411 ± 11,85512,271.8 ± 10,575.8
FD002FD003111.5 ± 7.279.8 ± 10.137.5 ± 1.540.03 ± 3.1917,406 ± 57027374.2 ± 1940.6
FD002FD00494.4 ± 6.743.6 ± 3.631.8 ± 1.631.68 ± 1.3466,305 ± 14,72319,085.6 ± 7653.3
FD003FD00190.5 ± 4.626.5 ± 1.931.7 ± 9.424.96 ± 2.515113 ± 48654644.4 ± 2542.3
FD003FD00280.8 ± 4.375.6 ± 9.544.6 ± 1.244.17 ± 0.6937,297 ± 15,57833,100.60 ± 34,567.30
FD003FD004102.6 ± 3.277.2 ± 9.147.9 ± 5.844.34 ± 1.24141,117 ± 66,21825,577.0 ± 14,002.3
FD004FD00185.6 ± 5.094.0 ± 8.831.5 ± 2.427.72 ± 1.817586 ± 273538,800.4 ± 23,873.2
FD004FD00280.8 ± 5.830.9 ± 1.424.9 ± 1.823.01 ± 1.5217,001 ± 12,9278588.20 ± 4675.55
FD004FD003102.9 ± 2.768.6 ± 11.227.8 ± 2.732.41 ± 1.95941 ± 179110,711.4 ± 2832.4
Average 96.5567.4037.0136.79 (−0.60%)40,338.9221,868.98 (−45.79%)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Duan, Y.; Xiao, J.; Li, H.; Zhang, J. Cross-Domain Remaining Useful Life Prediction Based on Adversarial Training. Machines 2022, 10, 438. https://doi.org/10.3390/machines10060438

AMA Style

Duan Y, Xiao J, Li H, Zhang J. Cross-Domain Remaining Useful Life Prediction Based on Adversarial Training. Machines. 2022; 10(6):438. https://doi.org/10.3390/machines10060438

Chicago/Turabian Style

Duan, Yuhang, Jie Xiao, Honghui Li, and Jie Zhang. 2022. "Cross-Domain Remaining Useful Life Prediction Based on Adversarial Training" Machines 10, no. 6: 438. https://doi.org/10.3390/machines10060438

APA Style

Duan, Y., Xiao, J., Li, H., & Zhang, J. (2022). Cross-Domain Remaining Useful Life Prediction Based on Adversarial Training. Machines, 10(6), 438. https://doi.org/10.3390/machines10060438

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop