Next Article in Journal
Use of New and Light Materials in Automotive Engineering for Towing System
Previous Article in Journal
POSS and PAG Dual-Containing Chemically Amplified Photoresists by RAFT Polymerization for Enhanced Thermal Performance and Acid Diffusion Inhibition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Method for Remaining Useful Life Prediction of Turbofan Engines Combining Adam Optimization-Based Self-Attention Mechanism with Temporal Convolutional Networks

1
Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650504, China
2
Information Technology Construction Management Center, Kunming University of Science and Technology, Kunming 650504, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(17), 7723; https://doi.org/10.3390/app14177723 (registering DOI)
Submission received: 21 June 2024 / Revised: 25 July 2024 / Accepted: 25 July 2024 / Published: 2 September 2024
(This article belongs to the Special Issue Deep Learning and Predictive Maintenance)

Abstract

:
Conducting the remaining useful life (RUL) prediction for an aircraft engines is of significant importance in enhancing aircraft operation safety and formulating reasonable maintenance plans. Addressing the issue of low prediction model accuracy due to traditional neural networks’ inability to fully extract key features, this paper proposes an engine RUL prediction model based on the adaptive moment estimation (Adam) optimized self-attention mechanism–temporal convolutional network (SAM-TCN) neural network. Firstly, the raw data monitored by sensors are normalized, and RUL labels are set. A sliding window is utilized for overlapping sampling of the data, capturing more temporal features while eliminating data dimensionality. Secondly, the SAM-TCN neural network prediction model is constructed. The temporal convolutional network (TCN) neural network is used to capture the temporal dependency between data, solving the mapping relationship of engine degradation characteristics. A self-attention mechanism (SAM) is employed to adaptively assign different weight contributions to different input features. In the experiments, the root mean square error (RMSE) values on four datasets are 11.50, 16.45, 11.62, and 15.47 respectively. These values indicate further reduction in errors compared to methods reported in other literature. Finally, the SAM-TCN prediction model is optimized using the Adam optimizer to improve the training effectiveness and convergence speed of the model. Experimental results demonstrate that the proposed method can effectively learn feature data, with prediction accuracy superior to other models.

1. Introduction

The turbofan engine is the primary power source for an aircraft. However, as the engine accumulates operating time, its internal components gradually suffer from factors such as wear, corrosion, and fatigue, leading to performance degradation and shortened lifespan. Conducting thorough maintenance checks on engines has become an essential aspect ensuring their healthy operation. Prognostics and health management (PHM) [1] technology is a predictive maintenance strategy that adapts to conditions. It aims to timely detect potential fault indicators through real-time monitoring, analysis, and evaluation of equipment’s operational status and to implement preventive maintenance measures, thereby enhancing equipment safety and operational efficiency. The RUL [2] prediction for turbofan engines refers to assessing and predicting the engine’s future remaining operational lifespan by analyzing its operating status, historical monitoring data, and other information using techniques such as mathematical models, statistical methods, or machine learning algorithms.
RUL prediction technology is a important step in implementing PHM, and therefore, predicting the RUL of engines holds significant research significance and practical value in ensuring flight safety and reducing operational costs for an aircraft. Despite the rapid development of RUL prediction technology in the aviation industry, it still faces several challenges, such as difficulty in extracting effective features from vast amounts of engine raw monitoring data and issues with low accuracy in predicting engine remaining lifespan. In conclusion, integrating advanced domain knowledge with innovative technologies to continually improve the practicality of engine RUL prediction remains a key challenge in the field of aviation maintenance.
Currently, engine RUL prediction methods include empirical methods, physics-based failure model methods, and data-driven methods [3]. Prediction methods based on physical failure models require utilizing the interactions among internal physical processes and components of the engine to establish mathematical models. However, constructing the physical model presents significant challenges [4]. Empirical prediction methods [5] rely on expert industry knowledge and involve analyzing the engine’s operational status and fault data to generate predictions. While simpler, these methods often yield lower accuracy in RUL prediction. Data-driven prediction methods [6] primarily rely on extensive raw engine monitoring data. They use advanced computer algorithms to uncover patterns in the data and construct predictive models for RUL estimation, achieving higher prediction accuracy. Compared to the other two methods, data-driven approaches are more versatile.
Although early machine learning methods such as random forest [7], support vector machine [8] and Markov models [9] have shown good capabilities in data processing, their relatively simple structures make it difficult to fully explore the hidden relationships among data, resulting in lower prediction accuracy. In contrast, deep learning methods can learn abstract feature representations of data, which give them an advantage in handling long-term data dependencies and complex dynamic patterns.
In the realm of deep learning methods, Caceres and others [10] have proposed a recurrent neural network (RNN) prediction method based on probabilistic Bayesian principles to address the uncertainty factors in RUL prediction. Babu et al. [11] were the first to apply convolutional neural network (CNN) methods to RUL prediction problems, enabling automatic learning from raw signals. Zhang et al. [12] developed a bidirectional long short-term memory (LSTM) neural network prediction model to assess engine remaining life. LYU et al. [13] used 1D-CNN to map multiple state variables’ long-term sequences directly to the engine’s remaining life, achieving more accurate RUL values. Although the aforementioned neural network learning methods have been extensively researched in engine remaining life prediction, certain limitations persist. For instance, RNN neural networks exhibit poor feature extraction ability, CNN neural networks fall short in handling time series data, and LSTM neural networks show low predictive accuracy under complex operating conditions. Considering these shortcomings, researchers [14] have introduced a TCN network. This neural network architecture is capable of effectively capturing local patterns in time series data. Huang [15] leveraged a multi-head probability sparse self-attention mechanism in deep learning to enhance data relevance and combined it with causal convolutions in TCN neural networks to boost the model’s learning capabilities. Song et al. [16] designed a MCA-DTCN network lifespan prediction model to extract key features from engine data. These research findings indicate that TCN prediction methods offer advantages in engine life prediction. However, considering that different sensor monitoring data may impact engine remaining life to varying degrees, it is essential to thoroughly investigate these issues during model training to prevent a decrease in prediction accuracy. Hence, when utilizing TCN neural network modeling, continuous optimization and improvement of the TCN neural network structure are necessary to enhance predictive performance.
To better address the above issue, this paper constructs a turbine engine RUL prediction method based on an Adam-optimized SAM-TCN. This approach aims to accurately capture significant feature data contributions to address the issue of low RUL prediction accuracy. The main contributions of this paper are as follows.
Considering that turbofan engine data used for experiments are multivariate time series, TCN neural networks are employed in this study to handle these temporal data sequences effectively, thereby avoiding the problems of gradient vanishing and slow convergence speed typically encountered in recurrent neural networks during training. Simultaneously, the SAM method is utilized to capture the impact magnitude of each feature sequence on the engine’s RUL, enabling the prediction model to focus more on useful feature data and enhancing its capability to extract key features. Additionally, the Adam optimization method is introduced to globally optimize the prediction model, thereby improving its training effectiveness and accuracy.
Experimental verification and analysis are conducted on the C-MAPSS dataset for turbofan engines (FD001-FD004 datasets). The RMSE values obtained are 11.50, 16.45, 11.62, and 15.47, respectively, and the Score values are 225.32, 1136.27, 259.79, and 1365.40, respectively. The results of these two evaluation metrics demonstrate that the proposed method in this paper achieves lower prediction errors and better performance compared to others in the literature, effectively validating the efficacy of the proposed approach.

2. Background

2.1. Self-Attention

The SAM is primarily used for handling complex data. The SAM can capture the relationships between the overall feature sequence and the current moment’s features, assigning different attention weights to input data features [17]. Compared to the attention mechanism, SAM not only swiftly captures internal feature relationships in data, thus reducing dependence on external information, but also allows the model to attend to all positions within the input sequence simultaneously during processing.
The three crucial vectors comprising the self-attention mechanism are Q i , K i , and V i vectors. These vectors are obtained through linear transformations of the input data X = x 1 , x 2 , , x i using corresponding learned weight matrices W q , W k , and W v .
To represent the attention level between each pair of input sequence elements x i and x j , attention scores need to be computed. These scores are initially calculated as the dot product of the Q i and the K i , which is divided by a scaling factor. According to reference [18], the calculation formula is as follows:
s c o r e ( Q i , K i ) = Q i K j d k
In the equation, d K denotes the dimensionality of the K vector.
To normalize the computed attention scores using function softmax—transforming them into values between 0 and 1 such that their sum equals 1—we obtain attention weight values. The calculation formula is
w i j = s o f t m a x s c o r e Q i , K i = s o f t m a x Q i K j d k
In the equation, W i j represents the attention weight.
We calculate the final output result:
z i = j = 1 n w i j V j
In the equation, Z i represents the final output result.
The principle structure diagram of self attention mechanism is shown in Figure 1 [19]. Introducing self-attention mechanisms in neural networks effectively addresses the issue of information overload in predictive models [20]. Self-attention allows the model to dynamically adjust attention distributions when processing input sequences, thereby reducing reliance on the entire sequence and improving computational efficiency.

2.2. Temporal Convolutional Network

Researchers introduced the TCN into their networks to address the inefficiency of CNNs in handling sequential data. This neural network not only inherits CNN’s powerful feature extraction capabilities but also leverages recurrent neural networks’ ability to store historical information. The TCN network can efficiently process input sequences of any length, effectively improving the efficiency and accuracy of prediction models.
Causal convolution predicts the output data x t + 1 at time t + 1 by learning input information x 1 , x 2 , , x t up to time t, effectively preventing future information leakage and addressing differences in input and output time steps during model training. When processing long-sequence input data, stacking more layers of networks is necessary, increasing the network’s complexity. To tackle these issues, dilated convolution is introduced to improve operational efficiency. Dilated convolution expands the receptive field by inserting zero elements between convolutional kernel elements. The dilated convolution of input data X = x 1 , x 2 , , x t at time t is formulated as [21]
F d X x t = k = 1 K f k x t ( K k ) d
In the formula, F ( f 1 , f 2 , , f k ) represents the filter; K is the size of the convolution kernel; d represents the dilation factor.
To effectively mitigate issues such as gradient explosion and vanishing gradients in deep networks, the residual neural network structure was introduced to learn residual errors of feature mappings across layers, thereby enhancing neural network stability. Residual connections combine with identity mapping functions to transmit information across layers in the network. The computational formula is
o = A c t i v a t i o n ( x + F ( x ) )
In the formula, x represents the input information; Activation represents the activation function; F ( x ) represents the output after residual connection.
The structure of the temporal convolutional network is shown in Figure 2 [22]. The output from the dilated causal convolutions is added to the input using 1 × 1 convolution operations. Weight normalization accelerates neural network convergence and improves training speed.

2.3. Adam Optimization Algorithm

The Adam algorithm [23] is a stochastic optimization algorithm designed to accelerate the optimization process and improve algorithm performance. It adjusts the learning rate adaptively to optimize the parameters of neural networks, demonstrating excellent performance in large-scale data and noisy environments.
Let the parameters at the current step t be θ t , the gradient of θ t at the current step be g t , the first moment estimate be m t , and the second moment estimate be v t . Their respective formulas are as follows [24]:
m t = β 1 m t 1 + 1 β 1 g 1
v t = β 2 v t 1 + 1 β 2 g t 2
In the equation, β 1 and β 2 represent adjustable exponential decay rates.
m ^ t = m t 1 β 1 t
v ^ t = v t 1 β 2 t
We update the model parameters θ t :
θ t = θ t 1 α m ^ t v ^ t + ε
In the equation, α denotes the learning rate; ε is a very small numerical value.

3. Related Work

Research on RUL prediction of turbofan engines is a crucial topic in the field of flight safety. Accurately predicting the engine’s remaining lifespan can assist airlines in optimizing maintenance schedules and reducing unscheduled maintenance incidents. Currently, machine learning and artificial intelligence technologies are predominantly used in the research of turbofan engine RUL prediction, supported by advanced data collection and sensor technologies.
In existing studies, Qiao et al. [25] discussed the data and results from RUL prediction experiments on turbofan engines, finding that indirect mapping approaches yield better results despite being more challenging and time-consuming to implement. Song et al. [26], considering the high dimensionality and volume of engine monitoring data, designed a bidirectional long short-term memory network for RUL prediction. Li et al. [27] respectively utilized CNN and LSTM networks to extract spatial features and perform data fusion, and they employed the SAM method to obtain feature weights. Zhen et al. [28] proposed a TCN-attention model for oil well production prediction to overcome issues with traditional neural networks such as poor data processing effects and gradient vanishing. Although this method mitigates the shortcomings of traditional neural networks, its prediction accuracy and overall performance still require improvement.
To address these challenges effectively, this study introduces a self-attention mechanism into the temporal convolutional network to obtain different feature weights and utilizes the Adam optimization algorithm to enhance the overall performance of the prediction model, significantly improving prediction accuracy.

4. The Proposed Method

4.1. SAM-TCN

To address the issue where TCN neural networks fail to sufficiently extract relationships between input data during model training, thereby resulting in poor performance, this paper introduces a SAM based on the TCN. The SAM method assigns varying attention weights to input features, thereby enhancing the TCN neural network’s feature extraction capabilities. Figure 3 shows the model structure diagram of the self-attention mechanism and temporal convolutional network. Using the Q and K vectors, attention scores are computed, normalized using the s o f t m a x function, and then multiplied by the V vector values for each data point before summing them up. The output data are activated using the ReLU activation function to obtain feature channel data weighted by different attention weights. These new feature data are then input into the TCN neural network, where dilated causal convolutions extract features from the inputted new features and produce the final output features.

4.2. The Adam-SAM-TCN Remaining Useful Life Prediction Model

Figure 4 shows the overall structure of the Adam SAM TCN prediction model. The specific process of the RUL prediction method is as follows:
(1)
From the 21-dimension raw data of datasets FD001 to FD004, 14 dimensions showing significant degradation trends were selected. These data were normalized and labeled with RUL, and initial values for sliding windows were set.
(2)
The SAM attention mechanism was introduced into the TCN to assign higher weights to input data that have a greater impact on the engine’s remaining useful life, thereby obtaining new feature data. We input these new features into the TCN network for training.
(3)
After initializing the parameters of the Adam optimizer, the parameters of the constructed SAM-TCN network were optimized to obtain the optimal parameters.
(4)
Utilizing the optimized Adam-SAM-TCN prediction model, the RUL prediction for the turbofan engine was conducted on the four datasets. The predicted results were analyzed and evaluated based on evaluation metrics.
The parameter settings of the proposed Adam-SAM-TCN neural network prediction model are shown in Table 1. Here, W represents the sliding window size, in denotes the number of input channels, o u t indicates the number of channels produced by convolution, s t r i d e represents the step size, S o f t m a x and R e L U are activation functions, k e r n e l s i z e denotes the size of the convolutional kernel, and p a d d i n g refers to the padding value on both sides of the input.

5. Experiment

5.1. Experimental Data

Validated using C-MAPSS data developed by the NASA’s Ames Research Center, we simulated the degradation of various components such as fan, turbine, and compressor in a turbofan engine under 9000 pounds of thrust. Parameters recorded include pressure, temperature, and rotational speed. This dataset is known as the C-MAPSS data [29]. Based on different engine operational conditions, this dataset was divided into four subsets. Among them, FD001 and FD003 contain single operational condition data, while FD002 and FD004 contain multi-operational condition data. Table 2 shows the descriptive information of the data, where fault mode 1 denotes a high-pressure compressor fault, and fault mode 2 denotes both high-pressure compressor and fan faults. The training set comprises data covering the entire lifecycle of the engine. The test set consists of monitoring data from early normal operation to a specific point before failure occurs.
The dataset includes data from three operational condition settings and 21 sensor measurements. Specific descriptions are provided in Table 3. Among them, °R denotes Rankine temperature units; p s i a represents pressure units in pounds per square inch absolute ( l b f / i n 2 ) ; r p m stands for rotational speed in revolutions per minute; l b m / s indicates flow rate in pounds per second.

5.2. Evaluation Metrics for Predicted Results

The definition of RUL for a turbofan engine is the remaining usable cycles from the current operating time of the engine until complete performance degradation and failure. The specific expression for RUL of a turbofan engine is [30]
T R U L ( t ) = t f t c t f t c
In the expression, t f denotes the time of engine degradation failure, and t c represents the current operating time of the engine.
This study used two evaluation indicators to analyze the prediction results. R M S E considers both the magnitude and direction of prediction errors, making it very useful for evaluating the overall performance of models [31]. The Score metric imposes greater penalties for subsequent fault predictions. The formulas for these two evaluation metrics are as follows:
R M S E = 1 n i n y ^ i y i 2
S c o r e = i = 1 n ( e ( y ^ i y i ) 13 1 ) ,   y ^ i y i < 0 i = 1 n ( e y ^ i y i 10 1 ) ,   y ^ i y i > 0 ,
In the above formulas, y i represents the actual RUL value of the ith engine, y ^ i represents the predicted RUL value of the ith engine, and n denotes the number of engines.

5.3. Experimental Result

5.3.1. Data Processing

Feature Selection

The C-MAPSS dataset includes a total of 21 monitored parameters collected by sensors. The trends of raw monitoring parameters from a single engine are illustrated in Figure 5. The trends of T 2 , P 2 , P 15 , e p r , B f a , N f d m d , and N R f d m d are not significant, indicating that these parameters have minimal influence on the remaining life of the engine. To reduce data redundancy, these seven parameters were removed. On the other hand, T 24 , T 30 , T 50 , P 30 , N f , N c , P s 30 , p h i , N R f , N R c , B P R , B l e e d , W 31 , and W 32 show clear monotonic trends with increasing engine operation cycles, suggesting a certain correlation with the engine degradation process. Therefore, these 14 monitored parameters were selected as input variables for the prediction model.

Data Normalization and RUL Labeling

Since the experimental data from CMAPSS consists of multiple parameters monitored by different sensors, these parameters exhibit varying units and significantly different scales. To mitigate the errors caused by dimensional discrepancies during model training, this study employed the MinMax scaling method to normalize the 14 monitored parameters in the dataset. The specific formula is as follows [32]:
x i = x i x i min x i max x i min
In the equation, x i and x i respectively denote the monitor data and the data after normalization of the ith monitored parameter from the sensor; x i max and x i min represent the maximum and minimum datas of the ith monitored parameter from the sensor.
Figure 6 illustrates the effects of normalization on the 14 types of monitoring data for a single engine.
The paper employed piecewise linear functions to describe the RUL of turbofan engines, as shown in Figure 7. In the initial stages of engine operation, wear was minimal, and the RUL remained stable. As the engine reached a certain degradation threshold, performance started to decline, causing the RUL to gradually decrease until reaching its end of life. The formula for calculating the segmented linear degradation curve of the engine RUL is as follows [33]:
R U L = R U L init ,   R U L i R U L init R U L i ,   R U L i < R U L init
In the equation, R U L init represents the initial degradation threshold of the engine, and R U L i denotes the RUL of the engine in the ith cycle period.

5.3.2. Predicted RUL Results

Different lengths of sliding window settings capture data patterns and features at different scales. Appropriate sliding window settings can enhance the overall accuracy of prediction models. To determine the optimal window length and maximize the performance of the prediction model, this study experimented with sliding window sizes of 10, 20, 30, 40, 50, and 60 for comparison of the experimental results. Figure 8 illustrates the R M S E values under different sliding window sizes across four datasets. As seen in Figure 8, the R M S E values of the model were minimized when the sliding window size was 40 for FD001 and FD002 and when it was 50 for FD003 and FD004.
Figure 9 shows the change in loss function throughout the entire iteration process during training of the SAM-TCN neural network. As seen in Figure 9, it is evident that across all four datasets that as the number of iterations increased, the predictive performance of the SAM-TCN neural network steadily improved. The objective loss function continuously decreased and eventually stabilized.
To explore the performance of the proposed Adam-SAM-TCN method across different datasets, the results are shown in Figure 10. In the figure, the horizontal axis represents the engine number, and the vertical axis represents the RUL of the engines. The solid blue line indicates the actual RUL values of the engines, while the dashed red line represents the predicted RUL results using the method proposed in this paper. As seen in Figure 10, it can be observed that the blue solid line fits closely with the red dashed line, indicating a small prediction error and hence a good predictive performance. The predictions for the single operating condition datasets FD001 and FD003 are closer to the actual values, with smaller prediction errors. In contrast, datasets FD002 and FD004, collected under multiple operating conditions, exhibited more complexity and posed greater challenges for prediction, resulting in lower prediction accuracy. Overall, the proposed method demonstrates a generally close alignment between the predicted results and actual values across all datasets.
To better analyze the predictive performance of the proposed method on individual engines, RUL predictions were conducted on the engines numbered 20, 185, 1, and 111 from the test set across four datasets. The predicted results are shown in Figure 11, where the actual RUL values of the engines closely align with the predictions, with the majority of predicted values falling within the 95% confidence interval. This indicates that the proposed method demonstrates effective predictive performance.
Figure 12 shows the distribution of prediction errors for the remaining useful life of engines on four datasets. The horizontal axis represents the size of prediction error. The vertical axis shows the frequency of occurrences within each prediction error interval. As seen in Figure 12, it is evident that the majority of prediction errors across all four datasets fell within the range of [−25, 25], indicating that the proposed method provides fairly accurate predictions of engine RUL.

6. Discussion

6.1. Experimental Results Dicussion

This paper proposed a hybrid network structure using Adam optimization and SAM-TCN for predicting the RUL of engines. It utilized a SAM to measure the contributions of different features to the engine’s remaining life. The results were evaluated using comprehensive performance metrics, with the RMSE and Score on four datasets reported as follows: 11.50, 16.45, 11.62, and 15.47 for the RMSE and 225.32, 1136.27, 259.79, and 1365.40 for the Score. From these metrics, it is evident that the proposed method performed well on the FD001 and FD003 datasets with smaller errors. This is primarily because the other two datasets contain more noise and complex operating conditions, making it challenging for the prediction model to capture these variations and resulting in less satisfactory performance. However, overall, the proposed model achieved good predictive accuracy. To comprehensively assess the prediction effectiveness of the proposed method, comparisons and analyses with methods from other literature are conducted in the next section.

6.2. Comparative Study

To better analyze the strengths and weaknesses of the proposed method, the Adam-SAM-TCN method was compared with recent methods. The RMSE and Score values of these methods on the dataset are presented in Table 4 and Table 5. As seen in Table 4, it can be observed that all prediction methods performed better on the datasets FD001 and FD003 compared to the datasets FD002 and FD004. Compared to the LSTM [34], CNN-BGRU-SA [35], TaFCN [36], Multi-attention-TCN [37], RCNN-Abi-LSTM [38], and GATA-TCN [39] methods proposed in other studies, the Adam-SAM-TCN method achieved the smallest RMSE values, indicating superior predictive performance in engine life prediction.
As seen in Table 5, it can be observed that, unlike the RMSE values in Table 4, the method from reference [37] achieved the lowest Score on the FD003 dataset. This is mainly due to Score’s penalties for overestimation and underestimation of the RUL. Apart from the FD003 dataset, the prediction method Adam-SAM-TCN obtained smaller Score values compared to the methods from other literature sources.

7. Conclusions

To address the challenges in predicting the RUL for turbofan engines under multi-operational conditions, characterized by difficulties and low accuracy, this paper proposed an Adam-optimized SAM-TCN neural network for engine RUL prediction. After experimental verification, the following conclusions have been obtained:
(1)
The Adam-SAM-TCN method introduces an SAM module on top of the TCN neural network for prediction modeling. This neural network structure enables the model to focus on local patterns during training while effectively capturing global patterns in data.
(2)
The method utilizes the Adam to optimize the prediction model, which adaptively adjusts the learning rate based on historical gradient information. The algorithm swiftly and accurately minimizes the loss function, thereby effectively enhancing the training effectiveness and generalization capability of the neural network.
(3)
By conducting training and testing on four datasets, experimental results show that the RMSE values for this paper came out to 11.50, 16.45, 11.62, and 15.47 respectively. Compared to existing prediction models, the evaluation metrics of the proposed method were consistently lower, demonstrating the effectiveness of the Adam-SAM-TCN prediction model proposed in this study.

Author Contributions

Conceptualization, H.W. and Y.L.; methodology, D.L.; software, D.L.; validation, D.L.; formal analysis, G.Z.; investigation, D.L., Y.L. and G.Z.; resources, H.W.; data curation, D.L.; writing—original draft preparation, D.L.; writing—review and editing, D.L.; visualization, G.Z. and R.L.; supervision, H.W.; project administration, Y.L. and R.L.; funding acquisition, H.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Funding Number: 61863016).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in this study are openly available in the NASA repository, and they are called Turbofan Engine Degradation Simulation Dataset and PHM08 Challenge Dataset: (https://www.nasa.gov/intelligent-systems-division, accessed on 16 February 2023).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ali, A.; Soheil, Z.; Amir, A.; Arash, M. A multimodal and hybrid deep neural network model for Remaining Useful Life estimation. Comput. Ind. 2019, 108, 186–196. [Google Scholar] [CrossRef]
  2. Zhao, Z.; Liang, B.; Wang, X.; Lu, W. Remaining useful life prediction of aircraft engine based on degradation pattern learning. Reliab. Eng. Syst. Saf. 2017, 164, 74–83. [Google Scholar] [CrossRef]
  3. Nei, L.; Cai, W.; Zhang, L.; Xu, S.; Wu, R.; Ren, Y. Remaining Useful Life Prediction of Turbofan Engine Based on SAE-SA-1D-CNN-BGRU. Aeroengine 2023, 49, 134–139. [Google Scholar]
  4. Liu, J.; Lei, F.; Pan, C.; Hu, D.; Zuo, H. Prediction of remaining useful life of multi-stage aero-engine based on clustering and LSTM fusion. Reliab. Eng. Syst. Saf. 2021, 214, 107807. [Google Scholar] [CrossRef]
  5. Wang, W. Life Prediction Based on Historical Data of Aircraft Engines. Master’s Thesis, Anhui University of Science and Technology, Anhui, China, 2023; pp. 1–69. Available online: https://kns.cnki.net/kcms2/article/abstract?v=sAMp-nZqXjwZCYfF-REwGuLAbZdXvzQDm2VfwUt_kLb3ZD1aDccK5FA4yn1fKiANeKFHsppVnA4Jb00RiYUURZtimcsCir0bAp5_ldutDjaHXX0erC4bNF01UyxiJohgQPsfu0pgLTl26eP8FKNMgrIK6SEfognrcBlWDCazg82D0SwEJTo91uH72pQd_xGzXwUabLZLGLM=&uniplatform=NZKPT&language=CHS (accessed on 18 May 2023).
  6. Kong, Z.; Jin, X.; Xu, Z.; Zhang, B. Spatio-temporal fusion attention: A novel approach for remaining useful life prediction based on graph neural network. IEEE Trans. Instrum. Meas. 2022, 71, 3515912. [Google Scholar] [CrossRef]
  7. Wang, C.; Lu, N.; Cheng, Y.; Jiang, B. A Data-Driven Aero-Engine Degradation Prognostic Strategy. IEEE Trans. Cybern. 2021, 51, 1531–1541. [Google Scholar] [CrossRef]
  8. Khelif, R.; Chebelmorello, B.; Malinowski, S.; Laajili, E.; Fnaiech, F.; Zerhouni, N. Direct remaining useful life estimation based on support vector regression. IEEE Trans. Ind. Electron. 2016, 64, 2276–2285. [Google Scholar] [CrossRef]
  9. Chadza, T.; Kyriakopoulos, K.G.; Lambotharan, S. Contemporary sequential network attacks predict-ion using hidden Markov model. In Proceedings of the 2019 17th International Conference on Privacy, Security and Trust (PST), Fredericton, NB, Canada, 26–28 August 2019; pp. 1–3. [Google Scholar] [CrossRef]
  10. Caceres, J.; Gonzalez, D.; Zhou, T.; Droguett-Enrique, L. A probabilistic Bayesian recurrent neural network for remaining useful life prognostics considering epistemic and aleatory uncertainties. Struct. Control Health Monit. 2021, 28, e2811. [Google Scholar] [CrossRef]
  11. Sateesh Babu, G.; Zhao, P.; Xiao, L. Deep convolutional neural network based regression approach for estimation of remaining useful life. In Proceedings of the International Conference on Database Systems for Advanced Applications, Dallas, TX, USA, 16–19 April 2016. [Google Scholar]
  12. Zhang, J.; Wang, P.; Yan, R.; Gao, R.X. Long short-term memory for machine remaining life prediction. J. Manuf. Syst. 2018, 48, 78–86. [Google Scholar] [CrossRef]
  13. Lyu, D.; Hu, Y. Remaining Useful Life Prediction of Aeroengine Based on Principal Component Analysis and One-Dimensional Convolutional Neural Network. Trans. Nanjing Univ. Aeronaut. Astronaut. 2022, 38, 867–875. [Google Scholar]
  14. Bai, S.; Kolter, J.Z.; Koltun, V. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv 2018, arXiv:1803.01271. [Google Scholar]
  15. Huang, J. Research and Implementation of Industrial Time Series Analysis Method Based on Machine Learning. Master’s Thesis, Civil Aviation Flight University of China, Deyang, China, 2023. [Google Scholar]
  16. Fu, S.; Lin, L.; Wang, Y.; Guo, F.; Zhao, M.; Zhong, B.; Zhong, S. MCA-DTCN: A novel dual-task temporal convolutional network with multi-channel attention for first prediction time detection and remaining useful life prediction. Reliab. Eng. Syst. Saf. 2024, 241, 109696. [Google Scholar] [CrossRef]
  17. Li, Y.; Baciu, G. SG-GAN: Adversarial Self-Attention GCN for Point Cloud Topological Parts Generation. IEEE Trans. Vis. Comput. Graph. 2021, 28, 3499–3512. [Google Scholar] [CrossRef] [PubMed]
  18. Ye, Q. Research on Maintenance Decision-making of Turbofan Engines Based on Attention-based Temporal Convolutional Network and Evolutionary Game. Master’s Thesis, Huazhong University of Science and Technology, Wuhan, China, 2021; pp. 1–102. [Google Scholar]
  19. Xu, Z.; Zhang, Y.; Miao, J.; Miao, Q. Global attention mechanism based deep learning for remaining useful life prediction of aero-engine. Measurement 2023, 217, 113098. [Google Scholar] [CrossRef]
  20. Zhang, H.; Zhang, Q.; Shao, S.; Niu, T.; Yang, X. Attention-based LSTM network for rotatory machineremaining useful life prediction. IEEE Access 2020, 8, 188–199. [Google Scholar] [CrossRef]
  21. Zhang, G. Rolling Bearing Life Prediction Based on Temporal Convolutional Network. Master’s Thesis, Nanjing University of Aeronautics and Astronautics, Nanjing, China, 2022; pp. 1–76. [Google Scholar]
  22. Sun, Y.; Ding, Q.; Xia, Y.; Li, C. Chiller fault diagnosis based on combination of multiblock and self-attention TCN. Chin. J. Process. Eng. 2024, 24, 162–171. [Google Scholar] [CrossRef]
  23. Agrawal, S.; Sarkar, S.; Srivastava, G.; Maddikunta, P.K.R.; Gadekallu, T.R. Genetically optimized prediction of remaining useful life. Sustain. Comput. Inform. Syst. 2021, 31, 10056. [Google Scholar] [CrossRef]
  24. Lu, Y. Prediction of Residual Life of Gearbox Bearing of Off shore Wind Turbine. Master’s Thesis, Jiangsu University of Science and Technology, Zhengjiang, China, 2022; pp. 1–75. [Google Scholar] [CrossRef]
  25. Qiao, X.; Jauw, V.L.; Seong, L.C.; Banda, T. Advances and limitations in machine learning approaches applied to remaining useful life predictions: A critical review. Int. J. Adv. Manuf. Technol. 2024, 133, 4059–4076. [Google Scholar] [CrossRef]
  26. Song, Y.; Shi, G.; Chen, L.; Huang, X.; Xia, T. Remaining Useful Life Prediction of Turbofan Engine Using Hybrid Model Based on Autoencoder and Bidirectional Long Short-Term Memory. J. Shanghai Jiaotong Univ. (Sci.) 2018, 23, 85–94. [Google Scholar] [CrossRef]
  27. Li, J.; Jia, Y.; Niu, M.; Zhu, W.; Meng, F. Remaining Useful Life Prediction of Turbofan Engines Using CNN-LSTM-SAM Approach. IEEE Sens. J. 2023, 23, 10241–10251. [Google Scholar] [CrossRef]
  28. Zhen, Y.; Fang, J.; Zhao, X.; Ge, J.; Xiao, Y. Temporal convolution network based on attention mechanism for well production prediction. J. Pet. Sci. Eng. 2022, 218, 111043. [Google Scholar] [CrossRef]
  29. Saxena, A.; Goebel, K.; Simon, D.; Eklund, N. Damage propagation modeling for aircraft engine run-to-failure simulation. In Proceedings of the 2008 International Conference on Prognostics and Health Management, Denver, CO, USA, 6–9 October 2008. [Google Scholar] [CrossRef]
  30. Shi, G. Research on Data-Driven Aero-Engine RUL Prediction Model. Master’s Thesis, Dalian University of Technology, Dalian, China, 2022. [Google Scholar]
  31. Wang, H.; Ye, X.; Li, Y.; Zhu, G. Remaining Useful Life Prediction for Lithium-Ion Batteries Based on Improved Mode Decomposition and Time Series. Sustainability 2023, 15, 9176. [Google Scholar] [CrossRef]
  32. Wang, H.; Li, D.; Li, D.; Liu, C.; Yang, X.; Zhu, G. Remaining Useful Life Prediction of Aircraft Turbofan Engine Based on Random Forest Feature Selection and Multi-Layer Perceptron. Appl. Sci. 2023, 13, 7186. [Google Scholar] [CrossRef]
  33. Wu, J.; Chen, B.; Chen, Z.; Wang, G.; Tian, Z.; Liu, Q. Application of SSA-TCN in Prediction of Remaining Useful Life of Turbofan Engine. J. China Three Gorges Univ. (Nat. Sci.) 2023, 45, 92–100. [Google Scholar]
  34. Rounak, B.; Manikandan, J. Prediction of Remaining Useful Life for Aero-Engines. In Proceedings of the 2021 IEEE International Conference on Aerospace Electronics and Remote Sensing Technology (ICARES), Bali, Indonesia, 3–4 November 2021. [Google Scholar] [CrossRef]
  35. Sun, J.; Zheng, L.; Huang, Y.; Ge, Y. Remaining Useful Life Prediction Based on CNN-BGRU-SA. In Proceedings of the 2022 International Conference on Electronics Technology and Artificial Intelligence (ETAI 2022), Chongqing, China, 16–18 September 2022. [Google Scholar] [CrossRef]
  36. Fan, L.; Yi, C.; Chai, Y. Trend attention fully convolutional network for remaining useful life estimation. Reliab. Eng. Syst. Saf. 2022, 225, 108590. [Google Scholar] [CrossRef]
  37. Shang, Z.; Zhang, B.; Li, W.; Qian, S.Q.; Zhang, J. Machine remaining life prediction based on multi-layer self-attention and temporal convolution network. Complex Intell. Syst. 2022, 8, 1409–1424. [Google Scholar] [CrossRef]
  38. Yan, X.; Liang, W.; Zhang, G.; She, B.; Tian, F. Prediction method for mechanical equipment based on RCNN-A BiLSTM. Syst. Eng. Electron. 2023, 45, 931–940. [Google Scholar] [CrossRef]
  39. Lin, L.; Wu, J.; Fu, S.; Zhang, S.; Tong, C.; Zu, L. Channel attention & temporal attention based temporal convolutional network: A dual attention framework for remaining useful life prediction of the aircraft engines. Adv. Eng. Inform. 2024, 60, 102372. [Google Scholar] [CrossRef]
Figure 1. Structure diagram of self-attention mechanism principle.
Figure 1. Structure diagram of self-attention mechanism principle.
Applsci 14 07723 g001
Figure 2. Temporal convolutional network structure diagram.
Figure 2. Temporal convolutional network structure diagram.
Applsci 14 07723 g002
Figure 3. Model structure diagram of self-attention mechanism and temporal convolutional network.
Figure 3. Model structure diagram of self-attention mechanism and temporal convolutional network.
Applsci 14 07723 g003
Figure 4. Overall flowchart of turbofan engine RUL prediction.
Figure 4. Overall flowchart of turbofan engine RUL prediction.
Applsci 14 07723 g004
Figure 5. Trend of raw monitoring parameters for single engine.
Figure 5. Trend of raw monitoring parameters for single engine.
Applsci 14 07723 g005
Figure 6. Effect of normalization on 14 types of monitoring data from a single engine.
Figure 6. Effect of normalization on 14 types of monitoring data from a single engine.
Applsci 14 07723 g006
Figure 7. Linear and segmented degradation curves of engine RUL.
Figure 7. Linear and segmented degradation curves of engine RUL.
Applsci 14 07723 g007
Figure 8. Impact of different sliding windows on prediction errors.
Figure 8. Impact of different sliding windows on prediction errors.
Applsci 14 07723 g008
Figure 9. Loss function variation of prediction model across four datasets.
Figure 9. Loss function variation of prediction model across four datasets.
Applsci 14 07723 g009
Figure 10. Comparison of actual and predicted RUL values across 4 datasets.
Figure 10. Comparison of actual and predicted RUL values across 4 datasets.
Applsci 14 07723 g010
Figure 11. Graph of predicted remaining life of individual engines.
Figure 11. Graph of predicted remaining life of individual engines.
Applsci 14 07723 g011aApplsci 14 07723 g011b
Figure 12. Distribution diagram of prediction error for remaining useful life of engine.
Figure 12. Distribution diagram of prediction error for remaining useful life of engine.
Applsci 14 07723 g012
Table 1. Configuration of Adam-SAM-TCN model structure parameters.
Table 1. Configuration of Adam-SAM-TCN model structure parameters.
StructureSpecific Parameters
Input W 1 = 40 ,   W 2 = 40 ,   W 3 = 50 ,   W 4 = 50
SAM layer i n = 40 ,   o u t = 40 ,   s t r i d e = 1 ,   s o f t m a x
TCN layer cov11 k e r n e l s i z e = 3 ,   s t r i d e = 1 ,   p a d d i n g = 2 ,   R e L U ,   D r o p o u t ( p = 0.2 )
TCN layer cov21 k e r n e l s i z e = 3 ,   s t r i d e = 1 ,   p a d d i n g = 2 ,   R e L U ,   D r o p o u t ( p = 0.2 )
Adam layer l r = 0.01 ,   b e t a s = ( 0.9 , 0.999 ) ,   e p s = 1 e 8
Table 2. C-MAPSS dataset information.
Table 2. C-MAPSS dataset information.
DatasetFD001FD002FD003FD004
Number of Training Samples100260100249
Number of Test Samples100259100248
Total Length of Training Set20,63053,75824,71961,248
Total Length of Test Set13,09533,99016,59541,213
Operating Conditions1616
Fault Mode1122
Table 3. Description of operational settings and sensor monitoring parameters.
Table 3. Description of operational settings and sensor monitoring parameters.
No.ParametersUnit
1Setting_1
2Setting_2
3Setting_3
4 T 2 °R
5 T 24 °R
6 T 30 °R
7 T 50 °R
8 P 2 psia
9 P 15 psia
10 P 30 psia
11 N f rpm
12 N c rpm
13 e p r
14Ps30psia
15Phi P p s / p s i
16 N R f rpm
17 N R c rpm
18BPR
19 B f a
20Bleed
21 N f _dmdrpm
22 N R f _dmdrpm
23 W 31 lbm/s
24 W 32 lbm/s
Table 4. Comparison of RMSE values of other RUL prediction methods on the C-MAPSS dataset.
Table 4. Comparison of RMSE values of other RUL prediction methods on the C-MAPSS dataset.
MethodFD001FD002FD003FD004Year
LSTM [34]14.1825.2512.7927.632021
CNN-BGRU-SA [35]13.8817.2514.8519.392022
TaFCN [36]13.9917.0612.0119.792022
Multi-attention-TCN [37]13.2519.5713.4321.692022
RCNN-Abi-LSTM [38]12.9819.1613.2422.292023
GATA-TCN [39]12.8017.6113.1621.042024
Adam-SAM-TCN11.5016.4511.6215.472024
Table 5. Comparison of Score values of other RUL prediction methods on the C-MAPSS dataset.
Table 5. Comparison of Score values of other RUL prediction methods on the C-MAPSS dataset.
MethodFD001FD002FD003FD004Year
CNN-BGRU-SA [35]248114029518402022
TaFCN [36]336194625136712022
Multi-attention-TCN [37]235165523924152022
RCNN-Abi-LSTM [38]258298024637952023
GATA-TCN [39]234.311361.23290.632303.422024
Adam-SAM-TCN225.321136.27259.791365.402024
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, H.; Li, D.; Li, Y.; Zhu, G.; Lin, R. Method for Remaining Useful Life Prediction of Turbofan Engines Combining Adam Optimization-Based Self-Attention Mechanism with Temporal Convolutional Networks. Appl. Sci. 2024, 14, 7723. https://doi.org/10.3390/app14177723

AMA Style

Wang H, Li D, Li Y, Zhu G, Lin R. Method for Remaining Useful Life Prediction of Turbofan Engines Combining Adam Optimization-Based Self-Attention Mechanism with Temporal Convolutional Networks. Applied Sciences. 2024; 14(17):7723. https://doi.org/10.3390/app14177723

Chicago/Turabian Style

Wang, Hairui, Dongjun Li, Ya Li, Guifu Zhu, and Rongxiang Lin. 2024. "Method for Remaining Useful Life Prediction of Turbofan Engines Combining Adam Optimization-Based Self-Attention Mechanism with Temporal Convolutional Networks" Applied Sciences 14, no. 17: 7723. https://doi.org/10.3390/app14177723

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop