Next Article in Journal
Pretreatment Method for DNA Barcoding to Analyze Gut Contents of Rotifers
Previous Article in Journal
Development of a Shallow-Depth Soil Temperature Estimation Model Based on Air Temperatures and Soil Water Contents in a Permafrost Area
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Aircraft Engines Remaining Useful Life Prediction with an Improved Online Sequential Extreme Learning Machine

1
Laboratory of Automation and Manufacturing Engineering, University of Batna 2, Batna 05000, Algeria
2
Department of Computer Science, University of Batna 2, Batna 05078, Algeria
3
SIME_Lab (LR 13ES03), University of Sousse–ESSTHS, Sousse 4011, Tunisia
4
Institut de Recherche Dupuy de Lôme (UMR CNRS 6027), University of Brest, 29238 Brest, France
5
Logistics Engineering College, Shanghai Maritime University, Shanghai 201306, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(3), 1062; https://doi.org/10.3390/app10031062
Submission received: 30 December 2019 / Revised: 16 January 2020 / Accepted: 1 February 2020 / Published: 5 February 2020
(This article belongs to the Section Electrical, Electronics and Communications Engineering)

Abstract

:
The efficient data investigation for fast and accurate remaining useful life prediction of aircraft engines can be considered as a very important task for maintenance operations. In this context, the key issue is how an appropriate investigation can be conducted for the extraction of important information from data-driven sequences in high dimensional space in order to guarantee a reliable conclusion. In this paper, a new data-driven learning scheme based on an online sequential extreme learning machine algorithm is proposed for remaining useful life prediction. Firstly, a new feature mapping technique based on stacked autoencoders is proposed to enhance features representations through an accurate reconstruction. In addition, to attempt into addressing dynamic programming based on environmental feedback, a new dynamic forgetting function based on the temporal difference of recursive learning is introduced to enhance dynamic tracking ability of newly coming data. Moreover, a new updated selection strategy was developed in order to discard the unwanted data sequences and to ensure the convergence of the training model parameters to their appropriate values. The proposed approach is validated on the C-MAPSS dataset where experimental results confirm that it yields satisfactory accuracy and efficiency of the prediction model compared to other existing methods.

1. Introduction

Health state prediction in complex equipment or subsystems based on conventional methods such as physical modeling approaches became a very difficult task. Due to the need for deep knowledge of system components and their inner interactions, the modeling process complexity increased and became almost impossible. Even if the final model can be prepared under limited conditions, results might be misleading the prediction or not accurate enough due to poor generalization.
Nowadays, machine learning applications for Remaining Useful Life (RUL) estimation has received growing importance due to the availability of heterogeneous data, which in turn motivated researchers to boost the conventional RUL prediction paradigm with a variety of prediction approaches.
Continuous improvements of machine learning models make its use in Prognostic Health Management (PHM) more relevant [1]. It allows modeling systems behavior by only extracting important patterns from their retrieved data even without a prior knowledge of their internal characteristics. Unlike conventional paradigms, machine learning techniques aims to reduce modeling complexity with less human intervention and less computational costs.
For instance, Xiong Li et al. [2] used a Hidden Semi Markov Model (HSMM) to train SVM for RUL prediction. García et al. [3] designed a hybrid SVM-based swarm intelligence technique searching for the best training coefficients during RUL prediction of aircraft engines. Saidi et al. [4] combined Spectral Kurtosis (SK) to SVM to be able to construct an RUL predictor with a more meaningful feature mapping for wind turbine high-speed shaft bearings health prognosis. Zheng et al. [5] based on a hybrid model constructed a training formwork of the ELM model. They adopted a time window features scaling as a preprocessing and appropriate features selection step to guarantee an accurate prediction of RUL. Ordóñez et al. [6] proposed a hybrid autoregressive model combined with an improved SVM using genetic algorithm to build several estimation algorithms for early RUL prediction. Chen et al. [7] used a SVM-based similarity approach for RUL prediction with the same C-MAPPS dataset. It should be mentioned that all cited works could be classified as a hybrid model that aims to accurately predict the RUL by an unsupervised training or preprocessing of training data before a “single-batch” supervised training. However, time-varying data and parameter update is not considered in these cases, which is incompatible with data-driven prediction models able to address dynamic health deterioration of equipment according to internal or environmental conditions.
Neural networks with mini batch training are able to interact with flowing sequences of data and update the prediction model to satisfy changes in newly arrived ones. In [8], a new data-driven approach is introduced by Ben Ali et al. by training a Simplified Fuzzy Adaptive Resonance Theory Map (SFAM) neural network with Weibull Distribution (WD) to avoid time domain fluctuation during RUL prediction. Al-Dulaimi et al. [9] developed a mini batch hybrid deep NN that takes two paths for RUL estimation; the multidimensional feature extraction based on Long Short Term Memory (LSTM) and convolutional NN and the prediction path via a fusion algorithm. Wen et al. [10] constructed feature mapping and training schemes based on ensemble residual CNN and validated their training model with the K-fold cross-validation method after constructing several learners to predict RUL accurately. Xiang et al. [11] simplified the RUL prediction process of aircraft engines by the direct use of raw sensors measurements without a prior expertise in signal processing using a novel deep convolutional NN combined with a time window feature extractor. The prediction approaches used in these works are based on mini batch training interacting with data-driven sequences. Many training models were developed with different paradigms such as hybrid, ensemble, and deep learning attempting to access more meaningful data representations through an accurate prediction. However, the effect of dynamic changes at the level of the training mini batches of data on the training models is not discussed in this case, which may allow the weights and biases to diverge in some mini batches to under-fit gradually the NN. In other cases, the use of old mini batch training algorithms for NN such as the Backpropagation (BP) algorithm can increase computational costs and time consumption.
Through this analysis, the main challenges for RUL predictions task are:
  • Update the training model based on data variation during time (adaptive learning).
  • Decide whether the newly arrived data is important (new) for the training model or not.
  • Make the training model in line with the actual health state of the equipment by giving more attention to newly acquired data and discarding gradually the old ones.
  • Reduce time consumption during training.
In the recent decade, ELM has been widely used for prediction purposes due to the fast training and fewer parameters tuning based on the best fit of a linear function approximation [12]. Zhou et al. [13] proposed a Stacked ELM (S-ELM) where a stack of small ELMs is specially designed for solving large and complex data problems. Ben Ali et al. [14] presents a new unsupervised learning classification tool of extracted data based on the Adaptive Resonance Theory 2 (ART2) for high-speed shaft bearing prognosis. Li et al. [15] proposed an improved OS-ELM which is one of the ELM variants with an adaptive forgetting factor and an update selection strategy to predict gas utilization ratio using time-varying data. Yang et al. [16] developed a new approach of regularization of the recursive learning of the OS-ELM to reduce over-fitting and both empirical and structural risks of the prediction model. Lu et al. [17] enhanced the recursive learning of the OS-ELM by introducing ensemble Kalman filter propagation for output weights tuning of NN under RUL prediction of aircraft engines. Yin et al. [18] proposed a (TD-RLS(λ)) with a forgetting factor to enhance the RLS adaptation to linear function approximation by updating its parameters based on environmental feedback.
OS-ELM can address online learning in real time such as our RUL prediction problem without iterative tuning, which strongly contribute to computational costs reduction. Training rules of OS-ELM allows recursive learning with any driven chunks of data easily even with different or fixed sizes. In general, one of the limitations of OS-ELM or ELM algorithms is the interpolation caused by the pseudo inverse of the matrix [19], which makes ELM variants suffering from structural risks.
In this paper, an improved data-driven approach based on OS-ELM is proposed for RUL estimation to enhance the novelty of the prediction algorithm and its adaptability to time-varying data. In fact, the main contributions of this work attempting to solve adaptive learning and ill-posed problems are:
  • SAEs based OS-ELM is modified for best features extraction and selection via unsupervised learning.
  • An OS-ELM is introduced.
Tikhonov regularization is involved to reduce structural risk and over-fitting by minimizing the norm of the training weights.
Dynamic forgetting of old data, USS and TD error objective functions are all integrated into both SAEs and OS-ELM to achieve better accuracy and adaptability to newly arrived data.
The proposed approach is applied to the public dataset of the turbofan engine [20], where obtained results compared to other recent used approaches in the literature show higher performance.
The remainder of this work is organized as follows: In Section 2, a general description of the turbofan engine dataset is presented. Section 3 elaborates on the proposed approach used in this paper. Section 4 illustrates and discusses the experimental results where the performances of the proposed learning scheme are showcased and compared to other data-driven methods. Finally, Section 5 concludes with a summary.

2. Dataset Description

The C-MAPSS dataset consists of four different subsets, where each dataset consists of multiple multivariate time series and describes different operating conditions and faults modes for different engines. Each dataset is divided into training and testing sets.
In our work, we have used the subset FD001 and only a single operating condition and one fault mode are considered. At the beginning of each time series, the engine is considered as normally functioning according to initial conditions and at a certain level, it starts deteriorating gradually by losing its performances until the failure.
Each dataset contains 26 columns which are considered as the input features of the prediction model, each column contains information about engine number, time cycles, sensors measurements, and operational conditions of the operating engine [21].
The dataset describes life cycles of 100 different engines with a big amount of training samples where their characteristics changes in time. The reason behind studying this type of data is to address dynamic programming and adaptive learning to solve higher dimensional space problems.
Figure 1 illustrates an example of sensors measurement variation and its corresponding RUL degradation function of the first engine from the chosen dataset, where the different colors represent different sensors.

3. The Proposed Approach

Figure 2 indicates the methodology adopted for RUL prediction in this work. Essentially there are two distinctive parts: Dataset preparation and prediction model training.

3.1. Dataset Preparing

The preparation of the dataset is firstly initialized with an appropriate feature selection where only the most significant and sensitive features will be chosen. Each one must describe a certain attitude of degradation level or gradual time-varying behavior in each time cycle. Measurements that do not represent any variation during operating cycles will not contribute to the learning process. In fact, their noncorrelation with the desired RUL can disturb the training model and makes it more sensitive to over-fitting and other ill-posed problems. In our experiments, we chose only 10 parameters for RUL prediction whose indices are: {7, 8, 9, 12, 14, 16, 17, 20, 25, 26}. After that these features are scaled in the interval [0–1] using a min-max normalization procedure before training or testing. Finally, each time series will be labeled according to the piecewise linear RUL function as proposed in [1,22] and shown previously in Figure 1.

3.2. Prediction Model Training

The training of the RUL prediction model passes by two different steps: Training of the SAEs and OS-ELM. In this paper, both unsupervised and supervised learning are performed based on the proposed algorithm. Therefore, before going any further, first, we will explain the fundamentals of basic OS-ELM. Then, the proposed OS-ELM will be explained in detail for both unsupervised and supervised learning strategies.

3.2.1. Basic OS-ELM

Since ELM learns a big amount of data at once, the output weights β of the hidden layer are also tuned for once and cannot be updated. Therefore, a new problem of weights updating appears when data change with time. Consequently, ELM releases the OS-ELM as one of its variants that depend on an online learning attitude with chunks in varied or fixed size, as it also could learn data that come consecutively in one by one instance [23].
OS-ELM experimentally has given the training rules as explained in the following steps:
For a given small dataset N 0 = { ( X 0 , T 0 ) } | X 0 = { x 1 , , x n } , T 0 = { t 1 , , t n } , initialization of the network parameters:
  • Generate randomly all the input weights and hidden layer biases { ( A , B I ) } | A = { a 1 , , a s } , B I = { b 1 , , b l } . The parameters of hidden nodes (weights and biases) must be normalized between {-1,1} [24].
  • Calculate the initial hidden layer output Hk (k = 0) as the same basic ELM theories as presented in Equation (1), where G can be generated independently from the training data according to any continuous bounded piecewise activation function.
  • Calculate the first output weight matrix βk (k = 0) as shown in Equation (2).
  • Set k = 0.
  • Recursive updating of β for new coming mini batches.
  • Calculate the new hidden layer Hk+1; Equation (1).
  • Update the output weight matrix βk+1; Equation (4).
  • Then, set k = k + 1.
H k = G ( A , B I , X k ) ,
β 0 = P 0 H 0 T 0 ,
P 0 = ( H 0 T H 0 ) 1 ,
β k + 1 = β k P k + 1 H k + 1 T e k + 1 ,
e k + 1 = T k + 1 H k + 1 T β k ,
P k + 1 = P k K k + 1 H k + 1 T P k ,
K k + 1 = P k H k + 1 H k + 1 T P k H k + 1 T ,

3.2.2. Proposed OS-ELM

In the proposed algorithm we have added a regularization parameter C to reduce empirical and structural risks [16] during the initialization phase. Therefore, Equation (3) will be changed into Equation (8) given below [25], where C [ 0 ; + ] .
After the training model is initialized, the USS-based TD objective function is proposed to adopt the RLS weights to the new data variations [18]. Furthermore, we proposed, based on [8,18,26], a new DFF formula that is proposed as illustrated in Equation (10), so Equations (4) and (5) will be changed, respectively, into (9) and (10).
P 0 = { ( 1 C + H 0 T H ) 1 , N l ( 1 C + H 0 T H ) 1 H T , N > l ,
β k + 1 = { β k + M k + 1 δ k + 1 , δ k + 1 > δ k β k , δ k + 1 δ k ,
δ k + 1 = T k + 1 ( H k + 1 γ H k + 2 ) T β k ,
P k + 1 = 1 λ ( P k k k + 1 ( H k + 1 γ H k + 2 ) T P k ) ,
K k + 1 = P k H k λ + 1 + ( H k + 1 γ H k + 2 ) T P k H k ,
λ k + 1 = λ min + ( 1 λ min ) e μ δ k + 1 ,
The dynamic forgetting factor changes according to the proposed formula in Equation (13) where λ min [ 0 , 1 ] and λ min λ 1 .

3.2.3. Training of the Proposed Network

SAEs are serially trained autoencoders. The encoded features of an autoencoder are considered as the inputs of the next one as shown in Figure 3.
Both of the unsupervised and supervised training in our proposed approach carried with fixed sizes of mini batches to verify the objective function of TD error in Equation (10) and to avoid a mismatch dimensions error for programming reasons. The results of the mapping of the SAEs are directly used for fine-tuning the RUL prediction network, the update of supervised and unsupervised training models is simultaneously achieved.
During the training of the SAEs, we have to change the targets in the above-proposed equations to be the same as the input [13]. Therefore, Equations (1), (2), and (10) must be changed to fit the training process of the SAEs.
Our main contribution to modify the autoencoder in this work is that the basic ELM theories for training the autoencoder show that: “After training the Autoencoder, we can encode features using the transpose of output weights matrix [13], Equation (14)”. However, mathematically it is proven that the best encoding can be obtained by using the inverse matrix of the transpose of output weights as shown in Equation (15). The proposed formula is tested experimentally and proved its accuracy [27].
X = H β H = X β T ,
H = ( β T ) 1 X ,
An important note that we should take into consideration during the training of the SAEs is that after the initialization phase, we are no longer in need of the randomly generated input weights and we will use the formula in Equation (16) instead of the one in Equation (1).
H k + 1 = G ( ( β k T ) 1 , B I , X k + 1 ) ,
where Xk+1 are the kth+1 mini batch of inputs of the training set.

4. Experiments, Results, and Discussion

The training parameters of the proposed algorithm are tuned after repeating the experiment in reason to attempt the best score value in Equation (17) and the minimum RMSE using Equation (18) for the test set in both Equations (17) and (18).
S = { i = 1 Z e d i / 13 1 , d i < 0 i = 1 Z e d i / 10 1 , d i 0
RMSE = 1 Z i = 1 Z d 2 ,
According to the PHM 2008 challenge [22], the scoring function penalizes the latent prediction d i 0 more than early ones d i < 0 for maintenance reasons. Too late prediction might delay maintenance operations, and too early prediction might not be harmful but consumes more maintenance resources. Figure 4 explains that the sparseness of the RMSE and score function towards higher values is less than the others towards zero, which confirms the credibility of the results.
Experimental results lead to the parameters shown in Table 1.
A set of training mini batches was ignored during training of the RUL prediction models with the contribution of the USS strategy. This can contribute to training time reduction and avoiding divergence of training weights.
Figure 5 illustrates the reaction of USS strategy towards the useless mini batches by showing the indexes of ignored and selected mini batches. It is obvious that USS conditions shown previously in Equation (9) are confirmed in Figure 5. The RMSE error of each selected chunk of data is bigger than the one before it.
During training, the DFF by the intervention of TD error minimization function-based RLS can contribute to the prediction algorithm adaptation towards variations in the newly arrived data. By adapting weights gradually to enhance the approximation function, the accuracy of the prediction will remain stable and always converges to the desired value. Figure 6 shows the variation of DFF during training. If the TD changes too much, it means that the new coming training data has also different characteristics from the first ones, which consequently forces the training weights β to adapt towards these new changes. The adaptation of β is a result of updating the forgetting factor λ, and the amount of changes is controlled by the user hyper-parameters expressed in Table 1.
The training model was tested using new data (unknown to the model) from a test set of FD001. Additionally, the example of curve fitting of the RUL target function in Figure 7 proves that the proposed approach has an acceptable performance. The response of the designed network in Figure 7 is plotted using a set of different life cycles of different engines from the test set.
The proposed approach performances are compared with a set of other recent approaches in the literature. The results from Table 2 indicate that the proposed algorithm has the ability to achieve a low score value depending on less training samples and less training time.

5. Conclusions

A new online sequential training scheme is presented in this work for the RUL prediction of aircrafts engines. The proposed training model explained that by accepting any training mini batch of time-varying data in every update without taking into consideration the divergence of the weights at a certain level, the training model will no longer behave positively to attain the needed approximation and generalization for newly arrived data.
So far, the proposed USS strategy is involved in over-fitting reduction and accurate feature selection. The DFF-based TD error estimation plays a key role in weights adaptation between the current and the next training mini batches. The accurate features mapping based on robust unsupervised learning can solve the higher dimensional data problems. The experimental results prove that the new learning scheme for higher dimensional dynamic data is fast and accurate making it a competitive candidate for time-varying data prediction problems.
In the literature, C-MAPPS dataset is studied and investigated with different machine learning models. One of the important issues that has not yet been discussed is that sensors measurements are contaminated with noise of unknown sources and behavior. Therefore, future investigations will be focused on enhancing the proposed approach accuracy considering noisy measurements.

Author Contributions

Conceptualization, T.B. and L.-H.M.; Methodology, T.B. and L.-H.M.; Software, T.B.; Validation, T.B., L.-H.M., O.K., L.S., and M.B.; Formal Analysis, T.B., L.-H.M., O.K., L.S., and M.B.; Investigation, T.B.; Writing—Original Draft Preparation, T.B.; Writing—Review & Editing, T.B., L.-H.M., O.K., L.S., and M.B.; Supervision, L.-H.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

C-MAPSSCommercial Modular Aero Propulsion System Simulation
DFFDynamic Forgetting Factor
ELMExtreme Learning Machine
NNNeural Networks
OS-ELMOnline Sequential ELM
RLSRecursive Least Squares
RMSERoot Mean Squared Error
RULRemaining Useful Life
SAEsStacked Autoencoders
SVMSupport Vector Machine
TDTemporal Difference
USSUpdated Selection Strategy
AInput weight matrix
BIVector of biases
CRegularization parameter
dPrediction difference
ePrediction error (matrix)
GActivation function
HHidden layer
kIndex of recursive learning
KGain matrix of RLS
lHidden nodes number
nSize of mini batch
NSize of training data
SScore function
TTarget samples
ZTesting samples number
(T)Transpose matrix
βOutput weights
λForgetting factor
γDiscounting factor of TD error
δTemporal difference
μSensitivity factor
(−1)Pseudo inverse of the matrix

References

  1. Babu, G.S. Deep Convolutional Neural Network Based Regression Approach for Estimation of Remaining Useful Life. Lect. Notes Comput. Sci. 2016, 9642, 214–228. [Google Scholar]
  2. Xiong, X.; Li, Q.; Cheng, N. Remaining useful life prognostics of aircraft engine based on fusion algorithm. In Proceedings of the 2016 IEEE Chinese Guidance, Navigation and Control Conference (CGNCC), Nanjing, China, 12–14 August 2017; pp. 628–633. [Google Scholar]
  3. Nieto, P.J.G.; García-Gonzalo, E.; Lasheras, F.S.; Juez, F.J.D. Hybrid PSO-SVM-based method for forecasting of the remaining useful life for aircraft engines and evaluation of its reliability. Reliab. Eng. Syst. Saf. 2015, 138, 219–231. [Google Scholar] [CrossRef]
  4. Saidi, L.; Ali, J.B.; Bechhoefer, E.; Benbouzid, M. Wind turbine high-speed shaft bearings health prognosis through a spectral Kurtosis-derived indices and SVR. Appl. Acoust. 2017, 120, 1–8. [Google Scholar] [CrossRef]
  5. Zheng, C.; Liu, W.; Chen, B.; Gao, D.; Cheng, Y.; Yang, Y.; Peng, J. A Data-driven Approach for Remaining Useful Life Prediction of Aircraft Engines. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems (ITSC), Maui, HI, USA, 4–7 November 2018; pp. 184–189. [Google Scholar]
  6. Ordóñez, C.; Lasheras, F.S.; Roca-Pardiñas, J.; Juez, F.J.d. A hybrid ARIMA–SVM model for the study of the remaining useful life of aircraft engines. J. Comput. Appl. Math. 2019, 346, 184–191. [Google Scholar] [CrossRef]
  7. Chen, Z.; Cao, S.; Mao, Z. Remaining useful life estimation of aircraft engines using a modified similarity and supporting vector machine (SVM) approach. Energies 2018, 11, 28. [Google Scholar] [CrossRef] [Green Version]
  8. Ali, J.B.; Chebel-Morello, B.; Saidi, L.; Malinowski, S. Accurate bearing remaining useful life prediction based on Weibull distribution and artificial neural network. Mech. Syst. Signal Process. 2014, 1–23. [Google Scholar]
  9. Al-Dulaimi, A.; Zabihi, S.; Asif, A.; Mohammadi, A. A multimodal and hybrid deep neural network model for Remaining Useful Life estimation. Comput. Ind. 2019, 108, 186–196. [Google Scholar] [CrossRef]
  10. Wen, L.; Dong, Y.; Gao, L. A new ensemble residual convolutional neural network for remaining useful life estimation. Math. Biosci. Eng. 2019, 16, 862–880. [Google Scholar] [CrossRef] [PubMed]
  11. Li, X.; Ding, Q.; Sun, J.Q. Remaining useful life estimation in prognostics using deep convolution neural networks. Reliab. Eng. Syst. Saf. 2018, 172, 1–11. [Google Scholar] [CrossRef] [Green Version]
  12. Huang, G.B. An Insight into Extreme Learning Machines: Random Neurons, Random Features and Kernels. Cognit. Comput. 2014, 6, 376–390. [Google Scholar] [CrossRef]
  13. Zhou, H.; Huang, G.-B.; Lin, Z.; Wang, H.; Soh, Y.C. Stacked Extreme Learning Machines. IEEE Trans. Cybern. 2014, 99, 1. [Google Scholar]
  14. Ben, J.; Harrath, S.; Bechhoefer, E.; Benbouzid, M. Online automatic diagnosis of wind turbine bearings progressive degradations under real experimental conditions based on unsupervised machine learning Online automatic diagnosis of wind turbine bearings progressive degradations under real experimental con. Appl. Acoust. 2018, 132, 167–181. [Google Scholar] [CrossRef]
  15. Li, Y.; Zhang, S.; Yin, Y.; Xiao, W.; Zhang, J. A novel online sequential extreme learning machine for gas utilization ratio prediction in blast furnaces. Sensors 2017, 17, 8. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Yang, L.; Yang, S.; Li, S.; Liu, Z.; Jiao, L. Incremental laplacian regularization extreme learning machine for online learning. Appl. Soft Comput. J. 2017, 59, 546–555. [Google Scholar] [CrossRef]
  17. Lu, F.; Wu, J.; Huang, J.; Qiu, X. Aircraft engine degradation prognostics based on logistic regression and novel OS-ELM algorithm. Aerosp. Sci. Technol. 2019, 84, 661–671. [Google Scholar] [CrossRef]
  18. Yin, B.; Dridi, M.; El Moudni, A. Recursive least-squares temporal difference learning for adaptive traffic signal control at intersection. Neural Comput. Appl. 2019, 31, 1013–1028. [Google Scholar] [CrossRef]
  19. Huang, G.; Member, S.; Zhou, H.; Ding, X.; Zhang, R. Extreme Learning Machine for Regression and Multiclass Classification. IEEE Trans. Syst. Man Cybern. Part B 2012, 42, 513–529. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Saxena, A.; Saad, A. Evolving an artificial neural network classifier for condition monitoring of rotating mechanical systems. Appl. Soft Comput. J. 2007, 7, 441–454. [Google Scholar] [CrossRef]
  21. Saxena, A.; Goebel, K.; Simon, D.; Eklund, N. Damage propagation modeling for aircraft engine run-to-failure simulation. In Proceedings of the 2008 IEEE ICPHM, Denver, CO, USA, 12–15 October 2008; pp. 1–9. [Google Scholar]
  22. Heimes, F.O. Recurrent neural networks for remaining useful life estimation. In Proceedings of the 2008 IEEE ICPHM, Denver, CO, USA, 12–15 October 2008; pp. 1–6. [Google Scholar]
  23. Huang, G.B.; Liang, N.Y.; Rong, H.N.; Saratchandran, P.; Sundararajan, N. On-line sequential extreme learning machine. In Proceedings of the 2005 IASTED ICCI, Calgary, AB, Canada, 4–6 July 2005; pp. 1–6. [Google Scholar]
  24. Lan, Y.; Soh, Y.C.; Huang, G.B. A Constructive Enhancement for Online Sequential Extreme Learning Machine. In Proceedings of the 2009 International Joint Conference on Neural Networks, Atlanta, GA, USA, 14–19 June 2009; pp. 1708–1713. [Google Scholar]
  25. Gu, Y.; Liu, J.; Chen, Y.; Jiang, X. Constraint Online Sequential Extreme Learning Machine for lifelong indoor localization system. In Proceedings of the 2014 International Joint Conference on Neural Networks, Beijing, China, 6–11 July 2014; pp. 732–738. [Google Scholar]
  26. Kiakojoori, S.; Khorasani, K. Dynamic neural networks for gas turbine engine degradation prediction, health monitoring and prognosis. Neural Comput. Appl. 2016, 27, 2157–2192. [Google Scholar] [CrossRef]
  27. Berghout, T. Autoencoders. 2019. Available online: https://www.mathworks.com/matlabcentral/fileexchange/66080-autoencoders (accessed on 21 October 2019).
Figure 1. (a) Normalized sensors measurement. (b) Corresponding RUL to sensors measurement.
Figure 1. (a) Normalized sensors measurement. (b) Corresponding RUL to sensors measurement.
Applsci 10 01062 g001
Figure 2. Flowchart of the proposed RUL prediction approach.
Figure 2. Flowchart of the proposed RUL prediction approach.
Applsci 10 01062 g002
Figure 3. Proposed NN (Neural Networks) for RUL prediction.
Figure 3. Proposed NN (Neural Networks) for RUL prediction.
Applsci 10 01062 g003
Figure 4. Scoring functions for the testing set results.
Figure 4. Scoring functions for the testing set results.
Applsci 10 01062 g004
Figure 5. The behavior of USS (Updated Selection Strategy) strategy towards the unused training mini batches.
Figure 5. The behavior of USS (Updated Selection Strategy) strategy towards the unused training mini batches.
Applsci 10 01062 g005
Figure 6. Dynamic forgetting function behavior during training.
Figure 6. Dynamic forgetting function behavior during training.
Applsci 10 01062 g006
Figure 7. RUL prediction function compared to desired function.
Figure 7. RUL prediction function compared to desired function.
Applsci 10 01062 g007
Table 1. Training parameters of the proposed algorithm.
Table 1. Training parameters of the proposed algorithm.
CλminλmaxγμlnG
1000.98110 × 10−60.98100205RELU
Table 2. Performance results compared to other recent RUL prediction approaches.
Table 2. Performance results compared to other recent RUL prediction approaches.
MethodsRMSEScoreTraining SamplesTraining Time [s]
ARIMA SVM [6] 39.6843-20631-
DCNN [1] 18.44801286.720631-
LSTM [5] 16.1733820631714.53
DCNN [11] 12.61273.717731-
WELM [5] 13.78267.31206315.04
HDNN [9] 13.01724520631-
Basic OS-ELM15.08221.7220631-
ResCNN [10] 12.16212.4820631-
Proposed Approach13.74199.1712102505.85

Share and Cite

MDPI and ACS Style

Berghout, T.; Mouss, L.-H.; Kadri, O.; Saïdi, L.; Benbouzid, M. Aircraft Engines Remaining Useful Life Prediction with an Improved Online Sequential Extreme Learning Machine. Appl. Sci. 2020, 10, 1062. https://doi.org/10.3390/app10031062

AMA Style

Berghout T, Mouss L-H, Kadri O, Saïdi L, Benbouzid M. Aircraft Engines Remaining Useful Life Prediction with an Improved Online Sequential Extreme Learning Machine. Applied Sciences. 2020; 10(3):1062. https://doi.org/10.3390/app10031062

Chicago/Turabian Style

Berghout, Tarek, Leïla-Hayet Mouss, Ouahab Kadri, Lotfi Saïdi, and Mohamed Benbouzid. 2020. "Aircraft Engines Remaining Useful Life Prediction with an Improved Online Sequential Extreme Learning Machine" Applied Sciences 10, no. 3: 1062. https://doi.org/10.3390/app10031062

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop