Next Article in Journal
Comparative Analysis of a New Class of Symmetric and Asymmetric Supercapacitors Constructed on the Basis of ITO Collectors
Previous Article in Journal
Empirical Study on the Low-Carbon Economic Efficiency in Zhejiang Province Based on an Improved DEA Model and Projection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Gas Turbine Model Identification Based on Online Sequential Regularization Extreme Learning Machine with a Forgetting Factor

College of Power Engineering, Naval University of Engineering, Wuhan 430033, China
*
Authors to whom correspondence should be addressed.
Energies 2023, 16(1), 304; https://doi.org/10.3390/en16010304
Submission received: 23 November 2022 / Revised: 11 December 2022 / Accepted: 14 December 2022 / Published: 27 December 2022

Abstract

:
Due to the advantages of high convergence accuracy, fast training speed, and good generalization performance, the extreme learning machine is widely used in model identification. However, a gas turbine is a complex nonlinear system, and its sampling data are often time-sensitive and have measurement noise. This article proposes an online sequential regularization extreme learning machine algorithm based on the forgetting factor (FOS_RELM) to improve gas turbine identification performance. The proposed FOS_RELM not only retains the advantages of the extreme learning machine algorithm but also enhances the learning effect by rapidly discarding obsolete data during the learning process and improves the anti-interference performance by using the regularization principle. A detailed performance comparison of the FOS_RELM with the extreme learning machine algorithm and regularized extreme learning machine algorithm is carried out in the model identification of a gas turbine. The results show that the FOS_RELM has higher accuracy and better robustness than the extreme learning machine algorithm and regularized extreme learning machine algorithm. All in all, the proposed algorithm provides a candidate technique for modeling actual gas turbine units.

1. Introduction

Gas turbine (GT) modeling is necessary to investigate and improve gas turbine performance [1]. Without an accurate mathematical model, the dynamic characteristics of the unit cannot be learned before the GT is installed, and the design of the control system cannot be carried out. The modeling of a gas turbine can be used to predict engine performance [2], fault diagnosis [3], and control strategy investigation [4]. An accurate mathematical model must genuinely reflect the gas turbine’s steady-state and dynamic response characteristics.
The methods of gas turbine modeling are divided into two categories: one of which is to build a component-level model based on the internal operation mechanism, energy and momentum conservation, and other physical laws of the engine. The other one is to identify the mathematical model of the engine based on the actual data during operation. Due to the consideration of physical laws, component-level modeling can explain the operating rules of gas turbines. Still, the internal complex mathematical derivation process also limits the scope of application of component-level modeling. The model identification method does not require coupling the complex internal aerodynamic thermal mechanism of an actual unit. It characterizes the operating characteristics as a mapping relationship between input and output parameters to create a black box model describing the system’s dynamic properties. The method has the effect of model dimensionality reduction compared to the component-level model and therefore has better real-time performance. Asgari et al. [5] developed a NARX-ANN algorithm for the offline identification of low-power gas turbines. The results showed that the NARX-ANN model outperformed the physics-based model. Talaat et al. [6] published an identification method for predicting a gas turbine’s performance degradation using a feedforward ANN algorithm, but this method is not applicable to control system design. Jurado F. [7] developed a Hammerstein model for a micro gas turbine, and the results showed that the proposed Hammerstein model could accurately describe the system’s dynamic characteristics. In this article, we focus on the application of model identification in the field of micro gas turbines.
Single hidden layer feedforward networks (SLFNs) can approximate any continuous function defined on a compact set and have been widely used in model identification. Traditional SLFNs training methods are usually based on gradients, such as backpropagation (BP) [8] and neuron-by-neuron algorithms [9]. However, because of the need to adjust all parameters of SLFNs during training, these algorithms expose the drawback of slow convergence when identifying objects in high-dimensional spaces. In addition, gradient-based algorithms are more accessible to converge to local extremes due to the overdependence on parameter initialization and the complexity of the feature space. To solve the above problems, Huang et al. [10] proposed a new training method for SLFNs called the extreme learning machine (ELM). Unlike gradient-based learning methods, an ELM randomly assigns the network’s input weights and the hidden layer’s biases, and the nonlinear activation function of the hidden layer guarantees the system’s nonlinearity. Therefore, the ELM algorithm must only train the weights between the hidden and output layers. This improvement reduces the computational burden of weight adjustment caused by the widely used gradient descent method and ensures the convergence speed of an ELM. At the same time, random hidden nodes guarantee the global convergence of the algorithm [11].
Due to its special training effect, the ELM has been widely used in time series analysis [12], control [13], system modeling [14] and prediction [15], fault detection and diagnosis [16], and so on. Although the ELM has been successful in practical applications, its unique learning approach, which brings many advantages, inevitably causes disadvantages as well. By only considering empirical risk minimization, an ELM may provide weaker stability, generalization performance, and sparsity results. Random initialization of the hidden layer node parameters makes the ELM converge faster but also leads to fluctuations in the identification effect. The performance of an ELM is highly dependent on the number of hidden layer nodes. Incorrect selection of the number of hidden layer nodes may lead to overfitting or underfitting [17]. Output weights are another parameter related to the performance of an ELM. It is easier to direct to overfitting because an ELM seeks the output weight values by minimizing the training error [11]. Deng et al. [18] pointed out that the ELM algorithm is based on the empirical risk minimization (ERM) principle and tends to overfitting. According to statistical learning theory, an ELM with good generalization performance should consider structural risk minimization (SRM) instead of ERM, which requires controlling the model complexity of the ELM.
In 1963, Tikhonov proposed a new method for solving the ill-posed problem known as regularization. To avoid the occurrence of overfitting, the regularization principle is often used to improve the generalization ability of an ELM. The regularization principle reduces the complexity of ELM networks by applying the norm of the output weights to the objective function, which avoids the occurrence of overfitting. Zhao et al. [14] proposed a robust regularized ELM to solve an unsatisfactory modeling problem of the original ELM in the presence of outliers and noisy environments. Huang et al. [19] introduced the L2 norm penalty term into the ELM algorithm, and the results show that the model minimizes the empirical risk.
Initially, ELM algorithms were designed for offline learning. However, offline learning does not address the time-varying nature presented by the gas turbine. The online sequential extreme learning machine (OS_ELM) proposed by Liang et al. [20] can address the above problem precisely. The OS_ELM acquires samples in the form of data blocks and retrains the entire dataset containing the old samples to purchase the output weights of the ELM when new samples arrive. In general, with the arrival of new samples, the old sample data contribute less and less to the model, but the traditional OS_ELM uses the same weights for the whole dataset. To reduce the impact of old samples on model training, an extreme learning machine with a forgetting mechanism algorithm was proposed in Ref. [21]. This algorithm improves the learning effect by quickly discarding the outdated data so that the weights of the old sample data are set to zero. In addition, Wang et al. [22] introduced the sliding window concept into the OS_ELM algorithm, which moves forward with time and gradually forgets the old sample dates.
The actual operating parameters of a gas turbine have two significant characteristics: On the one hand, due to the influence of measurement errors, operation errors, and changes in operating conditions, the acquired sample data usually contain noisy interference signals, and the presence of noise signals is more likely to lead to the occurrence of overfitting. On the other hand, the characteristics of a gas turbine are varied by environmental conditions, performance degradation, and other factors. Studies have shown that gas turbine performance is directly related to ambient temperature [23]. Suppose the ambient temperature is used as a state variable for identification. In that case, a large number of data containing all possible temperature ranges is required, and often the acquisition of sample data is difficult.
In this paper, we combine the advantages of the regularization principle and online learning mechanism to propose a regularized online sequential extreme learning machine based on an adaptive forgetting factor (FOS_RELM). Our main contributions include the following:
  • We improve the ELM algorithm using the Tikhonov regularization principle to reduce the complexity of the model and thus improve the algorithm’s performance in handling noisy data.
  • An ELM is a batch learning algorithm while data acquisition for gas turbines is a continuous process, and it is impossible to obtain the complete data set at one time. When some new data arrive, batch learning has to repeat the training with old and new data, so it takes a lot of time. This paper improves the ELM algorithm by using an online learning method to process the data sequentially, which saves training time.
  • A forgetting factor adaptive update strategy is designed to forget the old sample data selectively. This forgetting mechanism can eliminate the adverse effects of old data which do not match with current characteristics of model training and solve the problem of low training accuracy caused by the time-varying nature of gas turbines.
This article is organized as follows: Section 2 describes the basic structure of an ELM and regularized ELM (RELM). Section 3 introduces the principle of our proposed FOS_RELM algorithm. In Section 4, a comprehensive demonstration of the FOS_RELM algorithm is presented. Section 5 summarizes this paper’s work, and the application prospects of the FOS_RELM algorithm are discussed.

2. The Traditional Algorithms

This section briefly introduces the principles of the traditional ELM algorithm and RELM algorithm.
The extreme learning machine is a method for training single hidden layer feedforward networks (SLFNs). As shown in Figure 1, SLFNs consist of three parts: the input layer, the hidden layer, and the output layer.
With a training sample S = { ( x i , t i ) } i = 1 N , where x i = [ x i 1 , x i 2 , x i m ] R m is the ith input vector, and t i = [ t i 1 , t i 2 , t i n ] R n is the ith target vector. SLFNs with L hidden nodes can be described as Equation (1).
f ^ ( x ) = i = 1 L β i G ( a i , b i , x ) a i , x R m , b i R
where x R m is the input parameter of the SLFNs, β i is the output weight coefficient connecting the hidden layer and the output layer, G ( ) is the nonlinear activation function, and a i , b i are the randomly generated node parameters of the hidden layer.
The ELM algorithm reduces the training parameters by randomly initializing the parameters of hidden layer nodes so that the network between the input and hidden layers becomes a fixed nonlinear transformation. After the fixed nonlinear transformation of the hidden layer, the input parameters are mapped to a new linear space named the ELM feature space [24]. The ELM trains the network by minimizing the estimated and actual values error. Thus, the objective loss function of the model is shown as follows:
min β : H β T
where
H = [ G ( a 1 , b 1 , x 1 ) G ( a L , b L , x 1 ) G ( a 1 , b 1 , x N ) G ( a L , b L , x N ) ]
is the hidden nodes output matrix, β = [ β 1 , β 2 , β L ] T , and T = [ t 1 , t 2 , t N ] T . Combining Equations (1) and (2), the optimization problem of ELM can be expressed as Equation (4).
min β { J ELM = 1 2 E F 2 } T = H β + E
where E = [ e 1 , e 2 , e N ] T is the modeling error matrix, e i is the modeling error of the ith sample, and · F represents the Frobenius norm.
By solving Equation (4) using the least squares method, the expression for the output weight factor β is obtained as follows:
β = { ( H T H ) 1 H T T N L H T ( H H T ) 1 T N < L
Equation (3) minimizes the identification error on the training set as the objective function and trains the parameters of the network with the least squares principle. In this approach, only the empirical risk minimization is considered, and it is easy to overlearn the features of the training set when performing model identification, which results in overfitting. Deng et al. [18] used the regularization condition to improve the extreme learning machine, forming a regularized extreme learning machine (RELM) that integrates SRM and ERM. The objective function of RELM can be expressed as Equation (6).
min β f ( β ) = 1 2 β 2 + C 2 H β T = 1 2 β T β + C 2 ( H β T ) T ( H β T )
where C > 0 is the regularization factor for balancing the output weight matrix and the modeling error. Solving Equation (6) by setting the partial derivative of the function to zero gives the following expression:
f ( β ) β = β C H T ( T H β ) = 0 β = { ( H T H + I C ) 1 H T T N L H T ( H H T + I C ) 1 T N < L
where I is the unit matrix. It can be seen from Equation (7) that the RELM degenerates to an ELM when C . Hence, the ELM algorithm can be considered a particular case of the RELM algorithm. By adding the examination of the output weight coefficient β to the objective function of the RELM algorithm, we can finally obtain a model with a small value of β . It has been shown in [25] that the smaller the output weight β , the better the generalization performance of the network.

3. The Proposed Algorithm

The traditional ELM and gradient-based discrimination algorithms are offline discrimination methods. Offline identification methods can achieve satisfactory modeling results when sufficient sample excitation is obtained, and the system characteristics keep unchanged with time. However, a gas turbine is a complex power system with nonlinearity and high time lag. During actual operation, characteristics of a gas turbine will be changed by the external environment and performance degradation. For advanced control strategies, such as predictive and adaptive control, real-time acquisition of the object’s model is also required. Traditional identification methods can only establish a single mapping relationship, which cannot solve the time-varying problem of the gas turbine. Fortunately, the OS_ELM algorithm proposed by Liang et al. [20] can solve the above issues well. In this paper, we combine the advantages of the OS_ELM and the regularization principle to propose an FOS_RELM algorithm that applies to gas turbine model identification.
As shown in Figure 2, the training process for the FOS_RELM is mainly divided into a prior knowledge learning phase and an online updating phase. In the prior knowledge learning phase, the FOS_RELM algorithm trains the model according to the RELM algorithm. The online update phase updates the output weight coefficients of the model using new sample dates. The detailed steps of the FOS_RELM are described below.
4.
Prior knowledge learning phase:
The initialized training set S 0 = { ( x i , t i ) | x i R m , t i R n , i = 1 , 2 , 3 N 0 } is formed by taking N0 samples from the sample set. Initialized training samples should satisfy N0 > L. Consider regularization, and train the initialized output layer weight coefficients by Equation (8).
β 0 = ( H 0 T H 0 + I C ) 1 H 0 T T 0
Let P 0 = H 0 T H 0 + I C to obtain β 0 = P 0 1 H 0 T T 0 .
5.
Online updating phase:
When a new sample ( x N 0 + 1 , t N 0 + 1 ) is available, the implied layer matrix h 1 is expressed as Equation (9).
h 1 = [ G ( a 1 , b 1 , x N 0 + 1 ) G ( a L , b L , x N 0 + 1 ) ]
Therefore, the output weights can be calculated as follows:
β 1 = ( [ H 0 h 1 ] T [ H 0 h 1 ] + I C ) 1 [ H 0 h 1 ] T [ T 0 t 1 ] = ( H 0 T H 0 + h 1 T h 1 + I C ) 1 ( H 0 T T 0 + h 1 T t 1 )
Let P 1 = H 0 T H 0 + h 1 T h 1 + I C P 1 = P 0 + h 1 T h 1 .
From β 0 = P 0 1 H 0 T T 0 P 0 β 0 = H 0 T T 0 , the expression for β 1 can be rewritten as Equation (11).
β 1 = ( [ H 0 h 1 ] T [ H 0 h 1 ] + I C ) 1 [ H 0 h 1 ] T [ T 0 t 1 ] = P 1 1 ( ( P 1 h 1 T h 1 ) β 0 + h 1 T t 1 ) = β 0 + P 1 1 h 1 T ( t 1 h 1 β 0 )
From this, it can be deduced that the updated formula for the output weight coefficient β is shown in Equation (12).
{ P 0 = H 0 T H 0 + I C P k + 1 = P k + h k + 1 T h k + 1 β k + 1 = β k + P k + 1 1 h k + 1 T ( t k + 1 h k + 1 β k )
As seen in Equation (12), the updating process of the output weight coefficient β records the information of the whole sample. The data of the old sample will affect model identification accuracy when the features of the gas turbine are changed. Zhao et al. [21] modified the structure of the OS_ELM with the forgetting mechanism and defined the concept of the time window to distinguish the old and new observations. Observations in the time window are used to update the network, and the old sample data outside the time window are forgotten. However, this approach to bypassing the sample data cannot accurately reflect how much the training samples contribute to the network. In addition, it will decrease the utilization of sample data or affect the recognition accuracy of the network when the time window is not set correctly. The training error is the most accurate parameter to describe the contribution of data samples. If a small training error can still be achieved for the new sample data, it is proven that the original network can be well adapted to the latest data [24]. In this paper, we use the forgetting factor λ to adjust the weight values of the old and new samples. A strategy for adaptively changing the forgetting factor according to the training error e was designed, as shown in Figure 3. The training error is defined as e = 1 N i = 1 N ( t i t ^ i ) 2 . Defining a standard training error e t . The required adjustment strategy should be set with e t as the critical point for forgetting the old sample date. The correction function of the forgetting factor λ is shown in Equation (13). When e < e t , it is considered that the model has achieved a high identification effect at this time so that λ = 1. When e e t , it is proven that the model identified by the old sample data has failed to describe the characteristics of the new sample data. At this time, λ should tend to the set minimum forgetting factor λ min . Figure 3 is the schematic diagram of the correction function.
λ = { 1 0 < e e t 2 ( 1 λ min ) π [ π 2 atan ( e e t ) ] + λ min e t < e
where atan ( ) is the arctangent function. The updated formula of the output weight coefficient β is shown in Equation (14).
{ P 0 = H 0 T H 0 + I C P k + 1 = λ P k + h k + 1 T h k + 1 β k + 1 = β k + P k + 1 1 h k + 1 T ( t k + 1 h k + 1 β k ) ( 0 < λ 1 )
The forgetting factor λ determines the proportion of old and new samples in the online identification process, as seen from Equation (14). It indicates that all sample data are considered in the identification process when λ = 1. When λ < 1, the old sample data are gradually forgotten as the iterations increase.
The pseudocode of the proposed FOS_RELM algorithm is shown below (Algorithm 1).
Algorithm 1: FOS_RELM
Input: { ( x i , t i ) } i = 1 N , N0 ,   L ,   C ,   e t ,   λ min
Output: β
  • Random initialization hidden layer coefficients a i , b i .
  • Select initialize training dataset S 0 = { ( x i , t i ) | x i R m , t i R n , i = 1 , 2 , 3 N 0 } .
  • Calculate hidden layer output matrix H0 by Equation (2).
  • Calculate initialization output layer weight β 0 by Equation (8).
k = 1.
while ( x N 0 + k , t N 0 + k ) { ( x i , t i ) } i = 1 N do
5.
Calculate the hidden layer matrix h k with the new sample ( x N 0 + k , t N 0 + k ) by h k = [ G ( a 1 , b 1 , x N 0 + k ) G ( a L , b L , x N 0 + k ) ] .
6.
Calculate the objective function estimate t ^ N 0 + k = h N 0 + k β k 1 and Training error e.
7.
Update the value of λ through Equation (13).
8.
The updated value of the output weight coefficient β k can be obtained by Equation (14).
9.
k = k + 1.
end

4. Results and Discussion

4.1. Performance Evaluation of Simulation

In this section, the component-level mechanism model of the gas turbine is used to obtain sample data to verify the rationality and feasibility of the proposed FOS_RELM algorithm.

4.1.1. Input Signal Selection

During model training, the choice of the input excitation signal will directly affect the quality of the model [26]. The pseudorandom binary signal (PRBS) is a periodic signal with white noise characteristics, and it is usually used as an input signal in the model identification of linear systems. However, a gas turbine is a complex nonlinear system, and a PRBS does not cover all its relevant frequencies and attractive spaces [27]. Nelles et al. [28] improved the PRBS by assigning different amplitudes in the range Amin to Amax to each constant signal part to form the amplitude tunable pseudorandom binary signal (APRBS). Figure 4a shows that APRBS for Amin = 0 to Amax = 1. It has been demonstrated that an APRBS is the best excitation signal to obtain information about the process behavior of nonlinear systems [29]. It is worth noting that the minimum dwell time should be at least equal to the stability time of the system so that the system output has time to approach the new set point. Otherwise, only wrong output amplitudes will be generated. On the other hand, a very long minimum dwell time will lead to quasistatic excitation. Therefore, the minimum dwell time was selected to be 5 s in this paper.
In the actual operation of a gas turbine, we usually change the operating state of the gas turbine by controlling the load. We use the gas turbine power signal command to obtain the sample data as the input signal for the mechanism model simulation. The training samples were obtained using the input excitation, as shown in Figure 4a. To avoid the influence of the sample data on the identification effect, we used the data set mutually exclusive with the training set as the testing set. As shown in Figure 4b, we selected the combustion engine load command from no-load to 0.4 operating conditions to obtain the testing set. The sampling time was 0.01 s.

4.1.2. Network Architecture Design

To fully retain the system’s dynamic characteristics, the identified model should contain current and historical information on the essential state parameters. In this paper, the present and historical information on fuel, turbine output speed N t , exhaust temperature T e , and power P t were selected as input signals. The current moment information of N t , T e , and P t was used as the output signal to establish the structure of the identification model. All the state parameters are normalized. The model can be described as Equation (15).
T = f ( X ) { X = [ W f ( k ) W f ( k m 1 ) , N t ( k 1 ) N t ( k m 2 ) T e ( k 1 ) T e ( k m 3 1 ) , P t ( k 1 ) P t ( k m 4 1 ) ] T = [ N t ( k ) , T e ( k ) , P t ( k ) ]
To ensure the model’s accuracy, we need as much information about the input parameters as possible. However, excessive input parameters will increase the structural complexity of the network and quickly lead to the overfitting phenomenon. After debugging, we took m 1 = m 2 = m 3 = m 4 = 2 in this paper. The sigmoid function G ( x ) = 1 / ( 1 + exp ( x ) ) was chosen as the activation function. The number of nodes in the hidden layer was taken as L = 100 .
It can be seen in Equations (7) and (12), the value of the regularization coefficient C will affect the model’s learning accuracy. To understand the influence law of C on model identification, we compared the identification accuracy of the RELM algorithm and traditional ELM network when C takes different values. Parameter C took values in the set { 2 20 , 2 19 2 24 } , which has 45 elements. To avoid the influence of accidental factors, each algorithm was run 40 times independently, and the final results were judged by the root mean square error (RMSE). The comparison results are shown in Figure 5.
RMSE = 1 N i = 1 N ( t i t ^ i ) 2
As shown in Figure 5, when the regularization parameter C is smaller, the performance of the RELM identification network is worse. When C ∈ [210 224], the RELM’s training and testing error of the output speed N t and exhaust temperature T e are smaller than the ELM network, and both the RELM and ELM have the same identification accuracy of the power P t . When C = 219, the RELM has the highest identification accuracy for the output speed signal. From Figure 5b, we can see that the RELM has the smallest training error for the exhaust temperature signal when C = 221 and the minimum testing error when C = 219. Based on the comprehensive analysis of the results in Figure 5, it is concluded that the RELM performs best at C = 219. Therefore, parameter C was taken as C = 219 in the subsequent calculations and extended to the FOS_RELM algorithm. The parameters of the FOS_RELM algorithm are set as shown in Table 1.
The model identification results of the ELM, RELM, and FOS_RELM algorithms at C = 219 are recorded in Table 2. From Table 2, we can see that the FOS_RELM has the best identification effect, and the ELM has the worst impact. The ELM algorithm and RELM algorithms’ training times are similar since they have the same number of nodes in the hidden layer. The FOS_RELM requires iterative updates of the discriminated parameters, so the I’FOS_RELM has the longest training time.

4.1.3. Robustness Testing

The performance of a gas turbine is directly related to external environmental conditions [23]. When the temperature increases, the air density becomes less, and the characteristic line of the compressor is shifted, which changes the gas turbine’s thermal efficiency and power output. To verify the ability of the proposed FOS_RELM algorithm to cope with external environmental changes. The ambient temperature was changed from 22.6 °C to 17 °C to compare the identification effects of the FOS_RELM with the other algorithms.
The parameters of the FOS_RELM algorithm are chosen as follows: λ min = 0.98 , e t = 0.01 , C = 2 19 .
The model identification effect is shown in Figure 6. After the change in ambient temperature, both the ELM and RELM algorithms failed to describe the system’s dynamic characteristics accurately. In contrast, the FOS_RELM algorithm could still follow the changes in the system state parameters accurately. From this result, we can see that this method of online updating network parameters adopted by the FOS_RELM algorithm has a significant advantage in identifying the nonlinear system with time-varying nature, such as a gas turbine.
Further, to verify the improvement effect of the adopted adaptive update forgetting factor and regularization parameters, the FOS_RELM, FOS_ELM, and OS_ELM algorithms were used to identify the state parameters of the gas turbine when the ambient temperature was reduced from 26 °C to 17 °C, respectively. The operating data of the gas turbine at 17 °C were used for verification. The results are shown in Table 3. The proposed FOS_RELM algorithm has the highest recognition accuracy followed by the FOS_ELM algorithm, and the OS_ELM obtained the worst result.

4.2. Performance Evaluation of the Experiment

Section 4.1 compares the identification effects of the FOS_RELM algorithm with other algorithms at the simulation level. However, in actual operation, the placement of sensors and measurement errors can affect the data samples we obtain. The most obvious one is that noise interference signals are inevitably introduced in the measurement process. The noise signal coupled with the gas turbine’s characteristic data will affect the model identification’s effectiveness. The mechanical model does not capture these. Therefore, to evaluate the ability of the proposed FOS_RELM algorithm to resist noise signal interference, the measured data of the gas turbine was used as sample data to compare the identification effect of the FOS_RELM algorithm and the conventional OS_ELM algorithm.
The experimental micro gas turbine unit is shown in Figure 7. It contains a single-stage centrifugal compressor, a single-cylinder combustion chamber, a two-stage axial turbine, and a three-phase AC synchronous generator. Gas turbines support the combustion of various fuels [30,31]. For this article, we used diesel. The detailed parameters of the micro gas turbine are shown in Table 4. The experimental platform utilizes SUPCON’s real-time monitoring software AdvanTrol-Pro(V2) to record the experimental data. The AdvanTrol-Pro software records sampled data from the sensors at a frequency of one datum per second. Due to the inhomogeneity of the gas turbine temperature field, we used three thermocouple sensors to collect the gas turbine temperature signals. The average of the three sensor signals was used as the final output signal. We used magneto-electrical speed sensors to measure the output speed of the turbine. The experimental steps of this paper are as follows:
Process: Firstly, check whether the fuel and lubricating oil systems are standard, and determine whether the unit startup conditions are met. Secondly, start the gas turbine. The unit can realize two kinds of triggering modes, manual and automatic, and we used the automatic mode to accelerate the gas turbine to the no-load state. Thirdly, learn the variable working condition of the gas turbine operation. The gas turbine adopts the constant speed control mode; at the moment of load switching, the speed changes due to the power mismatch of the gas turbine, and the control system stabilizes the speed signal by adjusting the fuel injection. Finally, after all data measurements are completed, switch the gas turbine to shutdown mode.

4.2.1. Performance Test of Noise Resistance

To verify the ability of the FOS_RELM algorithm to resist noise interference, we independently ran the gas turbine twice with the same external conditions. The sampled data were selected as the training and testing sets, respectively. It is worth noting that the selected sample data exclude the startup phase data. The training set contains 1830 data points, as shown in Figure 8. The testing set includes 400 data points, as shown in Figure 9. All data are normalized. It can be seen that the sample data contain random noise signals.
The identification algorithm has 100 hidden nodes, and the sigmoid function was chosen as the activation function. The main parameters of the FOS_RELM algorithm were determined by debugging as λ min = 0.98 , e t = 0.01 , C = 2 10 .
Figure 10a–c depicts the effectiveness of the OS_ELM and FOS_RELM algorithms in identifying the gas turbine output speed, turbine exhaust temperature, and output power, respectively. In this paper, the sample data were obtained from the actual operation of the micro gas turbine. During the stable operation, the output speed of the gas turbine was the rated value, one after dimensionless. The speed of a gas turbine will fluctuate when the operating condition is changed. As shown in Figure 10a, the FOS_RELM can eliminate the interference of noise signals and accurately describe the dynamic characteristics of the gas turbine. However, the OS_ELM incorrectly treats the noise signal as an inherent characteristic of the gas turbine, which also appears as an overfitting phenomenon. As shown in Figure 10c, the stability of the OS_ELM algorithm is poor, and the OS_ELM does not accurately identify the output power when the operating condition of the gas turbine is changed. Compared with the OS_ELM algorithm, the FOS_RELM algorithm reduces the complexity of the model by using the regularization condition. From the perspective of numerical analysis, the FOS_RELM algorithm can obtain more stable results than the OS_ELM algorithm by reducing the value of the output weight coefficients.

4.2.2. Robustness Testing

The experimental unit shown in Figure 7 was run on 8 September 2021 (the ambient temperature was 28.5 °C) and 23 December 2021 (the ambient temperature was 14 °C) to obtain two operational data sets. The dynamic characteristics in the process of changing gas turbine operating conditions can affect the actions of the control system. If the identified parameters are not in line with the output parameters of the actual unit, it will mislead the controller to take wrong actions and lead to severe accidents. Therefore, we chose the operating data of the gas turbine variable condition process as the sample data for model identification. The model was first trained using the December sample data. Next, 800 data samples from the data collected in September were selected to update the model online. Finally, 202 samples from the data contained in September were used to test the model.
We compare the identification results of the OS_ELM and FOS_RELM in Figure 11. As shown in Figure 11, the FOS_RELM can still accurately describe the dynamic characteristics of the gas turbine when the environmental conditions changed. The speed of the gas turbine first dropped sharply and then slowly rose to the rated value during the ramp-up condition, as shown in Figure 11a. The FOS_RELM algorithm accurately tracks the speed variation law of the gas turbine. However, the OS_ELM algorithm was not accurate in estimating the variation law of the speed. From Figure 11c, it can be seen that the OS_ELM algorithm works poorly in identifying the power P t . The reason is that the old sample data cannot represent the gas turbine after changing external environmental conditions. However, the OS_ELM continues to use the old sample data in the updating process. However, the adaptive forgetting factor strategy applied in the FOS_RELM algorithm provides more attention to new sample observations which can characterize the gas turbine properties. As new sample data becomes available, the old observations are gradually forgotten, and the identification model can keep following the state changes of the gas turbine.

5. Conclusions

Compared with the traditional gradient descent method, the extreme learning machine has apparent advantages in computational efficiency and convergence accuracy. Still, it cannot handle time-varying and measurement noise in the data. Hence, a FOS_RELM algorithm was proposed in this paper. On the one hand, the Tikhonov regularization principle was used to control the complexity of the model and reduce the interference of noise signals to the model. On the other hand, a forgetting factor adaptive adjustment strategy was designed to gradually discard obsolete data in learning and reduce their adverse effects on the network. The FOS_RELM algorithm also inherits the advantages of the extreme learning machine in terms of training time and convergence accuracy.
To illustrate the effectiveness and feasibility of the proposed algorithm, validation experiments were conducted using a gas turbine simulation dataset. The results show that the identification accuracy of the proposed FOS_RELM algorithm is higher than ELM and RELM algorithms. When the gas turbine characteristics change, the identification accuracy of ELM and RELM algorithms deteriorates while the FOS_RELM algorithm still has a better identification effect. More importantly, when it is used to model a natural micro gas turbine system, it shows good robustness and stability and can maintain good identification accuracy when the training data are affected by noise interference and time variability. The proposed algorithm has a broad application prospect which can be used for both onboard modeling and fault diagnosis of the gas turbine.

Author Contributions

R.Y.: conceptualization, methodology, formal analysis, writing—original draft. Y.L.: methodology, visualization, supervision, funding acquisition. X.H.: investigation, data curation, software. Z.L.: writing editing, supervision, validation. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been supported by the Basic Research for National Science and Technology Major Project of China (Grant No. J2019-I-0012).

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

LNumber of nodes in the hidden layer
a i , b i The randomly generated node parameters of the hidden layer
β The output weight coefficient
G ( ) The nonlinear activation function
HThe hidden nodes output matrix
EThe modeling error matrix
e i The modeling error of the ith sample
mThe input nodes
CThe regularization factor
λThe forgetting factor
e t The standard training error
λ min The set minimum forgetting factor
W f Gas turbine fuel flow
N t Turbine output speed
T e Exhaust temperature
P t Turbine output power
STraining data set
TTarget value
S0The initialized training set
N0Number of initial training samples
NNumber of the training samples
RMSEThe root mean square error
FOS_RELMOnline sequential regularization extreme learning machine algorithm based on forgetting factor
GTGas turbine
SLFNsSingle hidden layer feedforward networks
ELMExtreme learning machine
OS_ELMOnline sequential extreme learning machine
ERMEmpirical risk minimization
SRMStructural risk minimization
RELMRegularized extreme learning machine
SVMsSupport vector machines
BPBackpropagation

References

  1. Mohamed, O.; Khalil, A. Progress in modeling and control of gas turbine power generation systems: A survey. Energies 2020, 13, 2358. [Google Scholar] [CrossRef]
  2. Liu, Y.; Meenakshi, V.; Karthikeyan, L.; Maroušek, J.; Krishnamoorthy, N.; Sekar, M.; Nasif, O.; Alharbi, S.A.; Wu, Y.; Xia, C. Machine learning based predictive modelling of micro gas turbine engine fuelled with microalgae blends on using LSTM networks: An experimental approach. Fuel 2022, 322, 124183. [Google Scholar] [CrossRef]
  3. Hanachi, H.; Mechefske, C.; Liu, J.; Banerjee, A.; Chen, Y. Performance-based gas turbine health monitoring, diagnostics, and prognostics: A survey. IEEE Trans. Reliab. 2018, 67, 1340–1363. [Google Scholar] [CrossRef]
  4. Mohamed, O.; Wang, J.; Khalil, A.; Limhabrash, M. Predictive control strategy of a gas turbine for improvement of combined cycle power plant dynamic performance and efficiency. SpringerPlus 2016, 5, 980. [Google Scholar] [CrossRef] [Green Version]
  5. Asgari, H. Modelling, Simulation and Control of Gas Turbines Using Artificial Neural Networks. Ph.D. Theis, University of Canterbury, Christchurch, New Zealand, 2014. [Google Scholar] [CrossRef]
  6. Talaat, M.; Gobran, M.H.; Wasfi, M. A hybrid model of an artificial neural network with thermodynamic model for system diagnosis of electrical power plant gas turbine. Eng. Appl. Artif. Intell. 2018, 68, 222–235. [Google Scholar] [CrossRef]
  7. Jurado, F. Modelling micro-turbines using Hammerstein models. Int. J. Energy Res. 2005, 29, 841–855. [Google Scholar] [CrossRef]
  8. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  9. Wilamowski, B.M.; Cotton, N.J.; Kaynak, O.; Dundar, G. Computing gradient vector and Jacobian matrix in arbitrarily connected neural networks. IEEE Trans. Ind. Electron. 2008, 55, 3784–3790. [Google Scholar] [CrossRef]
  10. Huang, G.-B.; Zhu, Q.; Siew, C. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No. 04CH37541), Budapest, Hungary, 25–29 July 2004; Volume 2. [Google Scholar] [CrossRef]
  11. Wang, J.; Lu, S.; Wang, S.-H.; Zhang, Y.-D. A review on extreme learning machine. Multimed. Tools Appl. 2021, 81, 41611–41660. [Google Scholar] [CrossRef]
  12. Cao, F.; Liu, B.; Park, D.S. Image classification based on effective extreme learning machine. Neurocomputing 2013, 102, 90–97. [Google Scholar] [CrossRef]
  13. Wong, K.I.; Wong, P.K. Adaptive air-fuel ratio control of dual-injection engines under biofuel blends using extreme learning machine. Energy Convers. Manag. 2018, 165, 66–75. [Google Scholar] [CrossRef]
  14. Zhao, Y.-P.; Hu, Q.-K.; Xu, J.-G.; Li, B.; Huang, G.; Pan, Y.-T. A robust extreme learning machine for modeling a small-scale turbojet engine. Appl. Energy 2018, 218, 22–35. [Google Scholar] [CrossRef]
  15. Berghout, T.; Mouss, L.-H.; Kadri, O.; Saïdi, L.; Benbouzid, M. Aircraft engines remaining useful life prediction with an improved online sequential extreme learning machine. Appl. Sci. 2020, 10, 1062. [Google Scholar] [CrossRef] [Green Version]
  16. Creech, G.; Jiang, F. The application of extreme learning machines to the network intrusion detection problem. In AIP Conference Proceedings; American Institute of Physics: College Park, MD, USA, 2012; Volume 1479, No. 1. [Google Scholar] [CrossRef]
  17. Martínez-Martínez, J.M.; Escandell-Montero, P.; Soria-Olivas, E.; Martín-Guerrero, J.D.; Magdalena-Benedito, R.; Gómez-Sanchis, J. Regularized extreme learning machine for regression problems. Neurocomputing 2011, 74, 3716–3721. [Google Scholar] [CrossRef]
  18. Deng, W.; Zheng, Q.; Chen, L. Regularized extreme learning machine. In Proceedings of the 2009 IEEE Symposium on Computational Intelligence and Data Mining, Nashville, TN, USA, 30 March–2 April 2009. [Google Scholar] [CrossRef]
  19. Huang, G.-B.; Zhou, H.; Ding, X.; Zhang, R. Extreme learning machine for regression and multiclass classification. IEEE Trans. Syst. Man Cybern. Part B 2011, 42, 513–529. [Google Scholar] [CrossRef] [Green Version]
  20. Liang, N.-Y.; Huang, G.-B.; Saratchandran, P.; Sundararajan, N. A fast and accurate online sequential learning algorithm for feedforward networks. IEEE Trans. Neural Netw. 2006, 17, 1411–1423. [Google Scholar] [CrossRef]
  21. Zhao, J.; Wang, Z.; Park, D.S. Online sequential extreme learning machine with forgetting mechanism. Neurocomputing 2012, 87, 79–89. [Google Scholar] [CrossRef]
  22. Wang, X.; Han, M. Online sequential extreme learning machine with kernels for nonstationary time series prediction. Neurocomputing 2014, 145, 90–97. [Google Scholar] [CrossRef]
  23. De Sa, A.; Al Zubaidy, S. Gas turbine performance at varying ambient temperature. Appl. Therm. Eng. 2011, 31, 2735–2739. [Google Scholar] [CrossRef]
  24. Han, M.; Liu, C. Endpoint prediction model for basic oxygen furnace steel-making based on membrane algorithm evolving extreme learning machine. Appl. Soft Comput. 2014, 19, 430–437. [Google Scholar] [CrossRef]
  25. Bartlett, P.L. The sample complexity of pattern classification with neural networks: The size of the weights is more important than the size of the network. IEEE Trans. Inf. Theory 1998, 44, 525–536. [Google Scholar] [CrossRef] [Green Version]
  26. Salehi, A.; Montazeri-Gh, M. Black box modeling of a turboshaft gas turbine engine fuel control unit based on neural NARX. Proc. Inst. Mech. Eng. Part M J. Eng. Marit. Environ. 2019, 233, 949–956. [Google Scholar] [CrossRef]
  27. Deflorian, M.; Zaglauer, S. Design of experiments for nonlinear dynamic system identification. IFAC Proc. Vol. 2011, 44, 13179–13184. [Google Scholar] [CrossRef]
  28. Nelles, O.; Isermann, R. Identification of nonlinear dynamic systems classical methods versus radial basis function networks. In Proceedings of the 1995 American Control Conference-ACC’95, Seattle, WA, USA, 21–23 June 1995; Volume 5. [Google Scholar]
  29. Nelles, O.; Isermann, R. Basis function networks for interpolation of local linear models. In Proceedings of the 35th IEEE Conference on Decision and Control, Kobe, Japan, 13 December 1996; Volume 1. [Google Scholar]
  30. Boomadevi, P.; Paulson, V.; Samlal, S.; Varatharajan, M.; Sekar, M.; Alsehli, M.; Elfasakhany, A.; Tola, S. Impact of microalgae biofuel on microgas turbine aviation engine: A combustion and emission study. Fuel 2021, 302, 121155. [Google Scholar] [CrossRef]
  31. Tan, D.; Wu, Y.; Lv, J.; Li, J.; Ou, X.; Meng, Y.; Lan, G.; Chen, Y.; Zhang, Z. Performance optimization of a diesel engine fueled with hydrogen/biodiesel with water addition based on the response surface methodology. Energy 2023, 263, 125869. [Google Scholar] [CrossRef]
Figure 1. The structure of SLFNs.
Figure 1. The structure of SLFNs.
Energies 16 00304 g001
Figure 2. The flow chart of the proposed FOS_RELM.
Figure 2. The flow chart of the proposed FOS_RELM.
Energies 16 00304 g002
Figure 3. Schematic diagram of the forgetting factor update function.
Figure 3. Schematic diagram of the forgetting factor update function.
Energies 16 00304 g003
Figure 4. (a) Training set input signal; (b) testing set input signal.
Figure 4. (a) Training set input signal; (b) testing set input signal.
Energies 16 00304 g004
Figure 5. RMSE of the identification model when C is taken for different values: (a) output speed signal; (b) exhaust temperature signal; (c) output power signal.
Figure 5. RMSE of the identification model when C is taken for different values: (a) output speed signal; (b) exhaust temperature signal; (c) output power signal.
Energies 16 00304 g005
Figure 6. The identification effects of FOS_RELM, RELM, and ELM after changing the ambient temperature.
Figure 6. The identification effects of FOS_RELM, RELM, and ELM after changing the ambient temperature.
Energies 16 00304 g006
Figure 7. Experimental micro gas turbine unit.
Figure 7. Experimental micro gas turbine unit.
Energies 16 00304 g007
Figure 8. Gas turbine model training data set: (a) Speed signal; (b) Exhaust temperature, fuel flow and power load signals.
Figure 8. Gas turbine model training data set: (a) Speed signal; (b) Exhaust temperature, fuel flow and power load signals.
Energies 16 00304 g008
Figure 9. Gas turbine model testing data set.
Figure 9. Gas turbine model testing data set.
Energies 16 00304 g009
Figure 10. The modeling results with OS_ELM and FOS_RELM: (a) Speed signal. (b) Exhaust temperature signal. (c) Power load signal.
Figure 10. The modeling results with OS_ELM and FOS_RELM: (a) Speed signal. (b) Exhaust temperature signal. (c) Power load signal.
Energies 16 00304 g010
Figure 11. The modeling results with OS_ELM and FOS_RELM under different environmental conditions: (a) Speed signal. (b) Exhaust temperature signal. (c) Power load signal.
Figure 11. The modeling results with OS_ELM and FOS_RELM under different environmental conditions: (a) Speed signal. (b) Exhaust temperature signal. (c) Power load signal.
Energies 16 00304 g011
Table 1. FOS_RELM algorithm parameter settings.
Table 1. FOS_RELM algorithm parameter settings.
FOS_RELM
L100
G ( ) sigmoid
C219
e t 0.01
λ min 0.98
Table 2. Model identification effects of ELM, RELM, and FOS_RELM algorithms.
Table 2. Model identification effects of ELM, RELM, and FOS_RELM algorithms.
ELMRELMFOS_RELM
Training RMSE N t 1.30 × 10−49.46 × 10−66.60 × 10−6
T e 1.12 × 10−42.78 × 10−52.67 × 10−5
P t 0.02260.02260.0226
Testing RMSE N t 1.29 × 10−41.16 × 10−51.01 × 10−5
T e 1.12 × 10−41.55 × 10−51.46 × 10−5
P t 0.01150.01150.0115
Training time (s)0.08090.08790.6609
Table 3. Identification effects of FOS_RELM, FOS_ELM, and OS_ELM algorithms.
Table 3. Identification effects of FOS_RELM, FOS_ELM, and OS_ELM algorithms.
AlgorithmOS_ELMFOS_ELMFOS_RELM
Testing RMSE N t 1.06 × 10−59.71 × 10−68.93 × 10−6
T e 2.18 × 10−41.17 × 10−41.19 × 10−4
P t 0.0110.0110.011
Training time (s)25.8225.5226.08
Table 4. Basic parameters of the micro gas turbine unit.
Table 4. Basic parameters of the micro gas turbine unit.
Rated SpeedRated Output PowerRated Power Generation EfficiencyDimensions (Length × Width × Height) Compressor TypeTurbine TypeCombustion Chamber TypeWeight
51,000 rpm125 kW15%2950 × 2200 × 2450 mm3 Single-stage centrifugal typeTwo-stage axial flow typeSingle cylinder type4.3 t
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, R.; Liu, Y.; He, X.; Liu, Z. Gas Turbine Model Identification Based on Online Sequential Regularization Extreme Learning Machine with a Forgetting Factor. Energies 2023, 16, 304. https://doi.org/10.3390/en16010304

AMA Style

Yang R, Liu Y, He X, Liu Z. Gas Turbine Model Identification Based on Online Sequential Regularization Extreme Learning Machine with a Forgetting Factor. Energies. 2023; 16(1):304. https://doi.org/10.3390/en16010304

Chicago/Turabian Style

Yang, Rui, Yongbao Liu, Xing He, and Zhimeng Liu. 2023. "Gas Turbine Model Identification Based on Online Sequential Regularization Extreme Learning Machine with a Forgetting Factor" Energies 16, no. 1: 304. https://doi.org/10.3390/en16010304

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop