Next Article in Journal
Influence of Eddy-Generation Mechanism on the Characteristic of On-Source Fire Whirl
Previous Article in Journal
A Novel Activity Recognition System for Alternative Control Strategies of a Lower Limb Rehabilitation Robot
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved Autoencoder and Partial Least Squares Regression-Based Extreme Learning Machine Model for Pump Turbine Characteristics

1
College of Automation, Huaiyin Institute of Technology, Huaian 223003, China
2
School of Hydropower and Information Engineering, Huazhong University of Science and Technology, Wuhan 430074, China
3
Division of Informatization Construction & Management, Huaiyin Institute of Technology, Huaian 223003, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(19), 3987; https://doi.org/10.3390/app9193987
Submission received: 20 August 2019 / Revised: 15 September 2019 / Accepted: 20 September 2019 / Published: 23 September 2019
(This article belongs to the Section Energy Science and Technology)

Abstract

:

Featured Application

Authors are encouraged to provide a concise description of the specific application or a potential application of the work. This section is not mandatory.

Abstract

Complete characteristic curves of a pump turbine are fundamental for improving the modeling accuracy of the pump turbine in a pump turbine governing system. In view of the difficulty in modeling the “S” characteristic region of the complete characteristic curves in the pump turbine, a novel Autoencoder and partial least squares regression based extreme learning machine model (AE-PLS-ELM) was proposed to describe the pump turbine characteristics. First, a mathematical model was formulated to describe the flow and moment characteristic curves. The improved Suter transformation was employed to transfer the original curves into WH and WM curves. Second, the ELM-Autoencoder technique and the partial least squares regression (PLSR) method were introduced to the architecture of the original ELM network. The ELM-Autoencoder technique was employed to obtain the initial weights of the Autoencoder based extreme learning machine (AE-ELM) model. The PLS method was exploited to avoid the multicollinearity problem of the Moore-Penrose generalized inverse. Lastly, the effectiveness of the proposed AE-PLS-ELM model has been verified using real data from a pumped storage unit in China. The results demonstrated that the AE-PLS-ELM model can obtain better modeling accuracy and generalization performance than the traditional models and, thus, can be exploited as an effective and sufficient approach for the modeling of pump turbine characteristics.

1. Introduction

As the demand for electricity and the requirements for developing a low-carbon economy continues, the driving force for energy development will gradually shift to renewable and clean energy such as photovoltaic power and wind power [1,2,3]. Due to its fast start-stop speed and flexible working conditions, the pumped storage units (PSUs) can quickly respond to the requirements of the power system for frequency and phase modulation, peak load shifting, rotation, and accident reserve, which can enhance the grid’s ability to absorb wind power and photoelectricity [4,5]. The pump turbine governing system (PTGS) is a complex hydraulic-mechanical-electrical-magnetic coupling system, which is time-varying, stochastic, and nonlinear [6,7]. The precise modeling of the PTGS is of great significance in analyzing the dynamic response of PSUs under complex operational conditions [8,9].
As a crucial part of PSU, an accurate pump turbine model is the key to the accurate modeling and simulation of PTGS [10]. In the current research work, the modeling of a pump turbine is mainly based on complete characteristic curves where different operational conditions of this model can be described [11,12,13]. The complete curves of pump turbine include the flow characteristic curve and the moment characteristic curve. These curves contain the inherent nonlinear characteristics of the pump turbine under various normal operational conditions or transient working conditions [14,15]. However, due to the existence of the “S” characteristic and hump area, the estimation of complete characteristic curves is highly complex [16,17]. It is easy for the complete characteristic curves to gather, cross, and twist in the pumps and anti-pump regions, which cause a huge interpolation error and non-convergence problem to the model when calculating the transient parameters of the runner boundary of the pump turbine. An in-depth and extensive research study has been conducted to extend the complete characteristic curves in recent years. The common methods to estimate the complete curves of pumped turbine mainly include the assistant mesh processing [18], the Suter transformation, its improved versions [19,20], and the 3D surface fitting [14].
The assistant mesh processing method divides the flow characteristic curve or moment characteristic curve into multiple sections by introducing the idea of piecewise linearization, and then the original curve is approximated using small line segments. An assistant mesh line, which is approximately orthogonal to the opening line, is drawn to facilitate calculation. The characteristic curves of the pump turbine have been obtained and the modeling accuracy has been improved through the assistant mesh processing technique [8]. However, the above method requires prior artificial meshing of which the workload is heavy and the piecewise approximation also causes errors in the orthogonal curvilinear grid.
The Suter transformation proposed by Suter [19] transformed the original complete characteristic curves into WH(x, y) and WM(x, y) curves using dimensionless parameters. The abscissa of the curves is determined according to the relative flow and the relative speed. Because the Suter transformation stretches the flow and moment characteristic curves to both sides of the coordinate axis, the difficulties of the interpolation calculation caused by curve bending, torsion, anti-S, and multi-value characteristics can be alleviated effectively. However, scholars have found that the complete curves using Suter transformation or improved Suter transformation still have the following drawbacks: the length of the opening lines is different, some opening lines have singularities and are not conductive, and the density of data points on the opening lines is different. These non-uniform curves will affect the interpolation accuracy and computational efficiency of the model. Therefore, it is necessary to establish neural network models for curve extension, bad point correction, and encryption processing of opening lines for the complete characteristic curves.
According to whether the flow and moment characteristic curves are pre-transformed, the 3D surface fitting technique can be mainly divided into two categories: the first category is based on the original complete curves of the pump turbine. The original complete curves are fitted using the least squares or neural network models directly. The second category is based on the pre-processed curves using Suter transformation. The pre-processed curves are then fitted using the least squares or neural network models. Li et al. [14] applied a backpropagation (BP) neural network to process the synthetic characteristic curve for the Francis turbine. Actual engineering applications validate that the proposed method can provide not only higher precision data for the transition process but also better investigation on the real hydraulic unit. Zhang et al. [21] applied a radial basis function (RBF) to process the synthetic characteristic curve of the Kaplan turbine. The results have demonstrated that the nonlinear model based on RBF networks can reflect the nonlinear operating characteristics of the Kaplan turbine with higher accuracy. Liu et al. [20] first employed a modified Suter transformation to pre-process the complete characteristic curve and then proposed an Adaboost-BP neural network ensemble model optimized by particle swarm optimization to describe the WH and WM characteristics of the pump turbine. The results show that the proposed model can obtain higher fitting accuracy and better generalization performance than a single BP neural network model.
Although the neural network model can fit and extend the complete characteristic curves and facilitate the calculation of the flow and moment characteristic curves for the pump turbine, the convergence speed of the traditional neural network model is slow and it is easy to fall into local minimum. As a novel single hidden layer feedforward neural network, the extreme learning machine (ELM) obtained the input weights and hidden layer biases through random initialization [22]. The output weights are directly obtained by calculating the generalized inverse matrix of the hidden layer output matrix [23,24]. The convergence rate is far faster than the traditional BP neural network. The ELM has been widely used in pattern recognition, statistical prediction, and classification and regression [25]. However, because of the random initialization strategy of input weights, the common ELM model fails to make full use of the inherent characteristics of the training data. In addition, the Moore-Penrose generalized inverse used in the ELM model may produce pathological solutions. Therefore, there exists multicollinearity in the output matrix, which affects the fitting and generalization performance of the model [26]. Therefore, there still exists much room to improve in describing pump turbine characteristics using ELM.
To improve the modeling accuracy of the pump turbine in the simulation of PTGS, an Autoencoder and partial least squares regression based extreme learning machine model (AE-PLS-ELM) is proposed to model the pumped turbine of a PSU. With the strong fitting ability of the AE-PLS-ELM model, the non-linear mapping relationship of the full characteristic curves of pump turbines is represented. The flow and moment characteristic curves of pump turbines are transformed into neural network models, which can be used for a real-time simulation. On the basis of curve pretreatment with improved Suter transformation, two AE-PLS-ELM models are used to model the characteristic curves. The automatic encoder technique (AE) and partial least squares regression algorithm (PLSR) are introduced to improve the performance of the ELM model. The rest of this paper is arranged as follows. Section 2 describes the model of pump turbine based on characteristic curves, Section 3 proposes an AE-PLS-ELM model, Section 4 provides the specific modeling process of the pump turbine characteristics based on the proposed AE-PLS-ELM model, Section 5 employs a numerical example to verify the performance of AE-PLS-ELM, Section 6 provided an additional test problem, and Section 7 gives the conclusions.

2. Nonlinear Modeling of the Pump Turbine

The most common method for the pump turbine nonlinear modeling is through the complete characteristic curves [14]. The main idea of the nonlinear modeling based on complete curves is to first extract certain discrete data points from the practical curves, and then the extracted data points are fitted or extended to obtain the modeling curves [14]. The mathematical model of the flow and moment characteristics to express the pump turbine characteristics is as follows [2].
{ M 11 = f M ( a , N 11 ) Q 11 = f Q ( a , N 11 )
where M11 represents the unit moment, Q11 denotes the unit flow, N11 is the unit speed, and a denotes the guide vane opening. In this study, the pump turbine complete characteristic curves of a hydropower station in China are employed as a case study. The practical complete characteristic curves are shown in Figure 1 as follows [2].
As can be seen in Figure 1, the original completed curves still have a significant twist curl when the unit speed is bigger than 80, which causes the multi-value phenomenon in the “S” characteristic area. For example, three flow (Figure 1a) and moment (Figure 1b) data points appear when the value of unit speed is 90 r/min and the value of the guide vane opening is 10. The interpolation error is large and the derivative is discontinuous when multiple values exist, which may cause an iteration error to occur in the PTGS. To avoid the multi-value phenomenon of the original curves, this study introduces the improved Suter transformation [20] to pre-process the original complete curves. The original flow and moment characteristic curves are changed to WH and WM curves, respectively, through the improved Suter transformation. The converting equations of the improved Suter transformation are expressed as follows.
{ W H ( x , y ) = h a 2 + q 2 + C h h ( y + C y ) 2 W M ( x , y ) = ( m + s 2 h ) a 2 + q 2 + C h h ( y + C y ) 2 x = arctan   [ ( q + s 1 h ) / a ] a 0 x = π + arctan   [ ( q + s 1 h ) / a ] a < 0
where a, q, h, and m denote the relative speed, relative flow, relative water head, and relative moment, respectively, x and y denote the relative flow angle and relative opening, respectively. s2 > |M11max|/M11r, s1 = 0.5~1.2, Cy = 0.1~0.3, and Ch = 0.4~0.6. The WH(x,y) and WM(x,y) curves based on an improved Suter transformation are given in Figure 2 as follows [2].

3. An Autoencoder and Partial Least Aquares Regression Based Extreme Learning Machine Model

An accurate pump-turbine model is the key to the modeling and simulation of PTGS. In this study, an autoencoder and partial least squares regression-based extreme learning machine model (AE-PLS-ELM) is proposed for the nonlinear modeling of pump turbine characteristics. The AE-PLS-ELM model is introduced in this section.

3.1. Extreme Learning Machine

ELM is a new type of single hidden layer forward neural network. It randomly generates the connection weights and biases between the input layer and the hidden layer. The connection weights and biases do not need to be adjusted during the training process. Once the number of hidden neurons is determined, the optimal solution can be obtained. Compared with the traditional training methods, ELM has the advantages of fast learning speed and good generalization performance [27].
Suppose that there are N samples ( x k , y k ) , k = 1 , 2 , , N , where x k R p and y k R q . The mathematical expression for an ELM model with L hidden neurons is as follows.
f L = i = 1 L β i g ( a i · x k + b i )   =   y ^ k k = 1 , 2 , , N
where y ^ k denotes the simulated output of the kth sample, a i and β i are the connection weights between the ith hidden neuron and the input layer and hidden layer, respectively, b i is the bias of the ith hidden neruon, and g ( ) is the activation function. Equation (3) can be reformulated below.
H β   =   Y ^
where
H = [ h ( x 1 )   h ( x N ) ] = [ g ( a 1 · x 1 + b 1 ) g ( a L · x 1 + b L ) g ( a 1 · x N + b 1 ) g ( a L · x N + b L ) ] N × L
β = [ β 1 T β L T ] L × m   a n d   Y = [ y 1 T y N T ] N × m
where H is known as the hidden layer output matrix, and β is the output weight matrix.
Based on the Moore-Penrose generalized inverse matrix theory, β can be calculated as:
β ^   =   H Y = ( H T H ) 1 H T Y
where H is the generalized inverse matrix of H . To improve the stability and the generalization of the ELM network, Huang et al. [28] added a positive constant 1 / C to the diagonal of H T H ( or   HH T ) . β can then be calculated as:
β = ( 1 C + H T H ) 1 H T Y

3.2. The ELM-Autoencoder Technique

The convergence speed of ELM is fast and the generalization capability is well compared with the traditional BP neural networks. However, the initial parameters of ELM are independent of the modeling data. Thus, the characteristics and internal relations of the modeling data cannot be effectively reflected. To obtain better initial parameters of ELM, the Autoencoder technology, which has been widely employed in deep learning, is introduced to ELM for modeling [29]. The traditional Autoencoder technology developed by Rumelhart et al. in 1986 is an unsupervised learning method based on a BP algorithm. The purpose of the Autoencoder technology is to approximate an identity function such that the output data is the same as the input data [30]. The ELM-Autoencoder technology based on the ELM algorithm is introduced in this study to avoid repeated iterative training of the BP network [31]. The Autoencoder based extreme learning machine (AE-ELM) proposed in this study can be implemented mainly in two stages. First, the ELM-Autoencoder technology is employed to establish the mapping relationship between X to X (X is the input data) using the ELM algorithm. The output weights of the ELM-Autoencoder is taken as the initial weights of the AE-ELM model. Second, the AE-ELM model, which takes X and Y (Y is the output data) as training data, is trained with the initial weights taken from the first stage.
The ELM-Autoencoder is a type of unsupervised network and the input weights and hidden biases should be orthogonal. Assume the number of input neurons is N p , and the number of the hidden neurons is N L . In this study, the number of input neurons is smaller than the number of hidden neurons. Therefore, a sparse ELM-Autoencoder architecture is adopted [32]. Given a set of N data samples, i.e., ( x k , y k ) for k = 1 , 2 , , N , the hidden layer output of the ELM-Autoencoder can be expressed as:
h ( x i ) = g ( a T x i + b ) i = 1 , 2 , , N
where h ( x i ) R N L denotes the hidden layer output with respect to the ith input, a T a = I ,   b T b = I , I is a unit matrix, a = [ a 1 , a 2 , , a N ] denotes the orthogonal weights connecting the input layer and the hidden layer, b = [ b 1 , b 2 , , b N ] denotes the orthogonal biases of the hidden nodes, and g ( ) is the activation function.
The output weight β is optimized to minimize the squared loss of the training error. The optimization problem can be expressed as the equation below.
min β O β = min β 1 2 β 2 + C 2 X H β 2
where X denotes the input data, H denotes the hidden layer outputs, and C denotes a penalty factor on the training error.
The first derivative of O β with respect to β can be denoted as:
Δ β = β C H T ( X H β ) = 0
The final hidden layer output weights can be calculated as:
β = ( I / C + H T H ) 1 H T X

3.3. Partial Least Squares Regression

PLSR developed by Wold is a multivariate statistical analysis method that takes advantage of both principal component analysis and the least square method [33]. Compared with the least square method, the PLSR method can deal with the multicollinearity problem. The PLSR method allows for regression modeling when the number of samples is smaller than the number of independent variables. In addition, the PLSR model considers information of both the independent and responsible variables, which makes it easier to identify system information and noises.
Given a set of N observed data samples, which were composed of p input and q output variables, i.e., S = ( x i , y i ) for i = 1 , 2 , , N , where x i R p denotes the independent variable and y k R q denotes the responsible variable. The objective of PLSR is modeling the linear relationship between the p independent variables and q response variables. The modeling process of PLSR can be denoted as follows [26].
First, the independent matrix X = ( x 1 , x 2 , , x p ) n × p and response matrix Y = ( y 1 , y 2 , , y q ) n × q are normalized into a zero mean and one variance. The normalized matrix of X and Y are denoted as E 0 and F 0 , respectively. Second, the PLSR method is applied to X and Y to extract the first pair of score vectors u 1 and v 1 , respectively. The score vectors u 1 and v 1 are the combination of the independent variable and responsible variables, respectively, and should contain the maximum variation information of them. The relationship between u 1 and v 1 should be as maximum as possible. Then the regression model between Y and u 1 is deduced. Third, the residual matrix E 1 and F 1 are calculated to replace E 0 and F 0 , respectively, to skip to the next step iteration until the residual matrix meets the stopping criteria.

3.4. The Proposed AE-PLS-ELM Model

In the Autoencoder based extreme learning machine (AE-ELM) model, the Moore-Penrose generalized inverse based on least square is exploited to calculate the output weights, which makes it possible to apply the PLSR method by replacing the least square method. The employment of PLSR in the AE-ELM model can avoid the multicollinearity problem when applying the Moore-Penrose generalized inverse especially when the hidden layer output matrix is highly correlated and contains noises [34]. Based on the description of AE-ELM and PLSR, the key to establish the AE-PLS-ELM model is to model the linear relationship between the hidden layer output matrix H and the output layer matrix Y using the partial peast squares regression (PLSR) method. As has been described in Section 3.2, the hidden layer output matrix H is an N × L dimension matrix with L hidden layer output variables. The output layer matrix Y is a N × q dimension matrix. The detailed modeling process of the AE-PLS-ELM model can be given as follows.
Step 1: Randomly assign the input weights a i , i = 1 , 2 , , n and hidden layer bias b i , i = 1 , 2 , , L .
Step 2: Calculate the hidden layer output matrix of the training data H, according to Equation (3).
Step 3: The hidden layer output matrix H is taken as the independent matrix of the PLSR model. Let E 0 = H and F 0 = Y .
Step 4: Extract the first pair of score vectors of E 0 and F 0 . The score vectors are denoted as u 1 and v 1 , respectively. u 1 and v 1 can be denoted below.
u 1 = E 0 ω 1 v 1 = F 0 c 1
where ω 1 = [ ω 11 , ω 12 , , ω 1 L ] T is the load factor of E 0 and ω 1 = 1 , and c 1 = [ c 11 , c 12 , , c 1 L ] T is the load factor of F 0 and c 1 = 1 .
The relevance between u 1 and F 0 should be as maximum as possible. Then u 1 and v 1 should satisfy the following two criterions.
(1) u 1 and v 1 should contain the maximum variation information of E 0 and F 0 .
(2) u 1 and v 1 should have maximum relevance.
The deduction of ω 1 and c 1 can be transformed to the following optimization problem.
max ( u 1 , v 1 ) = ( E 0 ω 1 , F 0 c 1 ) = ω 1 T E 0 T F 0 c 1 s . t . { ω 1 T ω 1 = ω 1 2 = 1 c 1 T c 1 = c 1 2 = 1
According to the Lagrange multiplier method, the following equation is true.
L = ω 1 T E 0 T F 0 c 1 λ 1 ( ω 1 T ω 1 1 ) λ 2 ( c 1 T c 1 1 )
The partial derivatives of L on ω 1 , c 1 λ 1 , and λ 2 can be expressed by the equation below.
{ L ω 1 = E 0 T F 0 c 1 2 λ 1 ω 1 = 0 L c 1 = F 0 T E 0 ω 1 2 λ 2 c 1 = 0 L λ 1 = ( ω 1 T ω 1 1 ) = 0 L λ 2 = ( c 1 T c 1 1 ) = 0
From Equation (16), it can be deducted that:
2 λ 1 = 2 λ 2 = ω 1 T E 0 T F 0 c 1 = E 0 ω 1 , F 0 c 1
Note that:
θ 1 = 2 λ 1 = 2 λ 2 = ω 1 T E 0 T F 0 c 1
θ 1 is the objective function of the optimization problem, and Equations (19) and (20) are true.
E 0 T F 0 c 1 = θ 1 ω 1
F 0 T E 0 ω 1 = θ 1 ω 1
From Equation (19) and Equation (20), it can be deducted that:
E 0 T F 0 F 0 T E 0 ω 1 = θ 1 2 ω 1
where ω 1 is the eigenvector of E 0 T F 0 F 0 T E 0 , θ 1 2 is the eigenvalue, and ω 1 and c 1 can be obtained according to Equation (19) and Equation (20). After obtaining ω 1 and c 1 , the score vectors u 1 and v 1 can be calculated according to Equation (13).
Step 5: Establish the linear regression model between E 0 and u 1 , and F 0 and v 1 according to the least square method.
{ E 0 = u 1 α 1 T + E 1 F 0 = v 1 γ 1 T + F 1
where α 1 = [ α 11 , α 12 , , α 1 L ] and γ 1 = [ γ 11 , γ 12 , , λ 1 L ] are regression coefficients, and E 1 and F 1 are the residual matrix. The least square estimation of α 1 and γ 1 can be denoted below.
{ α 1 = E 0 T u 1 u 1 2 , γ 1 = F 0 T v 1 v 1 2 .
Step 6: If F 1 satisfy the stopping criteria, Equation (22) is the final regression model and the iteration stops. Otherwise, replace E 0 and F 0 with E 1 and F 1 , respectively, and skip to Step 3 to get the second pair of score vectors.
u 2 = E 1 ω 2 v 2 = F 1 c 2
{ E 0 = u 1 α 1 T + u 2 α 2 T + E 2 F 0 = v 1 γ 1 T + v 2 γ 2 T + F 2
where α 2 and γ 2 are regression vectors and can be denoted as:
{ α 2 = E 1 T u 2 u 2 2 , γ 2 = F 1 T v 2 v 2 2 .
Step 7: Repeat Steps 4–5 until r principal components are calculated. The remaining m r components are small and are considered as noises. The residuals E r and F r are very small. E 0 and F 0 can be expressed by the equation below.
{ E 0 = u 1 α 1 T + u 2 α 2 T + + u r α r T + E r = U α T + E r , F 0 = v 1 γ 1 T + v 2 γ 2 T + + v r γ r T + F r = V γ T + F r
The relationship between u k and v k can be expressed by the equation below [1].
v k = u k b 1 ,   k = 1 , 2 , , r .
Then F 0 can be translated into:
F 0 = V γ T + F r = i = 1 r u 1 b 1 γ 1 T + u 2 b 2 γ 2 T + + u r b r γ r T + F r = U B γ T + F r
where U ^ = E 0 W , then the regression equation can be denoted as:
F ^ 0 = E 0 W B γ T + F r
According to the description above, the hidden layer output weight vector can be expressed by the equation below.
β ^ P L S = W B γ T
where W denotes the component matrix and B denotes the diagonal matrix.
Based on the above modeling process, the structure of the proposed AE-PLS-ELM model is shown in Figure 3 as follows.

4. Modeling Process of the Pump Turbine Based on AE-PLS-ELM

The proposed AE-PLS-ELM model is used to model the pump turbine characteristics of a PSU. The flow and moment characteristic curves of the pump turbine are preprocessed using an improve Suter transformation method. Two independent AE-PLS-ELM models are used to model the preprocessed complete curves. The preprocessed complete curves are then converted into target variables to construct a neural network model that can be used for real-time simulation. The input of the AE-PLS-ELM model is the relative flow angle x and the relative vane opening y. The full characteristic curves based on improved Suter transformation (Figure 2) are employed as data samples.
The specific steps of the modeling process of the pump turbine characteristics based on AE-PLS-ELM are as follows.
Step 1: Apply the improved Suter transformation to the complete characteristic curves of the pump turbine provided by the power station and the corresponding preprocessed WH and WM curves are obtained.
Step 2: Extract data points from the curves. Convert the relative flow angle x, the relative vane opening y, and the extracted data points into input and output sample pairs.
Step 3: Divide the above sample data into training data and test data. Since the dimensions and magnitudes of the data samples are different, the input and output data are normalized to facilitate the modeling and calculation process.
Step 4: Set the Sigmoid function as the activation function of the hidden layer and determine the range of the number of hidden layer nodes, according to the Kolmogorov empirical formula. The optimal number of hidden nodes is selected using a trial calculation modeling error.
Step 5: Import the normalized training data to the AE-PLS-ELM model for training, and a well-trained AE-PLS-ELM model is obtained.
Step 6: Import the normalized test data to the well-trained AE-PLS-ELM model and de-normalize the output data to obtain the test output.

5. Numerical Experiments and Analysis

In this section, a PSU in China is used as the research object to carry out the nonlinear modeling of the pump turbine [8]. A total number of 1125 data points are extracted from the pre-processed WH and WM characteristic curves using the improved Suter transformation (Figure 2). Thus, 1125 data pairs are generated for constructing an AE-PLS-ELM model. In addition, 90% of the data pairs (1012 points) are employed as the training samples and the remaining 10% (112 points) as the test sample.

5.1. Parameters Setting

To highlight the effectiveness of the proposed model, four conventional data-driven techniques named the Bagtree, the support vector regression (SVR), the BP neural network [35], and the ELM are employed as a control group to simulate the complete characteristic curve of the pump turbine. The Bagtree model is constructed using MATLAB’s “Bag” function. The parameters of the SVR model including the penalty factor C and the kernel parameter σ are obtained using the grid search (GS) algorithm. The search range of C is set as [2−8, 28], and the search range of σ is set as [2−5, 25]. The “trainlm” algorithm is employed in training the BP neural network. The maximum number of iterations is set as 500 and the target error is 1e-5. The number of the hidden nodes is selected using a trial-and-error method.

5.2. Comparative Analysis of the Results

To evaluate the performance of different models, four evaluation indices were used to include the root mean square error (RMSE), the mean absolute error (MAE), the mean absolute percent error (MAPE) [23,36], and the Nash-Sutcliffe Efficiency (NSE) [37] are employed. The four indices are defined as follows.
RMSE = 1 N i = 1 N ( f s ( i ) f o ( i ) ) 2
MAE = 1 N i = 1 N | f s ( i ) f o ( i ) |
MAPE = 1 N i = 1 N | f s ( i ) f o ( i ) | f o ( i ) × 100 %
NSE = 1 i = 1 N ( f s ( i ) f o ( i ) ) 2 i = 1 N ( f o ( i ) f o ( i ) ¯ ) 2
where f s ( i ) and f o ( i ) denote the simulated and observed value of the ith sample point, respectively, f o ( i ) ¯ denotes the mean observed value, and N denotes the size of the data set.
The 3D spatial surfaces for the WH and the WM characteristics based on the AE-PLS-ELM model are shown in Figure 4 and Figure 5, respectively. As can be seen from Figure 4 and Figure 5, the WH and WM spatial surfaces based on the AE-PLS-ELM model are smooth and uniform. In addition, the complete characteristic curves are continuous derivative, which make it easy to ensure the convergence of the water hammer calculation process. The AE-PLS-ELM model can also be used to encrypt and extend the WH and WM characteristic surfaces, according to the practical requirements for research and engineering applications. Therefore, the transition between different opening lines on the surface is smoother and it is convenient for operators to obtain the pump turbine characteristics of different working conditions.
The results of the Bagtree, SVR, BP, ELM, and AE-PLS-ELM models for WH characteristics are shown in Table 1. It can be seen from the table that the five models have good training and test accuracy and can model the WH characteristics of the PSU accurately. The AE-PLS-ELM model performs the best in terms of the four indices in both the training and test period among the five models, which indicates that the AE-PLS-ELM model can enhance the modeling accuracy of pump turbine for the WH characteristic curve effectively. Taking the RMSE value of the test period as an example, the RMSE value of ELM is 0.00297, which is lower than that of the Bagtree, SVR, and BP models. The neural network models as BP and ELM perform better than the Bagtree and SVR models. The performance of ELM is slightly better than BP and the Bagtree model performs the worst.
Compare the training and test results between the ELM and AE-PLS-ELM models, it can be found that the performance of the AE-PLS-ELM model is significantly better than the ELM model. For the training samples, the RMSE, MAE, MAPE, and NSE values of the AE-PLS-ELM model were 0.00229, 0.00130, 0.00599, and 0.99988, respectively, which were improved by 21.03%, 37.20%, 72.75%, and 0.007% compared with the 0.00290, 0.00207, 0.02198, and 0.99981 obtained by the ELM model. For the test samples, the RMSE, MAE, MAPE, and NSE values of the AE-PLS-ELM model were 0.00217, 0.00142, 0.00697, and 0.99988, respectively, which were improved by 26.93%, 39.32%, 71.75%, and 0.011% when compared with the 0.00297, 0.00234, 0.02467, and 0.99977 obtained by the ELM model. In a word, the proposed AE-PLS-ELM model overcomes the instability and multi-collinearity of the single ELM model, and can improve the generalization ability and fitting accuracy of the ELM for modeling the pump turbine characteristics.
The comparison of the residuals for the WH characteristics between the AE-PLS-ELM model and the Bagtree, SVR, BP, and ELM models are shown in Figure 6a–d, respectively. It can be seen from Figure 6 that the prediction accuracy of the AE-PLS-ELM model is significantly better than that of the Bagtree and SVR models at all test sample points. In addition, the residual of the AE-PLS-ELM model is generally smaller than the BP and ELM models at most of the test points.
The training and test results of the Bagtree, SVR, BP, ELM, and AE-PLS-ELM models for the WM characteristic are shown in Table 2. The comparison of the residuals for the WH characteristics between the AE-PLS-ELM model and the Bagtree, SVR, BP, and ELM models are shown in Figure 7a–d, respectively. The results obtained from Table 2 and Figure 7 are consistent with Table 1 and Figure 7. The neural network models that BP, ELM, and AE-PLS-ELM performed better than the Bagtree and the SVR models. The AE-PLS-ELM model can overcome the instability and multi-collinearity of the single ELM model, and obtain higher modeling accuracy.
The regression analysis scatter diagram of the residual and actual value for the WH and WM characteristics is shown in Figure 8. It can be seen from Figure 8 that the scatter plot of the single ELM model is more divergent around the axis, and the points of the AE-PLS-ELM model distributes more closely, which demonstrates the superiority of the AE-PLS-ELM model in modeling the complete characteristic curves of the pump turbine.

6. Additional Test Problem

To further demonstrate the effectiveness of the proposed AE-PLS-ELM model, a widely used nonlinear differential equation is studied as an additional test problem [38]. The nonlinear differential equation can be expressed by the equation below [9].
y ( k + 1 ) = y ( k ) y ( k 1 ) y ( k 2 ) u ( k 1 ) ( y ( k 2 ) 1 ) + u ( k ) 1 + y 2 ( k 2 ) + y 2 ( k 1 )
where u ( k ) is the control variable and is randomly generated in [−2,2] in the training period. Furthermore, 800 data pairs are generated for training. y ( k + 1 ) is taken as the output variable. y ( k ) , y ( k 1 ) , y ( k 2 ) , u ( k ) , u ( k 1 ) are taken as the input variables. In the test period, u ( k ) is generated using the following equation.
u ( k ) = { sin ( 3 π k / 250 ) , k 500 0.25 sin ( 2 π k / 250 ) + 0.2 sin ( 3 π k / 50 ) , k > 500
where k = 1 , 2 , , 800 , which means 800 data pairs are generated for testing. The RMSE, MAE, MAPE, and NSE for the test problem of the Bagtree, SVR, BP, ELM, and AE-PLS-ELM models are given Table 3. The test results of the five different models are shown in Figure 9. As can be seen in Table 3, the AE-PLS-ELM outperforms the other four models in terms of the four indices in the test period. The performance of the SVR model is the worst. The Bagtree model performed the second worst and the BP and ELM models perform better than the Bagtree model. It can also be observed from Table 3 that the SVR and ELM models encounter the overfitting problem during the test period. Their training performances are much better than the other models while the test performances are much worse.

7. Conclusions

PTGS plays an extremely important role in maintaining the safe and stable operation of the power system. However, it is a closed-loop control system with a complex structure, variable parameters, and strong nonlinearity. As a crucial part of PSU, an accurate pump turbine model is the key to the accurate modeling and simulation of PTGS. This study first introduced an improved Suter transformation to process the complete characteristic curves of the pump turbine. The crossing, aggregating phenomena, and multi-value problems in the “S” characteristic region of the pump turbine were reduced through the improved Suter transformation. Furthermore, an AE-PLS-ELM model was proposed to model the pump turbine characteristics precisely. The AE technique was introduced to the single ELM model for feature extraction of input data to improve its stability. In addition, the PLSR algorithm was employed to replace the Moore-Penrose generalized inverse in ELM to reduce the multicollinearity of the output weight. Results have shown that the proposed AE-PLS-ELM model has better fitting precision and generalization performance than traditional models such as Bagtree, SVM, BP, and ELM. Essentially, the proposed modeling framework is an effective technique in modeling the pump turbine characteristics and the proposed AE-PLS-ELM can be used in other regression problems in a future study. However, the performances of some other data-driven methods with different structures, such as the multivariate adaptive regression spline (MARS), the gene expression programming (GEP) [39], the general regression neural network (GRNN), the genetic programming (GP), and the cascaded neural network (CCNN) have not been studied. More attention will be paid to the performances of different data-driven methods for nonlinear modeling of pump turbine characteristics in the future study.

Author Contributions

C.Z. designed and performed the experiments and drew the figures. T.P. analyzed the results, and wrote the original draft. J.Z. provided the data, reviewed the paper, and gave constructive advice. J.J. checked the whole paper and improved the writing of this paper. X.W. collected relevant material and gave some advice for advisement.

Funding

This research received no external funding.

Acknowledgments

The Natural Science Foundation of Jiangsu Province(No. BK20191052), the Natural Science Foundation of the Jiangsu Higher Education Institutions of China (No. 19KJB470012), the Natural Science Foundation of the Jiangsu Higher Education Institutions of China (No. 19KJB480007), the National Natural Science Foundation of China (No. 51741907), the National Natural Science Foundation of China (NSFC) (No. 51709121), and the National Natural Science Foundation of China (No. 51709122) supported this work.

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

Pump turbine modelImproved Suter transformation
M11unit momentarelative speed
Q11unit flowqrelative flow
N11unit speedhrelative water head
aguide vane openingmrelative moment
xrelative flow angle
ELMyrelative opening
( x k , y k ) Training samplesAE-PLS-ELM
y ^ k Simulated output u i , v i Score vectors
NNumber of training samples ω i , c i Load factors
LNumber of hidden neurons λ 1 , λ 2 Lagrange multipliers
a i , β i Connection weights E r , F r Residual matrix
b i Bias α i , γ i Regression coefficients
g ( ) Activation function θ i Objective function
H Hidden layer output matrixWComponent matrix
H Generalized inverse matrix of H BDiagonal matrix
Abbreviations
AEAutoencoderGSGrid search
AE-ELMAutoencoder based extreme learning machineMAEMean absolute error
MAPEMean absolute percent error
AE-PLS-ELMAutoencoder and partial least squares regression based extreme learning machineNSENash-Sutcliffe Efficiency
PLSRPartial least squares regression
PSUPumped storage unit
BPBackpropagationPTGSPump turbine governing system
RBFRadial basis functionRMSERoot mean square error
ELMExtreme learning machineSVRSupport vector regression

References

  1. Li, C.; Zou, W.; Zhang, N.; Lai, X. An evolving T–S fuzzy model identification approach based on a special membership function and its application on pump-turbine governing system. Eng. Appl. Artif. Intell. 2018, 69, 93–103. [Google Scholar] [CrossRef]
  2. Zhou, J.; Zhang, C.; Peng, T.; Xu, Y. Parameter Identification of Pump Turbine Governing System Using an Improved Backtracking Search Algorithm. Energies 2018, 11, 1668. [Google Scholar] [CrossRef]
  3. Fu, W.; Wang, K.; Li, C.; Li, X.; Li, Y.; Zhong, H. Vibration trend measurement for a hydropower generator based on optimal variational mode decomposition and an LSSVM improved with chaotic sine cosine algorithm optimization. Measurement Science and Technology 2018, 30, 015012. [Google Scholar] [CrossRef]
  4. Xu, Y.; Zheng, Y.; Du, Y.; Yang, W.; Peng, X.; Li, C. Adaptive condition predictive-fuzzy PID optimal control of start-up process for pumped storage unit at low head area. Energy Convers. Manag. 2018, 177, 592–604. [Google Scholar] [CrossRef]
  5. Lai, X.; Li, C.; Zhou, J.; Zhang, N. Multi-objective optimization for guide vane shutting based on MOASA. Renew. Energy 2019, 139, 302–312. [Google Scholar] [CrossRef]
  6. Wang, L.; Han, Q.; Chen, D.; Wu, C.; Wang, X. Non-linear modelling and stability analysis of the PTGS at pump mode. IET Renew. Power Gener. 2017, 11, 827–836. [Google Scholar] [CrossRef]
  7. Zhang, H.; Chen, D.; Xu, B.; Patelli, E.; Tolo, S. Dynamic analysis of a pumped-storage hydropower plant with random power load. Mech. Syst. Signal Process. 2018, 100, 524–533. [Google Scholar] [CrossRef] [Green Version]
  8. Zhang, C.; Peng, T.; Li, C.; Fu, W.; Xia, X.; Xue, X. Multiobjective Optimization of a Fractional-Order PID Controller for Pumped Turbine Governing System Using an Improved NSGA-III Algorithm under Multiworking Conditions. Complexity 2019, 2019. [Google Scholar] [CrossRef]
  9. Zhang, C.; Li, C.; Peng, T.; Xia, X.; Xue, X.; Fu, W.; Zhou, J. Modeling and Synchronous Optimization of Pump Turbine Governing System Using Sparse Robust Least Squares Support Vector Machine and Hybrid Backtracking Search Algorithm. Energies 2018, 11, 3108. [Google Scholar] [CrossRef]
  10. Palikhe, S.; Zhou, J.; Bhattarai, K.P. Hydraulic Oscillation and Instability of a Hydraulic System with Two Different Pump-Turbines in Turbine Operation. Water 2019, 11, 692. [Google Scholar] [CrossRef]
  11. Zuo, Z.; Fan, H.; Liu, S.; Wu, Y. S-shaped characteristics on the performance curves of pump-turbines in turbine mode–A review. Renew. Sustain. Energy Rev. 2016, 60, 836–851. [Google Scholar] [CrossRef]
  12. Lima, G.M.; Júnior, E.L. Method to Estimate Complete Curves of Hydraulic Pumps through the Polymorphism of Existing Curves. J. Hydraul. Eng. 2017, 143, 04017017. [Google Scholar] [CrossRef]
  13. Huang, W.; Yang, K.; Guo, X.; Ma, J.; Wang, J.; Li, J. Prediction Method for the Complete Characteristic Curves of a Francis Pump-Turbine. Water 2018, 10, 205. [Google Scholar] [CrossRef]
  14. Li, J.; Han, C.; Yu, F. A New Processing Method Combined with BP Neural Network for Francis Turbine Synthetic Characteristic Curve Research. Int. J. Rotating Mach. 2017, 2017. [Google Scholar] [CrossRef]
  15. Qian, J.; Zeng, Y.; Guo, Y.; Zhang, L. Reconstruction of the complete characteristics of the hydro turbine based on inner energy loss. Nonlinear Dyn. 2016, 86, 963–974. [Google Scholar] [CrossRef] [Green Version]
  16. Tanaka, H.; Tsunoda, S. The development of high head single stage pump-turbines. In Proceedings of the 10th IAHR Symposium, Tokyo, Japan, 1–4 January 1980; pp. 429–440. [Google Scholar]
  17. Yang, J.; Pavesi, G.; Yuan, S.; Cavazzini, G.; Ardizzon, G. Experimental Characterization of a Pump–Turbine in Pump Mode at Hump Instability Region. J. Fluids Eng. 2015, 137, 051109. [Google Scholar] [CrossRef]
  18. Boldy, A.; Walmsley, N. Representation of the characteristics of reversible pump turbines for use in waterhammer simulations. In Proceedings of the 4th International Conference on Pressure Surges, Bath, UK, 21–23 September 1983; pp. 287–296. [Google Scholar]
  19. Suter, P. Representation of Pump Characteristics for Calculation of Water Hammer. Sulzer Tech. Rev. 1966, 4, 45–48. [Google Scholar]
  20. Liu, Z.M.; Zhang, D.H.; Liu, Y.Y.; Zhang, X. New Suter-transformation Method of Complete Characteristic Curves of Pump-turbines Based on the 3-D Surface. China Rural Water Hydropower 2015, 1, 143–145. [Google Scholar]
  21. Zhiyong, Z.; Bing, S.; Xu, L.; Xuerui, G. Nonlinear simulation of Kaplan turbine regulating system based on RBF networks. In Proceedings of the 2011 International Conference on Electric Information and Control Engineering, Wuhan, China, 15–17 April 2011; pp. 4302–4306. [Google Scholar]
  22. Peng, T.; Zhou, J.; Zhang, C.; Zheng, Y. Multi-step ahead wind speed forecasting using a hybrid model based on two-stage decomposition technique and AdaBoost-extreme learning machine. Energy Convers. Manag. 2017, 153, 589–602. [Google Scholar] [CrossRef]
  23. Zhang, C.; Zhou, J.; Li, C.; Fu, W.; Peng, T. A compound structure of ELM based on feature selection and parameter optimization using hybrid backtracking search algorithm for wind speed forecasting. Energy Convers. Manag. 2017, 143, 360–376. [Google Scholar] [CrossRef]
  24. Fu, W.; Wang, K.; Zhou, J.; Xu, Y.; Tan, J.; Chen, T. A Hybrid Approach for Multi-Step Wind Speed Forecasting Based on Multi-Scale Dominant Ingredient Chaotic Analysis, KELM and Synchronous Optimization Strategy. Sustainability 2019, 11, 1804. [Google Scholar] [CrossRef]
  25. Chang, C.-W.; Lee, H.-W.; Liu, C.-H. A Review of Artificial Intelligence Algorithms Used for Smart Machine Tools. Inventions 2018, 3, 41. [Google Scholar] [CrossRef]
  26. Han, M.; Zhang, R.; Xu, M. Multivariate Chaotic Time Series Prediction Based on ELM–PLSR and Hybrid Variable Selection Algorithm. Neural Process. Lett. 2017, 46, 1–13. [Google Scholar] [CrossRef]
  27. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the IEEE International Joint Conference on Neural Networks, Budapest, Hungary, 25–29 July 2004; pp. 985–990. [Google Scholar]
  28. Huang, G.B.; Zhou, H.; Ding, X.; Zhang, R. Extreme learning machine for regression and multiclass classification. IEEE Trans. Syst. Man Cybern. Part B 2012, 42, 513–529. [Google Scholar] [CrossRef] [PubMed]
  29. Li, K.; Xiong, M.; Li, F.; Su, L.; Wu, J. A novel fault diagnosis algorithm for rotating machinery based on a sparsity and neighborhood preserving deep extreme learning machine. Neurocomputing 2019, 350, 261–270. [Google Scholar] [CrossRef]
  30. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning internal representation by back-propagation of errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  31. Lekamalage, L.; Kasun, C.; Zhou, H.; Huang, G.-B.; Cambria, E. Representational learning with ELMs for big data. IEEE Intell. Syst. 2013, 11, 31–34. [Google Scholar]
  32. Sun, K.; Zhang, J.; Zhang, C.; Hu, J. Generalized extreme learning machine autoencoder and a new deep neural network. Neurocomputing 2017, 230, 374–381. [Google Scholar] [CrossRef]
  33. Wold, S.; Sjöström, M.; Eriksson, L. PLS-regression: A basic tool of chemometrics. Chemom. Intell. Lab. Syst. 2001, 58, 109–130. [Google Scholar] [CrossRef]
  34. Duzan, H.; Shariff, N.S.B.M. Ridge regression for solving the multicollinearity problem: Review of methods and models. J. Appl. Sci. 2015, 15, 392. [Google Scholar] [CrossRef]
  35. Guo, Y.; Zeng, Y.; Fu, L.; Chen, X. Modeling and Experimental Study for Online Measurement of Hydraulic Cylinder Micro Leakage Based on Convolutional Neural Network. Sensors 2019, 19, 2159. [Google Scholar] [CrossRef] [PubMed]
  36. Fu, W.; Wang, K.; Li, C.; Tan, J. Multi-step short-term wind speed forecasting approach based on multi-scale dominant ingredient chaotic analysis, improved hybrid GWO-SCA optimization and ELM. Energy Convers. Manag. 2019, 187, 356–377. [Google Scholar] [CrossRef]
  37. Nash, J.E.; Sutcliffe, J.V. River flow forecasting through conceptual models part I—A discussion of principles. J. Hydrol. 1970, 10, 282–290. [Google Scholar] [CrossRef]
  38. Vuković, N.; Miljković, Z. Robust sequential learning of feedforward neural networks in the presence of heavy-tailed noise. Neural Netw. 2015, 63, 31–47. [Google Scholar] [CrossRef] [PubMed]
  39. Mirabbasi, R.; Kisi, O.; Sanikhani, H.; Meshram, S.G. Monthly long-term rainfall estimation in Central India using M5Tree, MARS, LSSVR, ANN and GEP models. Neural Comput. Appl. 2018, 1–20. [Google Scholar] [CrossRef]
Figure 1. The characteristic curves of a pump turbine under different openings. (a) Flow characteristic curve and (b) moment characteristic curve.
Figure 1. The characteristic curves of a pump turbine under different openings. (a) Flow characteristic curve and (b) moment characteristic curve.
Applsci 09 03987 g001
Figure 2. WH and WM characteristic curves based on improved Suter transformation under different relative speed.
Figure 2. WH and WM characteristic curves based on improved Suter transformation under different relative speed.
Applsci 09 03987 g002
Figure 3. Structure of the proposed Autoencoder and partial least squares regression based extreme learning machine (AE-PLS-ELM) model. (a) ELM-AE, (b) PLSR, (c) AE-PLS-ELM.
Figure 3. Structure of the proposed Autoencoder and partial least squares regression based extreme learning machine (AE-PLS-ELM) model. (a) ELM-AE, (b) PLSR, (c) AE-PLS-ELM.
Applsci 09 03987 g003
Figure 4. 3D surface for WH characteristics based on the proposed AE-PLS-ELM model.
Figure 4. 3D surface for WH characteristics based on the proposed AE-PLS-ELM model.
Applsci 09 03987 g004
Figure 5. 3D surface for WM characteristics based on the proposed AE-PLS-ELM model.
Figure 5. 3D surface for WM characteristics based on the proposed AE-PLS-ELM model.
Applsci 09 03987 g005
Figure 6. Comparison of the test residuals for WH characteristics between Bagtree, SVR, BP, ELM, and AE-PLS-ELM. (a) Bagtree vs. AE-PLS-ELM, (b) SVR vs. AE-PLS-ELM, (c) BP vs. AE-PLS-ELM, (d) ELM vs. AE-PLS-ELM.
Figure 6. Comparison of the test residuals for WH characteristics between Bagtree, SVR, BP, ELM, and AE-PLS-ELM. (a) Bagtree vs. AE-PLS-ELM, (b) SVR vs. AE-PLS-ELM, (c) BP vs. AE-PLS-ELM, (d) ELM vs. AE-PLS-ELM.
Applsci 09 03987 g006
Figure 7. Comparison of the test residuals for WM characteristics between Bagtree, SVR, BP, ELM and AE-PLS-ELM. (a) Bagtree vs. AE-PLS-ELM, (b) SVR vs. AE-PLS-ELM, (c) BP vs. AE-PLS-ELM, (d) ELM vs. AE-PLS-ELM.
Figure 7. Comparison of the test residuals for WM characteristics between Bagtree, SVR, BP, ELM and AE-PLS-ELM. (a) Bagtree vs. AE-PLS-ELM, (b) SVR vs. AE-PLS-ELM, (c) BP vs. AE-PLS-ELM, (d) ELM vs. AE-PLS-ELM.
Applsci 09 03987 g007
Figure 8. Comparison of the Scatter diagram for the test residuals between ELM and AE-PLS-ELM.
Figure 8. Comparison of the Scatter diagram for the test residuals between ELM and AE-PLS-ELM.
Applsci 09 03987 g008
Figure 9. Comparison of the test results of the five different models for the nonlinear differential equation.
Figure 9. Comparison of the test results of the five different models for the nonlinear differential equation.
Applsci 09 03987 g009
Table 1. Statistics error indices for WH characteristics of different models.
Table 1. Statistics error indices for WH characteristics of different models.
CharacteristicsModelTrainingTest
RMSEMAEMAPENSERMSEMAEMAPENSE
WHBagtree0.018270.012140.134020.992320.020340.014040.164050.98938
SVR0.010190.007110.050710.996460.010690.007490.059630.99464
BP0.003170.002220.024380.999770.003240.002360.025240.99973
ELM0.002900.002070.021980.999810.002970.002340.024670.99977
AE-PLS-ELM0.002290.001300.005990.999880.002170.001420.006970.99988
Table 2. Statistics error indices for WM characteristics of different models.
Table 2. Statistics error indices for WM characteristics of different models.
CharacteristicsModelTrainingTest
RMSEMAEMAPENSERMSEMAEMAPENSE
WMBagtree0.182220.124220.143290.99315 0.211250.150820.187320.99001
SVR0.107300.074760.054110.99763 0.111930.078520.063720.99719
BP0.032270.022320.026800.99979 0.035080.026010.030500.99972
ELM0.026560.017130.015820.99985 0.029500.020810.021100.99981
AE-PLS-ELM0.023150.012690.005390.99989 0.020270.013060.006150.99991
Table 3. Statistics error indices for the test problem of different models.
Table 3. Statistics error indices for the test problem of different models.
ModelTrainingTest
RMSEMAEMAPENSERMSEMAEMAPENSE
Bagtree0.06928 0.00911 0.02077 0.98817 0.03897 0.02709 0.20570 0.99174
SVR0.00002 0.00001 0.00004 1.00000 0.05423 0.04805 0.41004 0.98400
BP0.00010 0.00007 0.00027 1.00000 0.02675 0.02348 0.20595 0.99611
ELM0.00004 0.00003 0.00007 1.00000 0.02506 0.02220 0.19029 0.99658
AE-PLS-ELM0.00010 0.00007 0.00021 1.00000 0.02038 0.01793 0.14956 0.99774

Share and Cite

MDPI and ACS Style

Zhang, C.; Peng, T.; Zhou, J.; Ji, J.; Wang, X. An Improved Autoencoder and Partial Least Squares Regression-Based Extreme Learning Machine Model for Pump Turbine Characteristics. Appl. Sci. 2019, 9, 3987. https://doi.org/10.3390/app9193987

AMA Style

Zhang C, Peng T, Zhou J, Ji J, Wang X. An Improved Autoencoder and Partial Least Squares Regression-Based Extreme Learning Machine Model for Pump Turbine Characteristics. Applied Sciences. 2019; 9(19):3987. https://doi.org/10.3390/app9193987

Chicago/Turabian Style

Zhang, Chu, Tian Peng, Jianzhong Zhou, Jie Ji, and Xiaolu Wang. 2019. "An Improved Autoencoder and Partial Least Squares Regression-Based Extreme Learning Machine Model for Pump Turbine Characteristics" Applied Sciences 9, no. 19: 3987. https://doi.org/10.3390/app9193987

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop