Next Article in Journal
Traceability of pH to the Mole
Next Article in Special Issue
Heuristic Methods for Reservoir Monthly Inflow Forecasting: A Case Study of Xinfengjiang Reservoir in Pearl River, China
Previous Article in Journal
Transferable Principles for Managing the Nexus: Lessons from Historical Global Water Modelling of Central Asia
Previous Article in Special Issue
Daily Runoff Forecasting Model Based on ANN and Data Preprocessing Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Daily Reservoir Runoff Forecasting Method Using Artificial Neural Network Based on Quantum-behaved Particle Swarm Optimization

1
Institute of Hydropower and Hydroinformatics, Dalian University of Technology, Dalian 116024, China
2
Department of Civil and Environmental Engineering, Hong Kong Polytechnic University, Hong Kong 999077, China
*
Author to whom correspondence should be addressed.
Water 2015, 7(8), 4232-4246; https://doi.org/10.3390/w7084232
Submission received: 30 June 2015 / Revised: 21 July 2015 / Accepted: 27 July 2015 / Published: 31 July 2015
(This article belongs to the Special Issue Use of Meta-Heuristic Techniques in Rainfall-Runoff Modelling)

Abstract

:
Accurate daily runoff forecasting is of great significance for the operation control of hydropower station and power grid. Conventional methods including rainfall-runoff models and statistical techniques usually rely on a number of assumptions, leading to some deviation from the exact results. Artificial neural network (ANN) has the advantages of high fault-tolerance, strong nonlinear mapping and learning ability, which provides an effective method for the daily runoff forecasting. However, its training has certain drawbacks such as time-consuming, slow learning speed and easily falling into local optimum, which cannot be ignored in the real world application. In order to overcome the disadvantages of ANN model, the artificial neural network model based on quantum-behaved particle swarm optimization (QPSO), ANN-QPSO for short, is presented for the daily runoff forecasting in this paper, where QPSO was employed to select the synaptic weights and thresholds of ANN, while ANN was used for the prediction. The proposed model can combine the advantages of both QPSO and ANN to enhance the generalization performance of the forecasting model. The methodology is assessed by using the daily runoff data of Hongjiadu reservoir in southeast Guizhou province of China from 2006 to 2014. The results demonstrate that the proposed approach achieves much better forecast accuracy than the basic ANN model, and the QPSO algorithm is an alternative training technique for the ANN parameters selection.

1. Introduction

Accurate daily runoff forecasting is extremely important for hydropower operation control and power grid operation scheduling [1,2,3,4,5]. Over the past decades, there have been abundant traditional classical research works for the daily runoff forecasting, which can be broadly separated into two categories: process-based model like Xin’anjiang model [6,7], and data-based model such as autoregressive model and moving average model [8,9]. These approaches entail exogenous input and rely on a number of assumptions for natural environment, leading to some deviation from the exact results in most cases. In recent years, with the booming development of heuristic methods, many researchers pay attention to applying them in daily runoff forecasting or the parameter selection of the hydrologic model, including artificial neural network [10,11], SCE-UA algorithm [12,13], support vector machine [14,15] and other hybrid methods [16,17].
As a typical artificial neural network, back propagation neural network (BP) can nearly simulate any complex linear or non-linear functional relationship without knowing the correlation between the input data and the output data [18,19]. After learning from the training data set, BP can be used to predict a new output data effectively with the corresponding input data. Compared with other methods, BP has high fault tolerance, strong robustness and easy adaptability to online learning. BP has been widely used in many practical areas, including load prediction, wind speed prediction and daily runoff forecasting. However, BP has some drawbacks, such as long computing time, slow convergence and easy to encounter local minimum. Hence, a variety of hybrid optimization methods using such global optimization algorithm like particle swarm optimization (PSO) are developed to improve the generalization ability of the artificial neural network [20,21]. These hybrid optimization methods can improve the BP forecasting performance in varying degrees. However, when PSO is applied for the ANN parameters selection, it may be trapped into the local optima of the objective function because PSO is restricted by search capability. Therefore, the promotion space is still large for ANN parameter selection using an evolutionary algorithm [22,23,24].
In recent years, a novel particle swarm optimization variant called quantum-behaved particle swarm optimization algorithm (QPSO) was proposed by Sun et al. [25] in 2004. In QPSO, the global optimal solution in the whole searching space can be guaranteed theoretically. Moreover, simulation results of numerous complex benchmark functions showed that QPSO has better global searching ability than the basic PSO [26,27]. Hence, QPSO are widely used to solve the complex optimization problems which includes hydrothermal scheduling and economic dispatch problem [28,29]. However, up to now, there are a few reports about using QPSO for parameter calibration of artificial neural network. Therefore, to improve the generalization ability and calculation efficiency of artificial neural network, a hybrid method, coupling artificial neural network and QPSO, is developed for daily reservoir runoff forecasting in this research, where QPSO algorithm is selected as training algorithm for artificial neural network to enhance hydrologic forecast accuracy.
The remaining is organized as follows. Artificial neural network is introduced briefly in Section 2. Section 3 explains the PSO and QPSO methods in details, then the proposed QPSO-ANN model is presented in Section 4. Section 5 provides the results of the proposed method and other methods and gives discussions on their performances. Finally, conclusions are summarized in Section 6.

2. Artificial Neural Network

In general, the relation between input vector and output vector of a nonlinear system can be expressed as Y = H(X), where X = [X1,…,Xi,…,Xn]T is input vector; Xi is the ith input data; n is the number of input data; Y = [Y1,…,Yj,…,Yu]T is output vector; Yj is the jth input data; u is the number of output data; H(·) denotes the complex nonlinear relation which can be estimated by some meta-heuristic methods like artificial neural network.
In our study, BP neural network is chosen as the basic approach to estimate input-output relations H(·) of the nonlinear hydrological system. The main advantage of BP is that it can reflect the complex nature of underlying process with less information than other traditional methods. A typical BP neural network consists of three layers, input layer, hidden layer and output layer, whose structure is shown in Figure 1. Each layer is composed of a series of interconnected processing nodes. As a specific neuron, each node in any layer uses a nonlinear transfer function to calculate the inner product of input vector and weight vector to get a scalar result. Two neighboring layers are connected via the weights of the nodes between these layers. The input layer receives and transmits input data to hidden layer. The hidden layer may contain a single layer or multi-layer that receives values from the previous layer. Each hidden layer is responsible for the input information conversion and then delivers them to the next hidden layer or output layer. The output layer presents the simulated results and has only one single layer with one or several nodes. In a single calculation, the BP neural network can obtain overall error between the estimated output values and the target output values, then loss function gradient is calculated. The gradient-descent algorithm is fed to update weights and thresholds to minimize loss function. The connection weights and thresholds between any two feed forward-connected neurons will be unceasingly adjusted until the error meets the termination conditions. Then, the optimized BP neural network can be used to forecast the target value with the corresponding input vector.
Figure 1. Schematic diagram of three-layer back propagation (BP) neural network.
Figure 1. Schematic diagram of three-layer back propagation (BP) neural network.
Water 07 04232 g001
Due to the defects of gradient descent method, the result may find a locally optimal solution instead of global optimum in most cases due to the existence of many local optima on the error surface. The convergence in back propagation learning cannot be guaranteed. Moreover, the computation speed is rather slow because the gradient descent algorithm requires small learning rates for stable learning.

3. Quantum-Behaved Particle Swarm Optimization

3.1. Particle Swarm Optimization

Particle Swarm Optimization (PSO), proposed by Kennedy and Eberhart in 1995 [30], is a well-known evolutionary population-based algorithm for global optimization problems [31,32]. The main concept of PSO is from analogy of biological and sociological feeding behavior of bird swarm, and its basic variant is a population (called a swarm) of candidate solutions (called particles). PSO is initialized with random solutions which search for optima by flying through problem space from generation to generation. The flight of each particle is guided continuously by its own best known position, pbest, and the best known position of the whole population, gbest. Each particle has position vector and velocity vector, and explores in the searching space by a few simple formulas.
There is extensive and profound homology or resemblance between PSO and other evolutionary computation techniques like genetic algorithm. Compared with genetic algorithm, PSO has faster convergence speed since it has no evolution operators like crossover and selection. Moreover, PSO has few parameters to adjust and depends directly upon function values rather than derivative information. In the past few years, PSO has been successfully applied to many research and application areas. However, the main defect of PSO is that global convergence cannot be guaranteed, especially when the number of decision variables or dimensions to be optimized are large. In other words, PSO is easily trapped into local optimum although it may have fast convergent rate [32,33].

3.2. Quantum-Behaved Particle Swarm Optimization

In order to deal with disadvantages of PSO, quantum-behaved particle swarm optimization (QPSO) was developed by Sun et al. [25] in 2004. From the quantum mechanics perspective, QPSO considers the particle possess quantum behavior and cannot determine the exact values of position vector and velocity vector simultaneously according to uncertainty principle [25,26]. Hence, there is no velocity vector in the particle of QPSO, and particle state is associated with an appropriate time-dependent Schrödinger equation and can be characterized by wave function ψ instead of position and velocity [27,28], where |ψ|2 is the probability density function of its position. Let M particles in d dimensional space with k ¯ maximum generations, the position vector of ith particle at kth generation can be expressed as xi(k) = [xi,1(k),xi,2(k),,xi,d(k)]T. There will be gbest(k) = [gbest1(k),gbest2(k),,gbestd(k)]T and pbesti(k) = [pbesti,1(k),pbesti,2(k),,pbesti,d(k)]T. Employing Monte-Carlo method, particle moves according to the following iterative equation:
x i , j ( k + 1 ) = { p i , j ( k + 1 ) + a ( k ) × | m b e s t j ( k + 1 ) x i , j ( k ) | × ln ( 1 / r 1 ) , i f r 2 0.5 p i , j ( k + 1 ) a ( k ) × | m b e s t j ( k + 1 ) x i , j ( k ) | × ln ( 1 / r 1 ) , i f r 2 < 0.5
p i , j ( k + 1 ) = r 3 × p b e s t i , j ( k ) + ( 1 r 3 ) × g b e s t j ( k )
m b e s t j ( k + 1 ) = 1 M i = 1 M p b e s t i , j ( k )
a ( k ) = ( a 1 a 2 ) × ( k ¯ k ) k ¯ + a 2
for i = 1,2,...,M; j = 1,2,...,d; k = 1,2,..., k ¯ . Where xi,j(k) is position for jth dimension of ith particle in kth generation; r1, r2, r3 are random variables distributed uniformly in [0,1]; a(k) is contraction-expansion coefficient in kth generation which controls the convergence speed of the particle; a1, a2 are maximum and minimum value of a(k), respectively; and there are usually a1 = 1.0, a2 = 0.5 [28]; pi,j(k) is the jth dimension of local attractor i in kth generation; mbest represents the mean best position defined as mean of all pbest position of the whole population.

4. Parameters Selection for Artificial Neural Network Based on QPSO Algorithm

In order to obtain better forecast accuracy, the novel QPSO algorithm is employed for parameters selection of BP neural network. In this paper, there is only one node in output layer, which is the daily runoff forecast value. If the node number of input layer and hidden layer are n and m, respectively, the architecture of the ANN neural network is n-m-1. The flow chart of the proposed method is shown in Figure 2, and the fundamental idea of the proposed method can be described as follows:
  • Step 0: Set basic parameters for the proposed method.
  • Step 0.1: Set maximize iterations k ¯ and population size M in QPSO.
  • Step 0.2: Divide data into training and testing sets.
  • Step 0.3: Define transfer function of neurons, which is a sigmoid function in this paper, i.e.:
    f [ x ] = 1 1 + e x
Step 1: The input and output data in both training and testing sets are normalized to ensure the quality of forecast results.
X = { X i ' } = a × X i X i min X i max X i min + b
where X i ' and Xi is the normalized value and real value of each vector, respectively; X i min and X i max are the minimum and maximum value of input or output arrays; a and b are the positive normalized parameters, respectively. Based on large numbers of numerical experiments, we found that when the variable a = 0.2 and b = 0.6 are adopted to normalize the raw data, the forecasting models performs better. Hence, we use the variable a = 0.2 and b = 0.6 for data normalization in this paper.
Step 2: The QPSO algorithm is employed to select the parameters of BP neural network. The ith particle in the kth generation is denoted by xi(k) = {wi, bi}. Here, wi and bi represent the connection weights and bias matrix between any two layers of the BP neural network, respectively.
Step 3: Set k = 1, and initialize the parameters w and b of every particle randomly in the searching space, which are the connection weight and bias on each node, respectively.
Step 4: Use the parameters to calculate the fitness of each particle. Here, the fitness is the forecasting error between the output values and the target ones.
Step 4.1: Calculate the outputs of all hidden layer nodes for each training sample.
y j = f [ i = 1 n x i w j i + b j ] , j = 1 , 2 , , m
where wji represents the connection weight from the input node i to the hidden node j, bj stands for bias of neuron j, yj is the output value of the hidden layer node j.
Step 4.2: Calculate the output data of the BP neural network for each training sample.
o 1 = f [ j = 1 m y i w 1 j + b 1 ]
where w1j represent the connection weight from hidden node j to the output node 1, b1 stands for the bias of the neuron; o1 stands for the output data of network.
Step 4.3: Step 4.1 and 4.2 are repeated until all the training set samples are calculated. Then the forecasting error F is regarded as the fitness of the particle xi(k).
F [ x i ( k ) ] = 1 s s = 1 S ( o s t s ) 2
where os and ts is the sth normalized output value and target value in the training data, respectively. S is the number of training set samples.
Step 5: Update the best known position of each particle and the best known position of the whole population according to the following two formulas:
p i ( k ) = { x i ( k ) i f ( k = 1 ) arg min { F [ p i ( k 1 ) ] , F [ x i ( k ) ] } o t h e r w i s e
g b e s t ( k ) = { arg min { min 1 i M { F [ p i ( k ) ] } , F [ g b e s t ( k 1 ) ] } i f ( k > 1 ) arg min { min 1 i M { F [ p i ( k ) ] } } o t h e r w i s e
Step 6: Calculate the mean best position and the contraction-expansion coefficient with Equations (3) and (4), respectively.
Step 7: Update current position of each particle by the Equation (1).
Step 8: Set k = k+1, if the maximum iterations k ¯ reached, the flow will go to Step 9, else go back to Step 4.
Step 9: Output the optimal parameter of the BP neural network, which will be used for new data forecasting process.
Step 10: Before starting the forecasting process, the input vector are needed to be normalized by Equation (6), then transmit the processed data into the calibrated artificial neural network model to obtain predictive value. The predictive data need to be renormalized to the original range of output data by Equation (12).
Y ^ = { Y ^ i } = ( Y i ' b ) × ( Y i max Y i min ) a + Y i min
where Y i ' and Y ^ i are normalized forecasting value and real forecasting value of the output vector, respectively; Y i min and Y i max are minimum and maximum value of the output arrays, respectively.
Figure 2. Flow chart of the proposed method.
Figure 2. Flow chart of the proposed method.
Water 07 04232 g002

5. Simulations

5.1. Study Area and Data Used

The study site is Hongjiadu reservoir in Wu River of Guizhou Province in southwest China. The Wu River is the biggest branch at southern bank of Yangtze River. Hongjiadu reservoir is the leading one with multi-year regulation ability of eleven cascade hydropower stations on Wu River. Rainfall produces most of the runoff. Its drainage area is 9900 km2 and the mean annual runoff is 155 m3·s−1 at the dam site. The total reservoir storage is 4.95 billion cubic meters and the regulated storage is 3.36 billion cubic meters. Locations of Wu River and Hongjiadu reservoir are shown in Figure 3.
Figure 3. Location of study area.
Figure 3. Location of study area.
Water 07 04232 g003
All the daily runoff data collected from Hongjiadu reservoir operators is from 1 January 2006 to 31 December 2013. The data set between 1 January 2006 and 31 December 2012 is used for parameters calibration, while that from 1 January 2013 to 31 December 2013 is for testing. Moreover, before applying the proposed method for daily runoff predication, the data should be normalized to avoid numerical difficulties during calculation and guarantee consistency of greater numeric ranges as well as smaller ones. In the modeling process, input and output data sets are linearly scaled to [0.2, 0.8] as shown in Equation (6).

5.2. Performance Assessment Measures

Many performance measure methods have been developed to assess the forecast accuracy. However, so far, there is no unified standard since each measure can reflect one or more characteristics of the forecasting method. Hence, four commonly used metrics are selected to evaluate the forecast results: coefficient of correlation (R), Nash-Sutcliffe efficiency coefficient (NSE), root mean squared error (RMSE) and mean absolute percentage error (MAPE). R evaluates the linear relation between two data sequences, while NSE shows the capability of the model in predicting values away from the mean. The larger the value of R and NSE, the better the performances of the forecasting model. RMSE and MAPE measure the residual error and the mean absolute percentage error between the observed and forecasted data, respectively. The smaller the value of RMSE and MAPE, the better the performances of the forecasting model.
R = i = 1 n [ ( Y i Y ¯ ) ( Y ^ i Y ˜ ) ] / i = 1 n ( Y i Y ¯ ) 2 ( Y ^ i Y ˜ ) 2
N S E = 1 i = 1 n ( Y i Y ^ i ) 2 / i = 1 n ( Y i Y ¯ ) 2
R M S E = 1 n i = 1 n ( Y i Y ^ i ) 2
M A P E = 1 n i = 1 n | Y i Y ^ i Y i | × 100
where Yi and Y ^ i are the observed value and predictive value of ith data, respectively. Y ¯ and Y ˜ represent the mean value of the observed value and predictive value, respectively. n is the total number of data set used for performance evaluation and comparison.

5.3. ANN Model Development

For ANN, proper selection of input variables will be helpful to find the best-fitted model. Based on the analysis of cross correlation coefficient and autocorrelation coefficient, four input combinations with different antecedent rainfalls and runoffs were developed for a comparative purpose. Table 1 summarizes these ANN models used for this study, where Runoff(t), Rainfall(t) are the runoff and rainfall value at the t-th period, respectively. To ensure the generalization capability of ANN, we use the trial and error method to determine the optimal network architecture. According to Chau et al. in 2005, the training process needs to be stopped when the error of the testing set starts to increase and that of the training set is still decreasing. Figure 4 shows the performance for the testing set against various numbers of neurons for model 1. The optimal ANN architecture adopted for model 1 is 2-4-1. The other three models have the same procedures as that for model 1. Table 2 shows the architecture and indices of various ANN forecasting models for Hongjiadu reservoir. We can find that antecedent two-day rainfalls and antecedent two-day runoff should be chosen as predictors, and ANN with the architecture of 4-7-1 performs best at this situation.
Table 1. Inputs and relation for various artificial neural network (ANN) forecasting models.
Table 1. Inputs and relation for various artificial neural network (ANN) forecasting models.
ModelInputsRelation between Output Variable and Input Variables
1Runoff(t-1),Rainfall(t-1)Runoff(t)=H[Runoff(t-1),Rainfall(t-1)]
2Runoff(t-1),Rainfall(t-1),Rainfall(t-2)Runoff(t)=H[Runoff(t-1),Rainfall(t-1),Rainfall(t-2)]
3Runoff(t-1),Runoff(t-2),Rainfall(t-1)Runoff(t)=H[Runoff(t-1),Runoff(t-2),Rainfall(t-1)]
4Runoff(t-1),Runoff(t-2),Rainfall(t-1),Rainfall(t-2)Runoff(t)=H[Runoff(t-1),Runoff(t-2),Rainfall(t-1),Rainfall(t-2)]
Figure 4. Performance of model 1 against different numbers of nodes in hidden layer.
Figure 4. Performance of model 1 against different numbers of nodes in hidden layer.
Water 07 04232 g004
Table 2. Architectures and indices of various ANN forecasting models for Hongjiadu reservoir.
Table 2. Architectures and indices of various ANN forecasting models for Hongjiadu reservoir.
ModelModel ArchitectureTrainingTesting
RNSERMSE (m3·s−1)MAPE (%)RNSERMSE (m3·s−1)MAPE (%)
12-4-10.8920.74082.49038.1500.8830.74763.56235.912
23-6-10.8910.73782.98937.9030.9030.75762.32138.417
33-5-10.8930.74282.09036.4930.9030.76161.79237.190
44-7-10.9070.78375.28634.8660.9040.77360.25235.680

5.4. Comparison of Different Methods

In order to verify the effectiveness of the proposed method, the same training and verification samples are used for these two models, ANN and ANN-QPSO, and the above four quantitative indexes are employed to evaluate their performances. According to the above-mentioned analysis, the neural network architectures of both ANN and ANN-QPSO are 4-7-1 for Hongjiadu reservoir. Moreover, for ANN-QPSO, the number of population is set to be 300 whilst the maximize iterations is 500. The two algorithms are implemented by adopting JAVA language.
Table 3 presents the statistics results using various models developed for Hongjiadu study area. It can be seen that these two methods have different performances during both training and testing periods. Compared to the basic ANN, the proposed method is able to produce better forecast results for the daily runoff forecasting in Hongjiadu reservoir. In the training phase, the ANN-QPSO model improved the ANN forecasting ability with about 28.18% and 48.08% reduction in RMSE and MAPE values, respectively. The improvements of the forecasting results regarding the R and NSE were approximately 3.97% and 13.41%, respectively. In the testing phase, when compared with that of ANN, the statistical values of R and NSE of the proposed method increases by 5.42% and 17.46% respectively, while the value of RMSE and MAPE decreases by 36.34% and 28.81%. Figure 5 shows the convergence characteristic for objective functions of two methods in Hongjiadu reservoir. The objective function of the proposed method uses about 4 s to converge to a small neighborhood of the final result, whilst the ANN nearly stops the searching process since 1 second. In addition, in term of total computing time from Table 3, the ANN-QPSO decreases by 66.56% when compared to that of ANN. Thus, it can be concluded that the proposed method needs less computation time and has higher forecasting accuracy degree and global search capability than conventional ANN.
Table 3. Performance indices of two methods for Hongjiadu reservoir.
Table 3. Performance indices of two methods for Hongjiadu reservoir.
MethodTime(s)TrainingTesting
RNSERMSE (m3·s−1)MAPE (%)RNSERMSE (m3·s−1)MAPE (%)
ANN-QPSO10.10.9430.88854.07418.1020.9530.90838.35425.401
ANN30.20.9070.78375.28634.8660.9040.77360.25235.680
Figure 5. Convergence characteristic of the objective function of two methods for Hongjiadu reservoir (a) ANN-QPSO; (b) ANN.
Figure 5. Convergence characteristic of the objective function of two methods for Hongjiadu reservoir (a) ANN-QPSO; (b) ANN.
Water 07 04232 g005
The observed peak flow and forecasted peak flow of two models for Hongjiadu reservoir are shown in Table 4. Both ANN-QPSO and ANN models forecast the maximum peak discharge as 1641.8 m3·s−1 and 1258.3 m3·s−1 instead of the observed 1696.4 m3·s−1, corresponding to about 3.2% and 25.8% underestimation, respectively. Furthermore, the absolute averages of the relative error of the ANN-QPSO and ANN models for forecasting the 8 peak flow are 11.6% and 30.9%, respectively. In summary, the ANN-QPSO method performs better than ANN in term of peak flow estimation.
Table 4. Observed peak flow and forecasted peak flow of two models for Hongjiadu reservoir.
Table 4. Observed peak flow and forecasted peak flow of two models for Hongjiadu reservoir.
PeriodDateObserved
Peak (m3·s−1)
Forecasted Peak (m3·s−1)Relative Error (%)
ANNP-QPSOQPSOANNP-QPSOQPSO
Training2006-06-30854.3792.7747.6−7.2−12.5
Training2007-07-301435.81234.21108.2−14.0−22.8
Training2008-06-221663.81514.5861.5−9.0−48.2
Training2009-08-04628.5456.4407.4−27.4−35.2
Training2010-07-111076.01053.8806.7−2.1−25.0
Training2011-06-23561.9471.6426.6−16.1−24.1
Training2012-07-261696.41641.81258.3−3.2−25.8
Testing2013-06-091343.01151.9622.5−14.2−53.6
Average (absolute)11.630.9
The above analysis indicates that, for daily runoff forecasting, the proposed model can obtain better results than basic ANN model with significant improvements in terms of four different statistical indicators. Figure 6 and Figure 7 demonstrate the scatter plots of observed data versus forecasted data using ANN and ANN-QPSO models during the training and testing period. The performances of both prediction models in the training period and testing period are respectively shown in Figure 8 and Figure 9. From Figure 6, Figure 7, Figure 8 and Figure 9, it can be clearly seen that the forecasting results of ANN model are inferior to that by the proposed ANN-QPSO model. For example, the model proposed in this paper can obtain the approximate maximum flows which is about 1700 m3·s−1 during the training period while an obvious deviation exists between the original observed data and the forecasted data in ANN model. Hence, the ANN-QPSO model can mimic daily runoff better than that by ANN model.
Figure 6. Scatter plots of observed data vs. forecasted data during the training period. (a) ANN-QPSO; (b) ANN.
Figure 6. Scatter plots of observed data vs. forecasted data during the training period. (a) ANN-QPSO; (b) ANN.
Water 07 04232 g006
Figure 7. Scatter plots of observed data vs. forecasted data during the testing period. (a) ANN-QPSO; (b) ANN.
Figure 7. Scatter plots of observed data vs. forecasted data during the testing period. (a) ANN-QPSO; (b) ANN.
Water 07 04232 g007
Figure 8. ANN, ANN-QPSO forecasted data and observed runoff data during the training period.
Figure 8. ANN, ANN-QPSO forecasted data and observed runoff data during the training period.
Water 07 04232 g008
Figure 9. ANN, ANN-quantum-behaved particle swarm optimization (QPSO) forecasted data and observed runoff data during testing period.
Figure 9. ANN, ANN-quantum-behaved particle swarm optimization (QPSO) forecasted data and observed runoff data during testing period.
Water 07 04232 g009

6. Conclusions

Globally, hydropower energy has become one of the promising growing clean and renewable energy sources. It is of great significance to accurately predict daily reservoir runoff for integration of hydropower energy in power system. In this research, a novel method called ANN-QSPO, which is based on artificial neural network (ANN) and quantum-behaved particle swarm optimization (QPSO), was developed for daily reservoir runoff forecasting to help reservoirs plan and manage in a more sustainable manner. In the proposed ANN-QSPO method, QPSO was employed to select the ANN optimal parameters and the ANN was used for the prediction after the training process. The proposed approach was compared with ANN model for daily runoff forecasting of Hongjiadu reservoir in southeast China. From the experiment, the results show that the proposed method achieves much better forecast accuracy than basic ANN model. Compared with the statistical values of R and NSE of ANN, the improvements of the proposed method were approximately 3.97% and 13.41% in the training phase, while respectively increases by 5.42% and 17.46% in the testing phase. In term of total computing time, the ANN-QPSO decreases by 66.56% when compared to that of ANN. Thus, QPSO algorithm can act as an alternative training algorithm for the ANN parameters selection.

Acknowledgments

This study is supported by the Major International Joint Research Project from the National Nature Science Foundation of China (51210014) and the National Basic Research Program of China (973 Program) (No. 2013CB035906).

Author Contributions

All authors contributed extensively to the work presented in this paper. Chun-tian Cheng contributed to the subject of the research and literature review. Wen-jing Niu and Zhong-kai Feng contributed to modeling, data statistical analysis and finalized the manuscripts. Jian-jian Shen contributed to the manuscripts review and commented. Kwok-wing Chau contributed to the manuscripts revision and supervision.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, J.; Cheng, C.T.; Liao, S.L.; Wu, X.Y.; Shen, J.J. Daily reservoir inflow forecasting combining QPF into ANNs model. Hydro. Earth Syst. Sci. 2009, 6, 121–150. [Google Scholar] [CrossRef]
  2. Wang, W.C.; Chau, K.W.; Xu, D.M.; Chen, X.Y. Improving forecasting accuracy of annual runoff time series using ARIMA based on EEMD decomposition. Water Resour. Manag. 2015, 29, 1–21. [Google Scholar] [CrossRef] [Green Version]
  3. Duan, Q.Y.; Sorooshian, S.; Gupta, V. Effective and efficient global optimization for conceptual rainfall-runoff models. Water Resour. Res. 1992, 28, 1015–1031. [Google Scholar] [CrossRef]
  4. Valipour, M.; Banihabib, M.E.; Behbahani, S.M.R. Comparison of the ARMA, ARIMA, and the autoregressive artificial neural network models in forecasting the monthly inflow of Dez dam reservoir. J. Hydrol. 2013, 476, 433–441. [Google Scholar] [CrossRef]
  5. Kneis, D.; Bürger, G.; Bronstert, A. Evaluation of medium-range runoff forecasts for a 500 km2 watershed. J. Hydrol. 2012, 414, 341–353. [Google Scholar] [CrossRef]
  6. Zhao, R.J. The Xinanjiang model applied in China. J. Hydrol. 1992, 135, 371–381. [Google Scholar]
  7. Cheng, C.T.; Ou, C.P.; Chau, K.W. Combining a fuzzy optimal model with a genetic algorithm to solve multi-objective rainfall-runoff model calibration. J. Hydrol. 2002, 268, 72–86. [Google Scholar] [CrossRef]
  8. Wang, W.C.; Chau, K.W.; Cheng, C.T.; Qiu, L. A comparison of performance of several artificial intelligence methods for forecasting monthly discharge time series. J. Hydrol. 2009, 374, 294–306. [Google Scholar] [CrossRef]
  9. Lin, G.F.; Chen, G.R.; Huang, P.Y. Effective typhoon characteristics and their effects on hourly reservoir inflow forecasting. Adv. Water Resour. 2010, 33, 887–898. [Google Scholar] [CrossRef]
  10. Hsu, K.; Gupta, H.V.; Sorroshian, S. Artificial neural network modeling of the rainfall-runoff process. Water Resour. Res. 1995, 31, 2517–2530. [Google Scholar] [CrossRef]
  11. Machado, F.; Mine, M.; Kaviski, E.; Fill, H. Monthly rainfall-runoff modelling using artificial neural networks. Hydrol. Sci. J. 2011, 56, 349–361. [Google Scholar] [CrossRef]
  12. Duan, Q.Y.; Sorooshian, S.; Gupta, V.K. Optimal use of the SCE-UA global optimization method for calibrating watershed models. J. Hydrol. 1994, 158, 265–284. [Google Scholar] [CrossRef]
  13. Duan, Q.Y.; Gupta, V.K.; Sorooshian, S. Shuffled complex evolution approach for effective and efficient global minimization. J. Optim. Theory Appl. 1993, 76, 501–521. [Google Scholar] [CrossRef]
  14. Lin, J.Y.; Cheng, C.T.; Chau, K.W. Using support vector machines for long-term discharge prediction. Hydrol. Sci. J. 2006, 51, 599–612. [Google Scholar] [CrossRef]
  15. Lima, A.R.; Cannon, A.J.; Hsieh, W.W. Nonlinear regression in environmental sciences by support vector machines combined with evolutionary strategy. Comput. Geosci. 2013, 50, 136–144. [Google Scholar] [CrossRef]
  16. Guo, W.J.; Wang, C.H.; Zeng, X.M.; Ma, T.F.; Yang, H. Subgrid parameterization of the soil moisture storage capacity for a distributed rainfall-runoff model. Water 2015, 7, 2691–2706. [Google Scholar] [CrossRef]
  17. Kamruzzaman, M.; Shahriar, M.S.; Beecham, S. Assessment of short term rainfall and stream flows in South Australia. Water 2014, 6, 3528–3544. [Google Scholar] [CrossRef]
  18. Chau, K.W.; Wu, C.L.; Li, Y.S. Comparison of several flood forecasting models in Yangtz River. J. Hydrol. Eng. 2005, 10, 485–491. [Google Scholar] [CrossRef]
  19. Piotrowski, A.P.; Napiorkowski, J.J. Optimizing neural networks for river flow forecasting evolutionary computation methods versus the Levenberg-Marquardt approach. J. Hydrol. 2011, 407, 12–27. [Google Scholar] [CrossRef]
  20. Chau, K.W. Particle swarm optimization training algorithm for ANNs in stage prediction of Shing Mun River. J. Hydrol. 2006, 329, 363–367. [Google Scholar] [CrossRef]
  21. Asadnia, M.; Chua, L.H.C.; Qin, X.S.; Talei, A. Improved particle swarm optimization based artificial neural network for rainfall-runoff modeling. J. Hydrol. Eng. 2014, 19, 1320–1329. [Google Scholar] [CrossRef]
  22. Liu, W.C.; Chung, C.E. Enhancing the predicting accuracy of the water stage using a physical-based model and an artificial neural network-genetic algorithm in a river system. Water 2014, 6, 1642–1661. [Google Scholar] [CrossRef]
  23. Wu, C.L.; Chau, K.W. Rainfall-runoff modeling using artificial neural network coupled with singular spectrum analysis. J. Hydrol. 2011, 399, 394–409. [Google Scholar] [CrossRef]
  24. Jeong, D.I.; Kim, Y.O. Combining single-value streamflow forecasts—A review and guidelines for selecting techniques. J. Hydrol. 2009, 377, 284–299. [Google Scholar] [CrossRef]
  25. Sun, J.; Feng, B.; Xu, W.B. Particle swarm optimization with particles having quantum behavior. In Congress on Evolutionary Computation, 2004 (CEC2004), Proceedings of the 2004 Congress on Evolutionary Computation, Portland, OR, USA, 19–23 June 2004; pp. 325–331.
  26. Fang, W.; Sun, J.; Ding, Y.; Wu, X.; Xu, W. A review of quantum-behaved particle swarm optimization. IETE Tech. Rev. 2010, 27, 336–348. [Google Scholar] [CrossRef]
  27. Xi, M.; Sun, J.; Xu, W. An improved quantum-behaved particle swarm optimization algorithm with weighted mean best position. Appl. Math. Comput. 2008, 205, 751–759. [Google Scholar] [CrossRef]
  28. Feng, Z.K.; Liao, S.L.; Niu, W.J.; Shen, J.J.; Cheng, C.T.; Li, Z.H. Improved quantum-behaved particle swarm optimization and its application in optimal operation of hydropower stations. Adv. Water Sci. 2015, 26, 413–422. (in Chinese). [Google Scholar]
  29. Sun, C.; Lu, S. Short-term combined economic emission hydrothermal scheduling using improved quantum-behaved particle swarm optimization. Expert Syst. Appl. 2010, 37, 4232–4241. [Google Scholar] [CrossRef]
  30. Eberhart, R.; Kennedy, J. A new optimizer using particle swarm theory. In Proceedings of the 6th International Symposium on Micro Machine and Human Science, New York, NY, USA, 4–6 October 1995; pp. 39–43.
  31. Gaing, Z.L. Particle swarm optimization to solving the economic dispatch considering the generator constraints. IEEE Trans. Pow. Syst. 2003, 18, 1187–1195. [Google Scholar] [CrossRef]
  32. Clerc, M.; Kennedy, J. The particle swarm-explosion, stability, and convergence in a multidimensional complex space. IEEE Trans. Evol. Comput. 2002, 6, 58–73. [Google Scholar] [CrossRef]
  33. Ren, C.; An, N.; Wang, J.; Li, L.; Hu, B.; Shang, D. Optimal parameters selection for BP neural network based on particle swarm optimization: A case study of wind speed forecasting. Knowl. Based Syst. 2014, 56, 226–239. [Google Scholar] [CrossRef]

Share and Cite

MDPI and ACS Style

Cheng, C.-t.; Niu, W.-j.; Feng, Z.-k.; Shen, J.-j.; Chau, K.-w. Daily Reservoir Runoff Forecasting Method Using Artificial Neural Network Based on Quantum-behaved Particle Swarm Optimization. Water 2015, 7, 4232-4246. https://doi.org/10.3390/w7084232

AMA Style

Cheng C-t, Niu W-j, Feng Z-k, Shen J-j, Chau K-w. Daily Reservoir Runoff Forecasting Method Using Artificial Neural Network Based on Quantum-behaved Particle Swarm Optimization. Water. 2015; 7(8):4232-4246. https://doi.org/10.3390/w7084232

Chicago/Turabian Style

Cheng, Chun-tian, Wen-jing Niu, Zhong-kai Feng, Jian-jian Shen, and Kwok-wing Chau. 2015. "Daily Reservoir Runoff Forecasting Method Using Artificial Neural Network Based on Quantum-behaved Particle Swarm Optimization" Water 7, no. 8: 4232-4246. https://doi.org/10.3390/w7084232

Article Metrics

Back to TopTop