Next Article in Journal
Multi-Input Nonlinear Programming Based Deterministic Optimization Framework for Evaluating Microgrids with Optimal Renewable-Storage Energy Mix
Next Article in Special Issue
Estimation of Fuzzy Parameters in the Linear Muskingum Model with the Aid of Particle Swarm Optimization
Previous Article in Journal
Design of Pumping Stations Using a Multicriteria Analysis and the Application of the AHP Method
Previous Article in Special Issue
Machine Learning Improvement of Streamflow Simulation by Utilizing Remote Sensing Data and Potential Application in Guiding Reservoir Operation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Novel Ensemble Forecasting of Streamflow Using Locally Weighted Learning Algorithm

1
State Key Laboratory of Hydrology-Water Resources and Hydraulic Engineering, Hohai University, Nanjing 210098, China
2
Forest Research Division, Research Institute of Forests and Rangelands, Agricultural Research, Education and Extension Organization (AREEO), Tehran 1496813111, Iran
3
Department of Civil Engineering, Amrita School of Engineering, Amrita Vishwa Vidyapeetham, Amritanagar, Coimbatore 641 112, India
4
Civil Engineering Department, Ilia State University, 0162 Tbilisi, Georgia
5
Agricultural Engineering Department, Faculty of Agriculture, Mansoura University, Mansoura 35516, Egypt
*
Authors to whom correspondence should be addressed.
Sustainability 2021, 13(11), 5877; https://doi.org/10.3390/su13115877
Submission received: 21 April 2021 / Revised: 20 May 2021 / Accepted: 20 May 2021 / Published: 24 May 2021

Abstract

:
The development of advanced computational models for improving the accuracy of streamflow forecasting could save time and cost for sustainable water resource management. In this study, a locally weighted learning (LWL) algorithm is combined with the Additive Regression (AR), Bagging (BG), Dagging (DG), Random Subspace (RS), and Rotation Forest (RF) ensemble techniques for the streamflow forecasting in the Jhelum Catchment, Pakistan. To build the models, we grouped the initial parameters into four different scenarios (M1–M4) of input data with a five-fold cross-validation (I–V) approach. To evaluate the accuracy of the developed ensemble models, previous lagged values of streamflow were used as inputs whereas the cross-validation technique and periodicity input were used to examine prediction accuracy on the basis of root correlation coefficient (R), root mean squared error (RMSE), mean absolute error (MAE), relative absolute error (RAE), and root relative squared error (RRSE). The results showed that the incorporation of periodicity (i.e., MN) as an additional input variable considerably improved both the training performance and predictive performance of the models. A comparison between the results obtained from the input combinations III and IV revealed a significant performance improvement. The cross-validation revealed that the dataset M3 provided more accurate results compared to the other datasets. While all the ensemble models successfully outperformed the standalone LWL model, the ensemble LWL-AR model was identified as the best model. Our study demonstrated that the ensemble modeling approach is a robust and promising alternative to the single forecasting of streamflow that should be further investigated with different datasets from other regions around the world.

1. Introduction

To understand the current state, potential, and prospects of water availability, systematic studies on all aspects of basin hydrology (e.g., precipitation, surface, and sub-surface water) and investigation of all indicators are required [1,2,3]. Streamflow is one such indicator that has a direct influence on local drinking water supply and the quantity of water available for irrigation, hydro-electricity generation, and other needs [4]. Indeed, projections have shown that 20% of the river discharge is controlled by human interventions [5]. Changes in land use and land cover over time, glaciers, snowfields, topographic boundaries, dams, and reservoir management are some of the key factors influencing streamflow trends [6]. Streamflow data are a very valuable asset if available over a long period of years. Advantages of streamflow forecasting include early flood warning and mitigation, reservoir planning and management, quantification of available water resources for water supply projects, etc. [7]. Accurate forecasting of streamflow is crucial for the efficient management of water reservoir systems, such as dams, under competing demand for water for irrigation, domestic use, and hydro-power generation activities while at the same time maintaining an adequate environment in the river (or stream) system [8]. In addition, both short-term and long-term streamflow forecasting is necessary pertaining to the optimization of the hydrological components of water resource systems mainly during flood or drought periods [9]. Early prediction of streamflow could provide an imminent warning to disaster management organizations to prepare in response to floods quite early thus preventing the costly socio-economic losses incurred from such extreme events [10].
Since streamflow is a derivative of a complex physical system, the predictions of streamflow using physical-based models generally have significant, inherent uncertainty caused by inaccurate or simple representation of hydrological processes, incomplete or incorrect antecedent conditions, bias or errors in the input variables, or uncertainty in the model parameters. In addition, the requirement for big data (n number of parameters) to simulate the hydrological process also restricts the application of physical models [11]. The application of statistics-based time series models such as the autoregressive integrated moving average (ARIMA) model and its derivatives such as periodic or seasonal ARIMA models and more complex multivariate models such as transfer function-noise (TFN) models have been particularly widely used in forecasting monthly streamflow [12]. However, these models are mostly built upon the assumption that the process follows a normal distribution, however the streamflow process is generally non-linear and stochastic in nature [13]. Machine learning (ML) models, which have been widely used in recent decades to model many real-world problems [14,15,16,17,18,19,20], have the unique ability to identify the complex non-linear relationships between the predictors (inputs) and targets (outputs) without the need for the physical characterization of the system or the requirement of making any underlying assumptions. Many hybrid ensemble ML models with the integration of different data preprocessing techniques such as wavelet transformations, empirical mode decomposition, etc. have very high efficiency in accurately forecasting the future streamflow using only antecedent streamflow time series data as input [12,21,22]. Examples of the most recent works on streamflow forecasting can be found in Adnan, Liang, Heddam, Zounemat-Kermani, Kisi and Li [2], Ferreira, et al. [23], Piazzi, et al. [24], Saraiva, et al. [25], and Tyralis, et al. [26].
Although several ML-derived models have been suggested and used to forecast streamflow, there is no model that can forecast streamflow without any biases or with utmost certainty based on the time series of antecedent streamflow values. While literature shows evidence that some single and hybrid ML models, such as OSELM, BGWO-RELM, MLR—KNN, RMGM-BP, RBF-ANN, and MARS-DE, are very effective in forecasting streamflow in river basins across the world, none of these models have been proven to forecast streamflow without any biases or with utmost certainty based on the time series of antecedent streamflow values [2,9,12,24,27]. Hence, the development and application of novel and sophisticated machine learning algorithms for streamflow forecasting are critical to overcoming such limitations in favor of improving the overall forecasting accuracy and model performance. Locally Weighted Learning (LWL) is one such novel machine learning algorithm that has proven efficient for modeling environmental problems. Recently LWL-based ensemble models have been successfully used to model groundwater potential [28] and forest fire susceptibility [29]. One unique advantage of LWL is that for each point of interest a local model is created based on neighboring data of the query point instead of building a whole global model for the entire functional space. Based on this strategy, data points closer to the query point receive a higher weight that can control overprediction. In this study, we combine the LWL algorithm with five ensemble learning techniques, that is, Additive Regression (AR), Bagging (BG), Dagging (DG), Random Subspace (RS), and Rotation Forest (RF), to develop five ensemble models for a novel ensemble forecasting of streamflow. We apply the models to the lagged streamflow time-series input derived from the antecedent streamflow data. To the best of our knowledge, the LWL technique has not yet been investigated for streamflow forecasting and this study is the first to use and compare different versions of the LWL-based ensemble models for this purpose.

2. Case Study

For this study, the Jhelum Catchment located in the western Himalayas in the north part of Pakistan was selected. This catchment originates from India and drains the southern slope of the Greater Himalayas and the northern slope of the Pir Punjal Mountains. The upstream side of the basin located in India is occupied with great glaciers. Due to climate change in recent years, this transboundary river in Pakistan side is greatly affected by glacier melt. Pakistan has a key reservoir (i.e., Mangla Reservoir) downstream of this basin. This reservoir is the second biggest reservoir in Pakistan with an installed capacity of 1000 MW and fulfills 6% of the electricity generation demand of the country. Therefore, precise estimation of this key catchment is very crucial for the economy and sustainability of water resources in Pakistan. This catchment mainly consists of two main sub-basins, that is, the Naran and Neelum basins. The catchment covers a drainage area of 33,342 km2 up to Mangla Dam with an elevation variation of 200 m to 6248 m. For accurate estimation of streamflow in this basin, the key hydraulic station, that is, Kohala station, at the main river Jhelum streamline after the confluence of both key tributaries (Neelum and Naran) was selected as shown in Figure 1. For model development, the monthly streamflow data of the selected station were obtained from the Water and Power Development Authority (WAPDA) of Pakistan for the duration of 1965 to 2012. For a robust data analysis with the models, a cross-validation scheme was applied. Therefore, data were divided into four equal datasets where each dataset was used for model testing whereas the other three datasets were set aside for model training.

3. Methods

3.1. Locally Weighted Learning (LWL) Algorithm

The locally weighted learning (LWL) algorithm is motivated by the classification of example-based approaches [30]. In this algorithm, the regression model is not processed unless the output value of the new vector is presented. This is required to correctly execute all learning at the prediction moment. LWL is an advanced type of M5 method in a way that suits both linear and non-linear regression in space for the unique fields of example [31]. Based on the weighted results, distance according to the questionnaire was used to allocate the weights to the training datasets and a regression equation is produced. There is a wide range of methods of distance-based weighting that can be used on the basis of the problem preference in LWL [32]. The statistical model for basic linear regression and the linear model of the multiple regression are presented, respectively, in Equations (1) and (2):
y i = β 0 + β 1 x i + ε i   i = 1 ,   2 ,   3 . . n
y i = β 0 + β 1 x i 1 + β 2 x i 2 + + β k x i k + ε i
where y is the response (dependent variable), x is the predictor (independent variable), y i and ε i represent random variables, and xi is constant. The linear existence of the model is due to β-parameters. The LWL objective function of squared error is expressed as follows:
Minimize   F = 1 2 N   K = 1 N w k ( K 0 + n = 1 M k n x k n + ε k y k ) 2
where F is the function of objective, w is the weight function matrix, M is total variables number, ε k   is   the random error, and K 0 . . k n are regression coefficients.

3.2. Bagging

Bagging or “Bootstrap Aggregating” is a method composed of two major steps for getting more stable, robust, and precise models [33,34]. Bagging is one of the stable ensemble learning techniques used for resampling the training dataset. The first phase consists of bootstrapping the raw data samples that make up the various sets of training data. From these training datasets, multiple models are created. Prediction is generated from the continuous training processes for datasets and multiple models. The underlying notion of the Bagging technique is straightforward. Instead of generating predictions from a standalone model that is appropriate for the actual data, the relationship between the input-output variables is defined by multiple models generated. Then using the weighted average in the Bagged algorithm, various models are coupled to form a single output [35,36]. This strategy can effectively reduce the possible uncertainties in the modeling process. Previous works prove that Bagging is a favorable choice for ensemble modeling of many environmental problems [29].

3.3. Additive Regression

Additive Regression was first developed by Stone [37] as a nonparametric method to approximate a multivariate function by using multiple unary functions. For the dependent variable Y and the independent variables X1, X2, …, Xp, the nonparametric additive model can be given by:
E ( Y | X 1 .   X 2 . X p =   + i = 1 P f i ( X i )  
where fi (Xi) is a unary nonparametric function. To satisfy the identifiable conditions, it is generally required that fi (Xi) = 0, i = 1, 2, …, p. Compared to traditional linear models, the nonparametric regression model does not pre-suppose the relationship between variables and the form of the regression function. Further, it is an adaptable and robust data-driven model that can yield a better approximation for nonlinear nonhomogeneous problems [38]. Given these advantages, many researchers applied this technique to study the linear and nonlinear relationships in environmental problems [39].

3.4. Random Subspace (RS)

Random Subspace (RS) was developed by Ho [40] as a new ensemble learning technique for resolving real-world problems. The numerous classifiers of this technique are combined and trained on an altered feature space to generate multiple training subsets for the classifiers, which are the training bases. RS applies multiple samples on function space, as opposed to the example space as in other ensemble models, as stated by Havlíček, et al. [41]. This strategy takes advantage of bootstrapping and grouping. The RS inputs are the training set (x), the base-classifier (w), and the subspaces number (L) [42]. It is strongly recommended by Pham, et al. [43] that this approach be used to prevent over-fitting issues and to cope with the most unnecessary datasets.

3.5. Dagging

Ting and Witten [44] pioneered the Dagging algorithm as a resampling ensemble technique that uses most votes to combine various classifiers to get improved prediction accuracy for the base classifier. Dagging generates multiple different samples instead of producing the bootstrap samples to acquire the base classifier. In recent years, it has been considered a promising machine learning algorithm for classification problems. In the real world, the Dagging ensemble technique has been applied to solve different classification problems. The development of an M dataset can occur with a specific training dataset containing N samples which may come from the existing training datasets [45,46]. There are many n (n < N) samples in any dataset that are distinct from each other. In the particular training datasets, the variables are not replaced and can be chosen as a part of the dataset specified where the size of sample datasets is expanded. According to that, a base classifier is installed on any sample dataset. Ultimately, depending on the training dataset, many classifiers can be acquired. The capability of Dagging has been frequently proven for obtaining improved predictive modeling of different classification problems [29,47].

3.6. Rotation Forest

Rotation Forest (RF) is an ensemble learning technique that independently trains L decision trees using, for each tree, a different set of extracted features. Suppose the x = (x1, … xn) T represents an example defined by n characteristics (attributes) and let X be an N × n matrix including examples of the training process. We assume that the actual class labels of all instances of training are also given. Let go of D = {D1, … DL} is the set of classifiers for L and F is the set of characteristics. The purpose of Rotation Forest is to create precise and diverse classifiers. As in Bagging, bootstrap samples are taken as the training collection for the individual classifiers. The key heuristic is to introduce extraction of features and to recreate a complete feature set for each classifier in the ensemble afterward [48]. The feature collection is randomly divided into K subsets to do this. The principal component analysis (PCA) is run on each subset separately, and a new set of n linear extracted features is constructed by pooling all main components. The data are translated into the new space of the function linearly. With this data collection, classifier Di is educated. Multiple splits of the collection of features will contribute to various extracted features, thereby leading to the diversity of the bootstrap sampling implemented.

4. Ensemble Forecasting

Ensemble forecasting of the monthly streamflow was performed using the LWL algorithm that was used as the base model and was combined with the Additive Regression (AR), Bagging (BG), Dagging (DG), Random Subspace (RS), and Rotation Forest (RF) ensemble techniques. This combination resulted in five ensemble models, namely the ensemble LWL-AR, LWL-BG, LWL-DG, LWL-RS, and LWL-RF models. In each model, the ensemble learning technique performs resampling of the training dataset to train the base LWL algorithm. Table 1 details the summary of statistical characteristics of the data used in this study. To build the models, we grouped the initial input parameters into four different scenarios of input data. They include:
(i)
Qt-1
(ii)
Qt-1, Qt-2
(iii)
Qt-1, Qt-2, Qt-3
(iv)
Qt-1, Qt-2, Qt-3, MN
where Qt-1 is the streamflow at 1 previous month and vice versa and MN is the month number of the streamflow.
In a cross-validation approach, data were divided into four equal sets such that three sets were used for model training and the remaining set was used for validation [49,50,51]. We used several performance metrics to measure the performance of the models during both training and validation phases. These metrics include: correlation coefficient (R) (Equation (5), root mean square error (RMSE) (Equation (6), mean absolute error (MAE) (Equation (7), relative absolute error (RAE) (Equation (8), and root-relative square error (RRSE) (Equation (9). A full description of these metrics can be found in the corresponding literature [2,24,52,53,54,55].
R = ( P i P ¯ ) ( T i T ¯ ) ( P i P ¯ ) 2 ( T i T ¯ ) 2  
R M S E = i = 1 n ( P i j T j ) 2 N  
M A E = i = 1 n | P i j T j | N
R A E = | P i j T j T j | × 100  
R R S E = i = 1 n ( P i j T j ) 2 i = 1 n ( T j T ¯ j ) 2  
where P is the value predicted, T is the target value,   P ¯ and T ¯ are the mean predicted and target values.
We developed the models using the open-source Weka software on an HP Laptop with an Intel(R) Core (TM) i3-3110M CPU @ 2.40GHz, 4 GB of RAM, an x64-based processor, and the Microsoft Windows 8.1 operating system. The optimum value for each model parameter was identified via a trial-and-error process. To do so, we arbitrarily entered different values until the best model performance was achieved [36,56]. Table 2 details the optimum parameter setting of each model.

5. Results

Table 3 shows the results of the single LWL model with different input combinations and datasets. Given the mean values of each metric obtained from each input combination and dataset, the model with input combination IV performed the best and achieved RMSE = 244.6 m3/s, MAE = 175 m3/s, RAE = 34.47 m3/s, RRSE = 40.90 m3/s, and R = 0.834 in the training phase and RMSE = 274.8 m3/s, MAE = 199.2 m3/s, RAE = 38.70 m3/s, RRSE = 44.08 m3/s, and R = 0.809 in the testing phase. Importing periodicity (i.e., MN) as an additional input variable into the model considerably improved both the training performance and prediction performance. A comparison between the results obtained from the input combinations III and IV revealed a significant performance improvement, that is, RMSE, MAE, RAE, and RRSE decreased up to 10.12, 14.59, 15.41, and 10.69% in the training phase and 6.17, 9.41, 8.40, and 9.56% in the testing phase, respectively. In terms of the R metric, the results showed 5.3 and 6.1% training and testing improvements when we used input combination IV. Further, the results revealed that the best and worst predictive performance (i.e., testing performance) was obtained with the datasets M3 and M2, respectively.
The results of the five ensemble models, that is, LWL-AR, LWL-BG, LWL-DG, LWL-RS, and LWL-RF, are summed up in Table 4, Table 5, Table 6, Table 7 and Table 8. Similar to the single LWL model, the performance of the ensembles models was predominantly influenced by the input combination and dataset. For example, RMSE of the testing phase ranged from 223.9 m3/s (M3-IV) to 407.8 m3/s (M2-I) for LWL-AR, from 255.3 m3/s (M3-IV) to 345.2 m3/s (M2-III) for LWL-BG, from 233.5 m3/s (M3-IV) to 390.6 m3/s (M2-I) for LWL-DG, from 242.8 m3/s (M3-IV) to 397.2 m3/s (M4-I) for LWL-RS, and from 229.4 m3/s (M3-IV) to 397 m3/s (M2-I) for LWL-RF. Given these values and also the values of other performance metrics, it is evident that the best performance of all models was achieved by the dataset M3 and the input combination IV (i.e., M3-IV).
A comparison between the results obtained from the single LWL model and its ensembles clearly indicates that the ensemble learning techniques considerably improved the training and testing performances of the base LWL algorithm. The ensemble models achieved greater training performance than the single LWL model by about 44.7, 44.7, 47.8, 44.7, and 13.9% in terms of the RMSE, MAE, RAE, RRSE, and R metrics, respectively. In the case of the testing performance, LWL-AR showed 53.3, 54.5, 55, 53.8, and 22.4% improvements. Similarly, testing performance improvements in the corresponding metrics are 8, 7.1, 8, 9, and 4.5% by applying LWL-BG, 12.5, 8.2, 8.3, 13.2, and 5.5% by applying LWL-DG, 3.1, 3.2, 1.9, 4.7, and 4.1% by applying LWL-RS, 12.6, 10, 11, 13.8, and 7.3% by applying LWL-RF, respectively.
A comparison of the models’ outcomes also reveals that the ensemble LWL-AR model performed better than the other models in both training and testing phases of the monthly streamflow modeling. The LWL-DG and LWL-RF models showed similar performance and ranked as the second-best models, followed by the LWL-RS model that was identified as the least effective ensemble model.
To further compare the models’ performance, we used time variation, scatter plots, and Taylor and violin diagrams to visualize the results obtained from the best input combination (i.e., M3-IV). Figure 2 shows that LWL-AR predictions are much closer to the observed values compared to the other models. Figure 3 reveals that the ensemble LWL-AR model performed better compared to other models in catching the extreme streamflow values (minimum and maximum), which is an important indicator in water resource management and for the evaluation of extreme events such as drought and flood.
Figure 4 compares the single LWL model with its ensemble models in low streamflow (i.e., lower than 500 m3/s) prediction and clearly demonstrates the superiority of LWL-AR in catching the minimums of streamflow. Figure 5 shows the scatter plots of the observed and predicted monthly streamflow for the best input combination (i.e., M3-IV). While the single LWL model resulted in a highly scattered prediction with R2 = 0.809, the LWL-AR ensemble model produced a fit line equation (y = 0.9401x + 56.669) close to the exact line (y = x) with the highest R2 value (0.867) compared to the other models.
Figure 6 shows the Taylor diagram of the models and indicates how well the models match each other in terms of their standard deviation and correlation difference. Among the different models, LWL-AR achieved a closer standard deviation to the observed data with the lowest square error and highest correlation, which is followed by the LWL-BG and LWL-DG models. Figure 7 shows the violin graph of the models and indicates that LWL-AR achieved a data distribution similar to the observed data, which is followed by the LWL-DG model.
Overall, our case study demonstrated that the ensemble models successfully outperformed the single LWL model and provided promising accuracy for streamflow forecasting. Due to the non-linear nature of many environmental processes and phenomena (e.g., streamflow), hybrid ensemble models that benefit from the advantages of multiple methods/models can better capture the complexity of these phenomena and often yield more accurate results than single simple models.

6. Discussion

In all ensemble models, considering periodicity (i.e., MN) as an additional input variable substantially improved both the training performance and predictive performance. During the testing phase, for the LWL-AR model, the improvements in RMSE, MAE, RAE, RRSE, and R were up to 13, 17.9, 17.5, 20.5, and 7%, respectively. For the LWL-BG model, the metrics improved up to 9.1, 14.5, 14.5, 8.9, and 5.9%, respectively. For the LWL-DG model, the metrics improved up to 8.6, 10.9, 10, 8.1, and 3.1%, respectively. For the LWL-RS model, the metrics improved up to 10.1, 12.3, 12.4, 10.3, and 9.3%, respectively. For the LWL-RF model, the metrics improved up to 13.8, 15.6, 15.6, 13.9, and 8.2%, respectively. These results are in agreement with the previous studies that reported on the improvement of predictive accuracy using the periodicity variable. For example, Kişi [57] demonstrated the improved performance of the three types of ANN models using the periodicity variable for the prediction of monthly streamflow of the Canakdere and Goksudere rivers, Turkey. Adnan, et al. [58] used the periodicity variable to improve the predictive capability of the FFNN, RBNN, GRNN, and ANFIS models for the prediction of the monthly streamflow of the Gilgit River, Pakistan. In a recent study, Adnan, Zounemat-Kermani, Kuriqi and Kisi [53] achieved an improved performance of the long short-term memory (LSTM), extreme learning machines (ELM), and random forest (RF) models for the monthly streamflow of the Kohala and Garhi Habibullah stations in Pakistan. They showed that the inclusion of the periodicity component (MN) decreased the RMSE of the optimal LSTM, ELM, and RF models by 11.9%, 6.9%, and 1% for the Garhihabibullah Station and by 20.8%, 20.5%, and 3.7% for the Kohala Station, respectively.
A comparison of the models’ outcomes revealed that the ensemble LWL-AR model performed better than the other models in both training and testing phases of the monthly streamflow modeling. The LWL-DG and LWL-RF models showed similar performance and ranked as the second-best models, followed by the LWL-RS model that was identified as the least effective ensemble model. The results of other modeling studies support our findings that the application of the ensemble learning techniques can considerably improve the capability of the base models for modeling different environmental problems [26,29,47,59]. Overall, our case study demonstrated that the ensemble models successfully outperformed the single LWL model and provided promising accuracy for streamflow forecasting. Due to the non-linear nature of many environmental processes and phenomena (e.g., streamflow), hybrid ensemble models that benefit from the advantages of multiple methods/models can better capture the complexity of these phenomena and often yield more accurate results than single simple models.

7. Conclusions

This study investigated the capability of five ensemble models, that is, LWL-AR, LWL-BG, LWL-DG, LWL-RS, and LWL-RF, for monthly streamflow forecasting. The results were validated using several performance metrics and compared to those of a single LWL model. Based on the results obtained, we conclude that:
  • The ensemble models are predominantly superior to the single LWL model for monthly streamflow forecasting.
  • Among the ensemble methods, the LWL-AR model surpasses the other models in both training and testing performances.
  • The most accurate models are developed when the periodicity variable (MN, month number) is incorporated into the modeling process.
  • Ensemble forecasting is a robust and promising alternative to the single forecasting of streamflow.
Although the developed ensemble models were verified using a regional-scale dataset from Pakistan, they are sufficiently general to be applied in any other region around the world with minor adjustments in the variables relative to local conditions. Future research can extend this ensemble forecasting approach by using other ensemble learning techniques (e.g., AdaBoost, MultiBoost, LogitBoost, Decorate, etc.) and, perhaps even more interesting, by testing various types of state-of-the-art machine learning methods as the base classifier. The idea of coupling machine learning methods with ensemble learning techniques with the aim of enhancing the computational performance and improving the predictive accuracy can be extended beyond forecasting monthly streamflow to solve many other complex geo-hydrology problems. In this study, previous streamflow values and periodicity information were considered as inputs to the ensemble models. In future works, streamflow forecasting considering the flood mitigation capacity of Mangla Dam can be investigated using ensemble models. Furthermore, by taking into account the landforms (the digital terrain model) and the dimensions of the river basin as inputs, the implemented methods may provide more accurate forecasting results.

Author Contributions

Conceptualization: R.M.A., O.K., and A.J.; formal analysis: A.J., A.E., O.K., and R.M.A.; validation: R.M.A., A.J., A.M., A.E., and O.K.; supervision: A.J. and O.K.; writing original draft: R.M.A., A.J., A.M., A.E., and O.K.; visualization: R.M.A., A.J., and A.E.; investigation: R.M.A., A.J., and A.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Key R&D Program of China. (2016YFC0402706).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study will be available on interested request from the corresponding author.

Conflicts of Interest

There is no conflict of interest in this study.

References

  1. Zhang, W.; Hu, Y.; Liu, J.; Wang, H.; Wei, J.; Sun, P.; Wu, L.; Zheng, H. Progress of ethylene action mechanism and its application on plant type formation in crops. Saudi J. Biol. Sci. 2020, 27, 1667–1673. [Google Scholar] [CrossRef] [PubMed]
  2. Adnan, R.M.; Liang, Z.; Heddam, S.; Zounemat-Kermani, M.; Kisi, O.; Li, B. Least square support vector machine and multivariate adaptive regression splines for streamflow prediction in mountainous basin using hydro-meteorological data as inputs. J. Hydrol. 2020, 586, 124371. [Google Scholar] [CrossRef]
  3. Yuan, X.; Chen, C.; Lei, X.; Yuan, Y.; Adnan, R.M. Monthly runoff forecasting based on LSTM–ALO model. Stoch. Environ. Res. Risk Assess. 2018, 32, 2199–2212. [Google Scholar] [CrossRef]
  4. Liu, J.; Liu, Y.; Wang, X. An environmental assessment model of construction and demolition waste based on system dynamics: A case study in Guangzhou. Environ. Sci. Pollut. Res. 2019, 27, 37237–37259. [Google Scholar] [CrossRef]
  5. Mehran, A.; AghaKouchak, A.; Nakhjiri, N.; Stewardson, M.J.; Peel, M.C.; Phillips, T.J.; Wada, Y.; Ravalico, J.K. Compounding Impacts of Human-Induced Water Stress and Climate Change on Water Availability. Sci. Rep. 2017, 7, 6282. [Google Scholar] [CrossRef] [Green Version]
  6. Zhang, C.; Zhang, B.; Li, W.; Liu, M. Response of streamflow to climate change and human activity in Xitiaoxi river basin in China. Hydrol. Process. 2014, 28, 43–50. [Google Scholar] [CrossRef]
  7. Adnan, R.M.; Liang, Z.; Parmar, K.S.; Soni, K.; Kisi, O. Modeling monthly streamflow in mountainous basin by MARS, GMDH-NN and DENFIS using hydroclimatic data. Neural Comput. Appl. 2021, 33, 2853–2871. [Google Scholar] [CrossRef]
  8. Gibbs, M.S.; Dandy, G.C.; Maier, H.R. Assessment of the ability to meet environmental water requirements in the Upper South East of South Australia. Stoch. Environ. Res. Risk Assess. 2013, 28, 39–56. [Google Scholar] [CrossRef]
  9. Kişi, Ö. Streamflow Forecasting Using Different Artificial Neural Network Algorithms. J. Hydrol. Eng. 2007, 12, 532–539. [Google Scholar] [CrossRef]
  10. Yossef, N.C.; Winsemius, H.; Weerts, A.; Van Beek, R.; Bierkens, M.F.P. Skill of a global seasonal streamflow forecasting system, relative roles of initial conditions and meteorological forcing. Water Resour. Res. 2013, 49, 4687–4699. [Google Scholar] [CrossRef] [Green Version]
  11. Aqil, M.; Kita, I.; Yano, A.; Nishiyama, S. A comparative study of artificial neural networks and neuro-fuzzy in continuous modeling of the daily and hourly behaviour of runoff. J. Hydrol. 2007, 337, 22–34. [Google Scholar] [CrossRef]
  12. Abudu, S.; Cui, C.-L.; King, J.P.; Abudukadeer, K. Comparison of performance of statistical models in forecasting monthly streamflow of Kizil River, China. Water Sci. Eng. 2010, 3, 269–281. [Google Scholar]
  13. Wang, W. Stochasticity, Nonlinearity and Forecasting of Streamflow Processes; IOS Press: Amsterdam, The Netherlands, 2006. [Google Scholar]
  14. Rajaee, T. Wavelet and Neuro-fuzzy Conjunction Approach for Suspended Sediment Prediction. CLEAN Soil Air Water 2010, 38, 275–286. [Google Scholar] [CrossRef]
  15. Mehdizadeh, S.; Fathian, F.; Safari, M.J.S.; Adamowski, J.F. Comparative assessment of time series and artificial intelligence models to estimate monthly streamflow: A local and external data analysis approach. J. Hydrol. 2019, 579, 124225. [Google Scholar] [CrossRef]
  16. Adnan, R.M.; Petroselli, A.; Heddam, S.; Santos, C.A.G.; Kisi, O. Short term rainfall-runoff modelling using several machine learning methods and a conceptual event-based model. Stoch. Environ. Res. Risk Assess. 2021, 35, 597–616. [Google Scholar] [CrossRef]
  17. Rahgoshay, M.; Feiznia, S.; Arian, M.; Hashemi, S.A.A. Simulation of daily suspended sediment load using an improved model of support vector machine and genetic algorithms and particle swarm. Arab. J. Geosci. 2019, 12. [Google Scholar] [CrossRef]
  18. Kim, C.M.; Parnichkun, M. Prediction of settled water turbidity and optimal coagulant dosage in drinking water treatment plant using a hybrid model of k-means clustering and adaptive neuro-fuzzy inference system. Appl. Water Sci. 2017, 7, 3885–3902. [Google Scholar] [CrossRef] [Green Version]
  19. Affes, Z.; Kaffel, R.H. Forecast Bankruptcy Using a Blend of Clustering and MARS Model—Case of US Banks. SSRN Electron. J. 2016, 281, 27–64. [Google Scholar] [CrossRef] [Green Version]
  20. Adnan, R.M.; Liang, Z.; Trajkovic, S.; Zounemat-Kermani, M.; Li, B.; Kisi, O. Daily streamflow prediction using optimally pruned extreme learning machine. J. Hydrol. 2019, 577, 123981. [Google Scholar] [CrossRef]
  21. Zhang, X.; Peng, Y.; Zhang, C.; Wang, B. Are hybrid models integrated with data preprocessing techniques suitable for monthly streamflow forecasting? Some experiment evidences. J. Hydrol. 2015, 530, 137–152. [Google Scholar] [CrossRef]
  22. Tongal, H.; Booij, M.J. Simulation and forecasting of streamflows using machine learning models coupled with base flow separation. J. Hydrol. 2018, 564, 266–282. [Google Scholar] [CrossRef]
  23. Ferreira, R.G.; da Silva, D.D.; Elesbon, A.A.A.; Fernandes-Filho, E.I.; Veloso, G.V.; Fraga, M.D.S.; Ferreira, L.B. Machine learning models for streamflow regionalization in a tropical watershed. J. Environ. Manag. 2021, 280, 111713. [Google Scholar] [CrossRef]
  24. Piazzi, G.; Thirel, G.; Perrin, C.; Delaigue, O. Sequential Data Assimilation for Streamflow Forecasting: Assessing the Sensitivity to Uncertainties and Updated Variables of a Conceptual Hydrological Model at Basin Scale. Water Resour. Res. 2021, 57, 57. [Google Scholar] [CrossRef]
  25. Saraiva, S.V.; Carvalho, F.D.O.; Santos, C.A.G.; Barreto, L.C.; Freire, P.K.D.M.M. Daily streamflow forecasting in Sobradinho Reservoir using machine learning models coupled with wavelet transform and bootstrapping. Appl. Soft Comput. 2021, 102, 107081. [Google Scholar] [CrossRef]
  26. Tyralis, H.; Papacharalampous, G.; Langousis, A. Super ensemble learning for daily streamflow forecasting: Large-scale demonstration and comparison with multiple machine learning algorithms. Neural Comput. Appl. 2021, 33, 3053–3068. [Google Scholar] [CrossRef]
  27. Zhang, K.; Ruben, G.B.; Li, X.; Li, Z.; Yu, Z.; Xia, J.; Dong, Z. A comprehensive assessment framework for quantifying climatic and anthropogenic contributions to streamflow changes: A case study in a typical semi-arid North China basin. Environ. Model. Softw. 2020, 128, 104704. [Google Scholar] [CrossRef]
  28. Yen, H.P.H.; Pham, B.T.; Van Phong, T.; Ha, D.H.; Costache, R.; Van Le, H.; Nguyen, H.D.; Amiri, M.; Van Tao, N.; Prakash, I. Locally weighted learning based hybrid intelligence models for groundwater potential mapping and modeling: A case study at Gia Lai province, Vietnam. Geosci. Front. 2021, 12, 101154. [Google Scholar] [CrossRef]
  29. Tuyen, T.T.; Jaafari, A.; Yen, H.P.H.; Nguyen-Thoi, T.; Van Phong, T.; Nguyen, H.D.; Van Le, H.; Phuong, T.T.M.; Nguyen, S.H.; Prakash, I.; et al. Mapping forest fire susceptibility using spatially explicit ensemble models based on the locally weighted learning algorithm. Ecol. Inform. 2021, 63, 101292. [Google Scholar] [CrossRef]
  30. Atkeson, C.G.; Moore, A.W.; Schaal, S. Locally Weighted Learning. Artif. Intell. Rev. 1997, 11, 11–73. [Google Scholar] [CrossRef]
  31. Ahmadianfar, I.; Jamei, M.; Chu, X. A novel Hybrid Wavelet-Locally Weighted Linear Regression (W-LWLR) Model for Electrical Conductivity (EC) Prediction in Surface Water. J. Contam. Hydrol. 2020, 232, 103641. [Google Scholar] [CrossRef] [PubMed]
  32. Kisi, O.; Ozkan, C. A New Approach for Modeling Sediment-Discharge Relationship: Local Weighted Linear Regression. Water Resour. Manag. 2016, 31, 1–23. [Google Scholar] [CrossRef]
  33. Chen, T.; Ren, J. Bagging for Gaussian process regression. Neurocomputing 2009, 72, 1605–1610. [Google Scholar] [CrossRef] [Green Version]
  34. Zhou, Y.; Tian, L.; Zhu, C.; Jin, X.; Sun, Y. Video Coding Optimization for Virtual Reality 360-Degree Source. IEEE J. Sel. Top. Signal Process. 2020, 14, 118–129. [Google Scholar] [CrossRef]
  35. Azhari, M.; Abarda, A.; Alaoui, A.; Ettaki, B.; Zerouaoui, J. Detection of Pulsar Candidates using Bagging Method. Procedia Comput. Sci. 2020, 170, 1096–1101. [Google Scholar] [CrossRef]
  36. Xue, X.; Zhang, K.; Tan, K.C.; Feng, L.; Wang, J.; Chen, G.; Zhao, X.; Zhang, L.; Yao, J. Affine Transformation-Enhanced Multifactorial Optimization for Heterogeneous Problems. IEEE Trans. Cybern. 2020, 1–15. [Google Scholar] [CrossRef] [PubMed]
  37. Stone, C.J. Additive Regression and Other Nonparametric Models. Ann. Stat. 1985, 13, 689–705. [Google Scholar] [CrossRef]
  38. Piegorsch, W.W.; Xiong, H.; Bhattacharya, R.N.; Lin, L. Benchmark Dose Analysis via Nonparametric Regression Modeling. Risk Anal. 2013, 34, 135–151. [Google Scholar] [CrossRef] [Green Version]
  39. Zhang, M.; Yang, Z.; Liu, L.; Zhou, D. Impact of renewable energy investment on carbon emissions in China—An empirical study using a nonparametric additive regression model. Sci. Total Environ. 2021, 785, 147109. [Google Scholar] [CrossRef]
  40. Ho, T.K. The random subspace method for constructing decision forests. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 832–844. [Google Scholar] [CrossRef] [Green Version]
  41. Havlíček, V.; Córcoles, A.D.; Temme, K.; Harrow, A.W.; Kandala, A.; Chow, J.M.; Gambetta, J.M. Supervised learning with quantum-enhanced feature spaces. Nat. Cell Biol. 2019, 567, 209–212. [Google Scholar] [CrossRef] [Green Version]
  42. Kuncheva, L.I.; Rodriguez, J.J.; Plumpton, C.O.; Linden, D.E.J.; Johnston, S.J. Random Subspace Ensembles for fMRI Classification. IEEE Trans. Med. Imaging 2010, 29, 531–542. [Google Scholar] [CrossRef]
  43. Pham, B.T.; Bui, D.T.; Prakash, I.; Dholakia, M. Hybrid integration of Multilayer Perceptron Neural Networks and machine learning ensembles for landslide susceptibility assessment at Himalayan area (India) using GIS. Catena 2017, 149, 52–63. [Google Scholar] [CrossRef]
  44. Ting, K.M.; Witten, I.H. Stacking Bagged and Dagged Models; University of Waikato: Hamilton, New Zealand, 1997. [Google Scholar]
  45. Yariyan, P.; Janizadeh, S.; Van Phong, T.; Nguyen, H.D.; Costache, R.; Van Le, H.; Pham, B.T.; Pradhan, B.; Tiefenbacher, J.P. Improvement of Best First Decision Trees Using Bagging and Dagging Ensembles for Flood Probability Mapping. Water Resour. Manag. 2020, 34, 3037–3053. [Google Scholar] [CrossRef]
  46. Zuo, C.; Chen, Q.; Tian, L.; Waller, L.; Asundi, A. Transport of intensity phase retrieval and computational imaging for partially coherent fields: The phase space perspective. Opt. Lasers Eng. 2015, 71, 20–32. [Google Scholar] [CrossRef]
  47. Tran, Q.C.; Minh, D.D.; Jaafari, A.; Al-Ansari, N.; Minh, D.D.; Van, D.T.; Nguyen, D.A.; Tran, T.H.; Ho, L.S.; Nguyen, D.H.; et al. Novel Ensemble Landslide Predictive Models Based on the Hyperpipes Algorithm: A Case Study in the Nam Dam Commune, Vietnam. Appl. Sci. 2020, 10, 3710. [Google Scholar] [CrossRef]
  48. Malek, A.G.; Mansoori, M.; Omranpour, H. Random forest and rotation forest ensemble methods for classification of epileptic EEG signals based on improved 1D-LBP feature extraction. Int. J. Imaging Syst. Technol. 2021, 31, 189–203. [Google Scholar] [CrossRef]
  49. Jiang, Q.; Shao, F.; Lin, W.; Gu, K.; Jiang, G.; Sun, H. Optimizing Multistage Discriminative Dictionaries for Blind Image Quality Assessment. IEEE Trans. Multimed. 2018, 20, 2035–2048. [Google Scholar] [CrossRef]
  50. Pham, B.T.; Jaafari, A.; Avand, M.; Al-Ansari, N.; Du, T.D.; Yen, H.P.H.; Van Phong, T.; Nguyen, D.H.; Van Le, H.; Mafi-Gholami, D.; et al. Performance Evaluation of Machine Learning Methods for Forest Fire Modeling and Prediction. Symmetry 2020, 12, 1022. [Google Scholar] [CrossRef]
  51. Zhang, K.; Zhang, J.; Ma, X.; Yao, C.; Zhang, L.; Yang, Y.; Wang, J.; Yao, J.; Zhao, H. History Matching of Naturally Fractured Reservoirs Using a Deep Sparse Autoencoder. SPE J. 2021, 1–22. [Google Scholar] [CrossRef]
  52. Zhao, C.; Li, J. Equilibrium Selection under the Bayes-Based Strategy Updating Rules. Symmetry 2020, 12, 739. [Google Scholar] [CrossRef]
  53. Adnan, R.M.; Zounemat-Kermani, M.; Kuriqi, A.; Kisi, O. Machine Learning Method in Prediction Streamflow Considering Periodicity Component. In Understanding Built Environment; Springer: Berlin/Heidelberg, Germany, 2020; pp. 383–403. [Google Scholar]
  54. Kisi, O.; Shiri, J.; Karimi, S.; Adnan, R.M. Three different adaptive neuro fuzzy computing techniques for forecasting long-period daily streamflows. In Big Data in Engineering Applications; Springer: Singapore, 2018; pp. 303–321. [Google Scholar]
  55. Alizamir, M.; Kisi, O.; Muhammad Adnan, R.; Kuriqi, A. Modelling reference evapotranspiration by combining neuro-fuzzy and evolutionary strategies. Acta Geophys. 2020, 68, 1113–1126. [Google Scholar] [CrossRef]
  56. Zhao, J.; Liu, J.; Jiang, J.; Gao, F. Efficient Deployment with Geometric Analysis for mmWave UAV Communications. IEEE Wirel. Commun. Lett. 2020, 9, 1. [Google Scholar] [CrossRef]
  57. Kişi, Ö. River flow forecasting and estimation using different artificial neural network techniques. Hydrol. Res. 2008, 39, 27–40. [Google Scholar] [CrossRef]
  58. Adnan, R.M.; Yuan, X.; Kisi, O.; Yuan, Y.; Tayyab, M.; Lei, X. Application of soft computing models in streamflow forecasting. In Proceedings of the Institution of Civil Engineers—Water Management; Thomas Telford Ltd.: London, UK, 2019; Volume 172, pp. 123–134. [Google Scholar] [CrossRef]
  59. Pham, B.T.; Jaafari, A.; Van Phong, T.; Yen, H.P.H.; Tuyen, T.T.; Van Luong, V.; Nguyen, H.D.; Van Le, H.; Foong, L.K. Improved flood susceptibility mapping using a best first decision tree integrated with ensemble learning techniques. Geosci. Front. 2021, 12, 101105. [Google Scholar] [CrossRef]
Figure 1. Location map of study area.
Figure 1. Location map of study area.
Sustainability 13 05877 g001
Figure 2. Time variation graphs of the observed and predicted streamflow by different LWL-based ensemble models in the testing phase using the best input combination (M3-IV).
Figure 2. Time variation graphs of the observed and predicted streamflow by different LWL-based ensemble models in the testing phase using the best input combination (M3-IV).
Sustainability 13 05877 g002
Figure 3. Comparison of different LWL-based ensemble models for forecasting peak streamflow in the testing phase using the best input combination (M3-IV).
Figure 3. Comparison of different LWL-based ensemble models for forecasting peak streamflow in the testing phase using the best input combination (M3-IV).
Sustainability 13 05877 g003
Figure 4. Comparison of different LWL-based ensemble models for forecasting low streamflow in the testing phase using the best input combination (M3-IV). Notice that only the streamflow values lower than 500 m3/s are shown.
Figure 4. Comparison of different LWL-based ensemble models for forecasting low streamflow in the testing phase using the best input combination (M3-IV). Notice that only the streamflow values lower than 500 m3/s are shown.
Sustainability 13 05877 g004
Figure 5. Scatterplots of the observed and predicted streamflow by (a) LWL, (b) LWL-AR, (c) LWL-BG, (d) LWL-DG, (e) LWL-RS, (f) LWL-RF ensemble models in the testing phase using the best input combination (M3-IV).
Figure 5. Scatterplots of the observed and predicted streamflow by (a) LWL, (b) LWL-AR, (c) LWL-BG, (d) LWL-DG, (e) LWL-RS, (f) LWL-RF ensemble models in the testing phase using the best input combination (M3-IV).
Sustainability 13 05877 g005
Figure 6. Taylor diagram of different LWL-based ensemble models in the testing phase using the best input combination (M3-IV).
Figure 6. Taylor diagram of different LWL-based ensemble models in the testing phase using the best input combination (M3-IV).
Sustainability 13 05877 g006
Figure 7. Violin diagrams of different LWL-based ensemble models in the testing using the best input combination (M3-IV).
Figure 7. Violin diagrams of different LWL-based ensemble models in the testing using the best input combination (M3-IV).
Sustainability 13 05877 g007
Table 1. An overview of the statistical characteristics of the data used.
Table 1. An overview of the statistical characteristics of the data used.
StatisticsWhole Dataset (m3/s) 1965 to 2012M1 Dataset
(m3/s) 2001 to 2012
M2 Dataset
(m3/s) 1989 to 2000
M3 Dataset
(m3/s) 1977 to 1988
M4 Dataset
(m3/s) 1965 to 1976
Mean772.9794.0783.7835.8678.0
Min.110.7112.3134.9127.0110.7
Max.28242824242627732014
Skewness0.8860.9310.7160.8450.888
Std. dev.609.2645.1600.6651.7514.1
Variance371,069416,106360,780424,712264,330
Table 2. Optimum parameter setting of the models.
Table 2. Optimum parameter setting of the models.
ParameterModel
LWLARBGDGRSRF
DebugFalseFalseFalseFalseFalseFalse
Search algorithmLinear NN search-----
Weighting kernel0-----
Number of iterations-1412101011
Shrinkage-0.1----
Bag size percent--100---
Seed--1111
Number of folds---10--
Verbose---False--
Number of boosting iterations-30----
Subspace size----0.5-
Max group-----3
Min group-----3
Number of groups-----False
Projection filter-----PCA
Removed percentage-----50
Table 3. Results of the single LWL model.
Table 3. Results of the single LWL model.
MetricData SetTrainingTesting
Input CombinationInput Combination
IIIIIIIVIIIIIIIV
RMSEM1358.6300.3295.5255.9365.8308.6311.4295.4
M2358.7303.7275.5242.1397.0370.2369.5328.0
M3358.8283.8271.5244.0382.1303292.9274.8
M4362.3306.5300.4252.3397.9342.2312.4277.7
Mean359.6298.6285.7248.6385.7331.0321.6294.0
MAEM1282.6227.7226.0183.5271.0231.3241.4207.7
M2279.9227.0210.1178.8306.9263.9265.0222.2
M3274.4213.8204.9175.0291.5228.8219.9199.2
M4281.5230.8229.0183.5309.8257.2240.9200.0
Mean279.6224.8217.5180.2294.8245.3241.8207.3
RAEM152.2442.0941.7835.9257.5749.1251.2744.12
M255.6744.1441.3935.5756.6848.7448.9541.03
M353.3542.5140.7534.4756.0143.9542.2538.70
M455.4745.4744.5333.6757.5647.7944.7538.16
Mean54.1843.5542.1134.9156.9647.4046.8140.50
RRSEM158.8047.4246.6439.9465.8157.858.3255.32
M260.5149.6246.1940.8563.6056.3456.2449.91
M358.6347.8845.8040.9060.4250.4348.7444.08
M460.7451.3849.0941.7261.6252.9948.3846.24
Mean59.6749.0846.9340.8562.8654.3952.9248.89
RM10.6590.7760.7830.8410.5940.6720.6760.746
M20.6420.7500.7920.8340.6120.6870.6940.759
M30.6580.7730.7920.8340.6290.7460.7620.809
M40.6340.7360.7590.8260.6190.7230.7570.789
Mean0.6480.7590.7820.8340.6140.7070.7220.776
The best performance is shown in bold.
Table 4. Results of the ensemble LWL-AR model.
Table 4. Results of the ensemble LWL-AR model.
MetricDatasetTrainingTesting
Input CombinationInput Combination
IIIIIIIVIIIIIIIV
RMSEM1321.0184.5162.4143.7327.5292.8293.3261.9
M2310.0183.8170.3128.3407.8334.4315.2273.5
M3306.9174.0152.1138.4373.6264.4258.8223.9
M4314.2193.3169.1139.1377.6294.6284.2242.9
Mean313.0183.9163.5137.4371.6296.6287.9250.6
MAEM1248.1135.8115.897.47247.2199.8195.9171.1
M2241.6130.212088.47310.9224.2209.4168.6
M3232.2125.1104.895.72292.1191.7183.8150.7
M4242.6136.6117.895.72295.1200.4198.6156.0
Mean241.1131.9114.694.30286.3204.0196.9161.6
RAEM145.8725.1121.4317.7052.5142.4441.6133.90
M248.0425.3123.3217.6057.4341.4138.6731.13
M345.1524.8720.8318.9556.1136.8235.3128.95
M447.7926.9223.2018.6054.8337.2336.8931.80
Mean46.7125.5522.2018.2155.2239.4838.1231.45
RRSEM150.6929.1325.6421.8661.3354.8354.9345.49
M252.2930.0327.8321.6462.0650.9047.9641.62
M350.1529.3625.6523.4862.1844.0043.0823.26
M452.6732.428.3423.3158.4745.6344.0240.57
Mean51.4530.2326.8722.5761.0148.8447.5037.74
RM10.7430.9160.9350.9530.6210.7400.7330.823
M20.7280.9100.9240.9530.6120.7430.7730.828
M30.7500.9140.9350.9450.6160.8080.8210.867
M40.7230.9030.9220.9470.6580.7940.8060.835
Mean0.7360.9110.9290.9500.6270.7710.7830.838
The best performance is shown in bold.
Table 5. Results of the ensemble LWL-BG model.
Table 5. Results of the ensemble LWL-BG model.
MetricDatasetTrainingTesting
Input CombinationInput Combination
IIIIIIIVIIIIIIIV
RMSEM1363.6290.0272.9240.9345.5294.0274.3261.2
M2352.3285.8266.5237.9398.1340.6345.2306.5
M3336.6262.0250.7229.4359.4292.8276.1255.3
M4342.8289.5272.6250.5376.4319.1294.7258.8
Mean348.8281.8265.7239.7369.9311.6297.6270.5
MAEM1284.4224.2212.6177.4269.0226.6218.5187.7
M2273.6217.4202.1174.8310.9243.5247.3208.0
M3265.9202.0191.0169.8279.4223.6209.5191.7
M4270.6220.8205.1183.0300.5248.6225.3182.8
Mean273.6216.1202.7176.3290.0235.6225.2192.6
RAEM152.5841.4539.3132.9057.1348.1446.4038.83
M253.1942.2739.8334.4557.4244.9745.6938.43
M352.8840.1637.9833.7753.6842.9540.2434.88
M453.3243.5039.8835.5755.8346.1941.8636.83
Mean52.9941.8539.2534.1756.0245.5643.5537.24
RRSEM157.4245.7943.1038.0464.7155.0551.3748.46
M257.5646.7144.6739.8860.5851.8352.5446.65
M356.7744.1942.2938.7059.8248.7345.9640.45
M457.4748.5344.5440.9458.3049.4245.6442.48
Mean57.3146.3143.6539.3960.8551.2648.8844.51
RM10.6720.7940.8170.8590.5900.6940.7430.781
M20.6690.7830.8030.8450.6270.7360.7310.796
M30.6790.8080.8240.8540.6460.7620.7890.845
M40.6710.7660.8030.8340.6610.7600.7990.821
Mean0.6730.7880.8120.8480.6310.7380.7660.811
The best performance is shown in bold.
Table 6. Results of the ensemble LWL-DG model.
Table 6. Results of the ensemble LWL-DG model.
MetricDatasetTrainingTesting
Input CombinationInput Combination
IIIIIIIVIIIIIIIV
RMSEM1369.2310.1279.1241.0320.3274.1259.0249.4
M2355.4296.3270.9264.8390.6335.8326.4298.8
M3338.3285.6262.6234.5349.6288.2253.9233.5
M4346.6299.3271.7239.5337.6324.9286.0247.1
Mean352.4297.8271.1245.0349.5305.8281.3257.2
MAEM1286.0236.6211.7171.3248.0217.4206.0191.1
M2275.2225.8201.7194.0299.3246.8229.3209.4
M3262.2220.1200.1166.9274.3227.2197.8181.3
M4267.5225.2207.3181.4298.8252.2220.8179.4
Mean272.7226.9205.2178.4280.1235.9213.5190.3
RAEM152.8743.7539.1431.6752.6746.1843.7639.85
M253.5143.8939.2137.7155.2845.5942.3638.68
M352.1543.7739.8032.8952.6943.6637.9933.34
M452.7044.3740.8436.0755.5246.8541.0236.72
Mean52.8143.9539.7534.5954.0445.5741.2837.15
RRSEM158.3048.9744.0738.0659.9851.3248.5044.56
M258.0748.4244.2743.2859.4451.1049.6745.48
M357.0648.1844.3039.3058.1847.9642.2638.26
M458.0950.1845.5440.4058.6450.3244.2941.52
Mean57.8848.9444.5540.2659.0650.1846.1842.46
RM10.6630.7660.8140.8670.6230.7240.7530.803
M20.6630.7710.8060.8150.6430.7530.7780.797
M30.6760.7740.8150.8480.6630.7740.8240.847
M40.6630.7530.7970.8410.6590.7640.8210.828
Mean0.6660.7660.8080.8430.6470.7540.7940.819
The best performance is shown in bold.
Table 7. Results of the ensemble LWL-RS model.
Table 7. Results of the ensemble LWL-RS model.
MetricDatasetTrainingTesting
Input CombinationInput Combination
IIIIIIIVIIIIIIIV
RMSEM1371.3329.9287.0270.4351.4302.1274.1248.7
M2358.7319.8301.1282.0397.0362.1371.6345.5
M3359.8317.2279.3268.2382.1326.8302.7242.8
M4362.3296.7319.8280.4397.9344.3319.2302.5
Mean363.0315.9296.8275.3382.1333.8316.9284.9
MAEM1282.6261.1225.2200.3271.0248.1221.8192.7
M2279.9238.0231.8207.2306.9295.5274.1243.0
M3274.4245.1215.7205.9291.5263.8238.5191.4
M4281.5248.6238.0219.3309.8276.7240.7228.3
Mean279.6248.2227.7208.2294.8271.0243.8213.9
RAEM152.2546.7641.6437.0957.5652.6847.1140.66
M255.6849.1345.0640.2956.6853.4350.6344.88
M353.3550.6442.5140.5756.0151.2745.8137.03
M455.4748.7849.1343.6257.5654.4144.7342.41
Mean54.1948.8344.5940.3956.9552.9547.0741.25
RRSEM158.6349.8845.3242.7065.8161.7151.3245.47
M260.5155.9749.2046.0860.4258.5456.5652.58
M358.8054.8946.8244.9663.6057.6850.3941.39
M460.7454.8355.9747.3061.6257.7249.4346.86
Mean59.6753.8949.3345.2662.8658.9151.9346.58
RM10.6590.6760.8060.8370.5940.6370.7360.796
M20.6420.7140.7690.8140.6290.6590.7020.773
M30.6590.6820.7900.8210.6120.6760.7500.848
M40.6340.6790.7140.7920.6190.6710.7690.815
Mean0.6490.6880.7700.8160.6140.6610.7390.808
The best performance is shown in bold.
Table 8. Results of the ensemble LWL-RF model.
Table 8. Results of the ensemble LWL-RF model.
MetricDatasetTrainingTesting
Input CombinationInput Combination
IIIIIIIVIIIIIIIV
RMSEM1371.3259.9261.7225.4351.4289.8278.3232.4
M2359.8271.7269.4229.1397.0307.4336.3300.2
M3358.7242.9253.0213.8382.1265.6266.3229.4
M4362.3271.0311.5230.6397.9297.3311.4266.4
Mean363.0261.4273.9224.7382.1290.0298.1257.1
MAEM1282.6196.2200.0167.4271.0218.0212.8195.7
M2274.4203.4205.2165.0306.9223.1237.4201.7
M3279.9184.2193.0160.0291.5195.5195.6173.0
M4281.5204.1237.8167.3309.8227.8237.8175.7
Mean279.6197.0209.0164.9294.8216.1220.9186.5
RAEM152.2536.2736.9730.9657.5746.3045.1937.31
M253.3539.5539.9032.0756.6841.2043.8637.25
M355.6836.6238.3831.3656.0137.5637.5733.23
M455.4740.2244.1832.9657.5642.3244.1836.37
Mean54.1938.1739.8631.8456.9641.8542.7036.04
RRSEM158.6341.0341.3335.5965.8154.2752.1142.96
M258.8044.4144.0337.4460.4246.7851.1745.68
M360.5140.9842.6836.0663.6044.2144.3238.68
M460.7445.4248.2438.6661.6246.0448.2441.26
Mean59.6742.9644.0736.9462.8647.8348.9642.15
RM10.6590.8340.8300.880.5940.7140.7530.821
M20.6590.8050.8060.8690.6290.7870.7500.806
M30.6420.8350.8190.8820.6120.8080.8050.858
M40.6340.7960.7710.8560.6190.7990.7710.846
Mean0.6490.8180.8070.8720.6140.7770.7700.833
The best performance is shown in bold.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Adnan, R.M.; Jaafari, A.; Mohanavelu, A.; Kisi, O.; Elbeltagi, A. Novel Ensemble Forecasting of Streamflow Using Locally Weighted Learning Algorithm. Sustainability 2021, 13, 5877. https://doi.org/10.3390/su13115877

AMA Style

Adnan RM, Jaafari A, Mohanavelu A, Kisi O, Elbeltagi A. Novel Ensemble Forecasting of Streamflow Using Locally Weighted Learning Algorithm. Sustainability. 2021; 13(11):5877. https://doi.org/10.3390/su13115877

Chicago/Turabian Style

Adnan, Rana Muhammad, Abolfazl Jaafari, Aadhityaa Mohanavelu, Ozgur Kisi, and Ahmed Elbeltagi. 2021. "Novel Ensemble Forecasting of Streamflow Using Locally Weighted Learning Algorithm" Sustainability 13, no. 11: 5877. https://doi.org/10.3390/su13115877

APA Style

Adnan, R. M., Jaafari, A., Mohanavelu, A., Kisi, O., & Elbeltagi, A. (2021). Novel Ensemble Forecasting of Streamflow Using Locally Weighted Learning Algorithm. Sustainability, 13(11), 5877. https://doi.org/10.3390/su13115877

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop