Next Article in Journal
Assessing Climate Change Impacts on Spring Discharge in Data-Sparse Environments Using a Combined Statistical–Analytical Method: An Example from the Aggtelek Karst Area, Hungary
Previous Article in Journal
Assessing Urban Lake Performance for Stormwater Harvesting: Insights from Two Lake Systems in Western Sydney, Australia
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A DeepAR-Based Modeling Framework for Probabilistic Mid–Long-Term Streamflow Prediction

1
Water Resources Department, Changjiang River Scientific Research Institute, Wuhan 430010, China
2
Hubei Key Laboratory of Water Resources & Eco-Environmental Sciences, Changjiang River Scientific Research Institute, Wuhan 430010, China
3
Research Center on the Yangtze River Economic Belt Protection and Development Strategy, Wuhan 430010, China
4
Hubei Key Laboratory of Intelligent Yangtze and Hydroelectric Science, China Yangtze Power Co., Ltd., Yichang 443300, China
*
Authors to whom correspondence should be addressed.
Water 2025, 17(17), 2506; https://doi.org/10.3390/w17172506
Submission received: 14 July 2025 / Revised: 4 August 2025 / Accepted: 21 August 2025 / Published: 22 August 2025
(This article belongs to the Section Hydrology)

Abstract

Mid–long-term streamflow prediction (MLSP) plays a critical role in water resource planning amid growing hydroclimatic and anthropogenic uncertainties. Although AI-based models have demonstrated strong performance in MLSP, their capacity to quantify predictive uncertainty remains limited. To address this challenge, a DeepAR-based probabilistic modeling framework is developed, enabling direct estimation of streamflow distribution parameters and flexible selection of output distributions. The framework is applied to two case studies with distinct hydrological characteristics, where combinations of recurrent model structures (GRU and LSTM) and output distributions (Normal, Student’s t, and Gamma) are systematically evaluated. The results indicate that the choice of output distribution is the most critical factor for predictive performance. The Gamma distribution consistently outperformed those using Normal and Student’s t distributions, due to its ability to better capture the skewed, non-negative nature of streamflow data. Notably, the magnitude of performance gain from using the Gamma distribution is itself region-dependent, proving more significant in the basin with higher streamflow skewness. For instance, in the more skewed Upper Wudongde Reservoir area, the model using LSTM structure and Gamma distribution reduces RMSE by over 27% compared to its Normal-distribution counterpart (from 1407.77 m3/s to 1016.54 m3/s). Furthermore, the Gamma-based models yield superior probabilistic forecasts, achieving not only lower CRPS values but also a more effective balance between high reliability (PICP) and forecast sharpness (MPIW). In contrast, the relative performance between GRU and LSTM architectures was found to be less significant and inconsistent across the different basins. These findings highlight that the DeepAR-based framework delivers consistent enhancement in forecasting accuracy by prioritizing the selection of a physically plausible output distribution, thereby providing stronger and more reliable support for practical applications.

1. Introduction

The combined impacts of climate change and human activities are profoundly altering the processes of runoff generation and confluence, resulting in increasing uncertainty in the evolution of water resources [1,2,3,4]. At the same time, socio-economic drivers such as population growth and urbanization are intensifying the demand for water supply [5,6,7,8]. Against this backdrop, mid–long-term streamflow prediction (MLSP), which refers to forecasts at ten-day, monthly, seasonal, and annual scales with lead times ranging from three days to one year, has become increasingly vital for effective water resource management and integrated utilization, as it provides valuable insights into future runoff patterns [9,10,11,12,13,14]. Consequently, MLSP is attracting growing attention in both research and practical applications [11,12,13,14].
Many models have been developed and applied in MLSP to improve the predictive performance and provide information for the comprehensive utilization of water resources [11,12,13,14,15,16,17,18]. These models can be broadly divided into physical-based models, which simulate the streamflow based on the runoff generation and confluence equations, and data-driven models, which directly simulate the relationship between streamflow and predictors including precipitation, temperature, and other factors [13,14,19,20,21,22]. Along with the development of artificial intelligence (AI) methods, the AI-based data-driven models, including support vector regression (SVR), artificial neural network (ANN), gated recurrent unit neural network (GRU), long short-term memory network (LSTM) and so on, can obtain better predictive performance than traditional models and have become predominant in the MLSP [10,20,23,24,25,26,27,28]. For instance, many studies applied SVR models in MLSP, and the results demonstrate that the SVR models can generate more accurate forecasts than linear models and ANN models [29,30,31,32,33]. Xie et al. (2024) compared five AI-based models, and the results demonstrate that the LSTM model outperformed other models in forecasting monthly streamflow in 37 basins [14]. In addition, some hybrid models which incorporate preprocessing methods into the AI methods are developed to integrate the advantages of different base models and have obtained better forecasting skill [19,34,35,36,37].
The proposed AI-based models in MLSP demonstrate significant improvements in predictive performance, but their limited capacity to characterize future water resource uncertainties constrains practical applications [38,39,40,41]. To overcome this issue, many post-processing methods are adopted to produce ensemble forecasting results capable of characterizing predictive uncertainties [40,41,42,43]. For example, Liang et al. (2018) proposes the hydrological uncertainty processor to post-process the deterministic outputs from the SVR model to quantify prediction uncertainties [39]. Mo et al. (2023) applies the generalized autoregressive conditional heteroskedasticity model to identify time-varying forecasting errors to improve the predictive performance [44].
Although the post-processing approach in MLSP demonstrates predictive capability, it exhibits two fundamental limitations: (1) inability to directly generate probability distributions and (2) failure to preserve the inherent statistical characteristics of streamflow. To overcome these constraints in MLSP, this study adopts a DeepAR-based modeling framework. The DeepAR architecture exhibits two key capabilities: (1) direct prediction of the predictand’s distribution parameters and (2) flexible selection of probability distributions for the target variable [45,46]. It has been applied in many studies, and empirical validation across five time series datasets demonstrates its superior performance over other existing state-of-the-art methods [45,47,48,49]. However, the DeepAR’s robustness under heavy-tailed streamflow distributions requires further verification, and the selection criteria for appropriate probability distributions lack systematic guidance. Therefore, the objectives of this study are (1) to develop a DeepAR-based probabilistic forecasting framework for MLSP, (2) to validate modeling framework’s applicability through implementation in two case studies, and (3) to systematically evaluate impacts of base model architecture and distribution type selection on predictive skill.

2. Materials and Methods

2.1. DeepAR Model

DeepAR is a probabilistic forecasting framework developed by Amazon Research, which combines recurrent neural networks (RNNs) with parametric probability distributions to generate time-series predictions [45]. Unlike traditional point-forecasting models, DeepAR directly outputs the parameters of user-specified distributions (e.g., Gaussian for real-valued data, Negative Binomial for positive count data, and Beta for data in the unit interval), enabling native uncertainty quantification.
Let z t denote the streamflow value at time t, and x t represent the vector of predictor variables at time t. The DeepAR model estimates the conditional distribution of future streamflow values:
P ( z t 0 + 1 : t 0 + L | z t 0 H + 1 : t 0 , x t 0 H + 1 : t 0 + L )
where t 0 is the forecast initialization time, H is the length of the target time series at prediction time, and L is the prediction length.
For a trained DeepAR model, the distribution parameters θ t at time t are computed as a function of the hidden state h t and model parameters Θ :
θ t = f ( h t , Θ )
where f ( · ) is a function used to map the hidden state to distribution parameters and the hidden state h t evolves recursively via
h t = g ( h t 1 , z t 1 , x t , Θ )
where g ( · ) is a nonlinear transition function implemented as a multi-layer RNN (LSTM or GRU) parametrized by Θ , and h t 1 is the hidden state from the previous time step.
Then, the simulated or forecasted streamflow value at time t can be sampled by
z t ~ ~ p ( · | f ( h t , Θ ) )

2.1.1. Training

Given a streamflow time series z t t = 1,2 , , T and associated predictor variables x t , the DeepAR model parameters Θ , including the parameters of both f · and g ( · ) , can be learned by maximizing the log-likelihood, as shown below:
L = t 0 t = t 0 + 1 t 0 + L log p ( z t | f ( h t , Θ ) )

2.1.2. Prediction

Given an observed streamflow sequence z t 0 H + 1 : t 0 and corresponding predictor variables x t 0 H + 1 : t 0 + L , the trained DeepAR model generates probabilistic forecasts for future streamflow values z t 0 + 1 : t 0 + L through the following procedure:
(1)
The hidden state h t 0 is obtained by recursively processing the historical streamflow z t 0 H + 1 : t 0 and predictors x t 0 H + 1 : t 0 through the RNN transition function in Equation (3);
(2)
Initial conditions are set as h t 0 ~ = h t 0 and z t 0 ~ = z t 0 ;
(3)
For each subsequent time step t = t 0 + 1 to t 0 + L , the hidden state h t ~ is updated using g ( h t 1 ~ , z t 1 ~ , x t , Θ ) , and the forecast z t ~ is sampled from p ( · | f ( h t ~ , Θ ) ) ;
(4)
Step (3) is repeated N times to produce ensemble forecasts z i , t 0 + 1 : t 0 + L i = 1,2 , , N , providing a Monte Carlo approximation of the predictive distribution.

2.1.3. Likelihood Model

The likelihood p ( · | f ( h t , Θ ) ) , which defines the target distribution, must be carefully selected to match the statistical characteristics of predictand. While Salinas et al. (2020) [45] recommended Gaussian distributions for real-valued data and Negative Binomial distributions for positive count data, these choices prove suboptimal for streamflow forecasting due to (1) heavy-tailed characteristics of streamflow, and (2) the strictly positive nature of streamflow values. Therefore, to better characterize the statistical properties of streamflow, this study employs Gamma distribution. This choice is substantiated by its strong physical basis, as the Gamma function can be derived from conceptual watershed runoff models, distinguishing it from alternative distributions, which are applied primarily for their statistical fitting capabilities [50]. Its proven ability to accurately represent diverse streamflow regimes has also led to its widespread acceptance and recommendation in hydrological frequency analysis [51]. In addition to the physically grounded Gamma distribution, this study also employs Student’s t-distribution, which is suitable for extreme events, and conventional Gaussian distribution, which has been widely proposed.
For Student’s t-distribution, the likelihood and function f ( · ) are as below:
p S T z μ , σ , ν = Γ ν + 1 2 Γ ν 2 ν π σ 1 + z μ 2 ν σ 2 ν + 1 2 μ h t = w μ T h t + b μ σ h t = s o f t p l u s w σ T h t + b σ ν h t = 2 + s o f t p l u s w ν T h t + b ν
where the softplus activation function s o f t p l u s x = log ( 1 + e x ) is applied to enforce positivity constraints on the distribution parameters.
For Gamma distribution, the likelihood and function f ( · ) are as below:
p G a m m a z α , β = β α Γ α z α 1 e β z α h t = s o f t p l u s w α T h t + b α β h t = s o f t p l u s w β T h t + b β

2.2. DeepAR-Based Modeling Framework

This study develops a DeepAR-based probabilistic forecasting framework for MLSP (Figure 1), which consists of four main steps: (1) data preparing, involving data collection and optimal distribution selection; (2) data splitting, which spites data into training, validation and test datasets; (3) model calibration, involving model training and selection; and (4) model evaluation, involving model performance assessment [52,53].

2.2.1. Data Preparing

Data preparing and analysis are proposed to collect streamflow and corresponding predictors (precipitation), and select optimal distribution for the streamflow. To select the most appropriate distribution from these candidates, both the Akaike Information Criterion (AIC) and Bayesian Information Criterion (BIC) are employed to evaluate the goodness-of-fit between the observed data and theoretical distributions. Then, the optimal distribution is selected by minimizing both AIC and BIC values, which are computed as
A I C = 2 l n ( L i ) + 2 k A I C = 2 l n ( L i ) + k l n ( n )
where Li is the maximized likelihood value, k denotes the parameter number, and n is the length of streamflow time series.

2.2.2. Data Splitting

Data splitting is an important step in the data-driven modeling process, through which the available data is divided into training, validation, and test datasets [52,53,54]. In this study, the data after a specific time point is first separated as the test set to ensure no overlap between the test data and other data. Subsequently, the remaining data is randomly shuffled and split into training and validation datasets in a specific ratio.

2.2.3. Model Calibration

Model calibration is used to optimize model architecture and parameters in order for the model to represent the underlying relationships between predictors and predictand. First, the optimal streamflow distribution obtained by minimizing the AIC/BIC metrics is adopted to define the model’s output distribution type. Furthermore, taking into account the specific architecture of the DeepAR model, this study utilizes two types of RNN structures (i.e., LSTM and GRU), the detailed structures and gating equations of which can be seen in Demir (2025) [55]. Then, multiple model variants are generated by varying input conditions, and parameters are optimized using a training dataset by the ADAM optimizer, which is selected for its proven efficiency and robust performance in handling complex models through adaptive learning rates [56,57]. Finally, the model variants are used in a validation dataset and compared to select the best model in terms of their predictive performance.

2.2.4. Model Evaluation

Model evaluation is used to assess the predictive performance of the selected model over an independent dataset (i.e., test dataset) [58]. The deterministic and probabilistic predictions are generated, and two types of metrics are proposed to evaluate the predictive performance [22,53,59,60]. The performance of the deterministic predictions is evaluated using the following three metrics: the root mean square error (RMSE) and mean absolute error (MAE), which reflect the overall prediction accuracy, and the Nash–Sutcliffe efficiency (NSE) coefficient, which indicates the overall model effectiveness. The performance of the probabilistic predictions is evaluated using the following three metrics: (1) continuous ranked probability score (CRPS), which measures the overall probabilistic predictive performance; (2) prediction interval coverage probability (PICP), which measures the proportion of observations that fall within their predicted probability intervals; and (3) mean prediction interval width (MPIW), which indicates the average width of the prediction intervals and thus the sharpness of the forecast. These six metrics can be calculated according to the following equations:
R M S E = 1 n i = 1 n Q i ^ Q i 2 M A E = 1 n i = 1 n | Q i ^ Q i | N S E = 1 i = 1 n Q i ^ Q i 2 i = 1 n Q i 1 n i = 1 n Q i 2 C R P S = 1 N i = 1 N F i x H x Q i 2 d x H x Q i = 0 x < Q i 1 x Q i P I C P = 1 n i = 1 n c i × 100 %         c i = 1 L i Q i U i 0 o t h e r w i s e         M P I W = 1 n i = 1 n U i L i
where n is the sample size in the test dataset, i is the sample index, Q i is the observed streamflow, Q i ^ is the ensemble mean prediction, L i and U i are the lower and upper bounds of the prediction interval, and F i ( ) is the cumulative distribution function (CDF) of the probabilistic forecast.

2.3. Case Study and Data

The performance of the DeepAR-based modeling framework is examined in two basins upstream of the Wudongde (WDDR) and Sanxia (SXR) reservoirs, as illustrated in Figure 2. The available data and statistical characteristics for the two basins is presented in Table 1, including ten-day naturalized streamflow (Figure 3) and areal mean precipitation records spanning January 1980 to September 2022. To ensure data compatibility with neural network requirements, all variables are first normalized before model processing and then inversely transformed to their original scales. For the streamflow variable, the data is standardized using the Z-score method and then shifted by a constant to ensure all values were above 0, as required by the Gamma distribution. For the precipitation variable, min–max normalization is used to scale its values into the range 0–1.

2.4. Experiment Setup

The overall experimental setup, based on the DeepAR-based modeling framework, is illustrated in Figure 4. Three candidate distribution types—Gaussian, Gamma, Student’s t-distribution—are provided to account for streamflow characteristics. Then, the data after January 2018 is separated as the test dataset, while the other data is randomly split into training and validation datasets at an 8:2 ratio, following the recommendations in previous studies [52,53]. This splitting method effectively prevents data leakage between the test dataset and the training/validation datasets. The resulting training/validation and test datasets are visualized in Figure 3. As shown in Figure 3, the characteristics of the streamflow in the test dataset are largely consistent with those in the training/validation datasets for both the Upper WDDR and Upper SXR areas. Additionally, it is evident that the streamflow in neither basin exhibits a statistically significant increasing or decreasing linear trend over time. The coefficients of determination R2 with time are 0.000 for both areas, with corresponding p-values of 0.876 and 0.681, respectively.
During the model calibration phase, different models for predicting streamflow in the next 18 ten-day periods, are trained based on varying input conditions that incorporate combinations of three precipitation input scenarios (temporal lags: [0], [0, 1] or [0, 1, 2]) and two streamflow input scenarios (temporal lags: [1], [1, 2]), with the model training hyperparameters detailed in Table A1. The optimal input combination is then selected based on a comparative evaluation of all model variants’ predictive performance (i.e., RMSE) on the validation dataset.
After data preparing and model calibration, the final model combining the optimal input configuration and the probability distribution output is established and applied to produce deterministic predictions and probabilistic predictions with 100 members in this study. In order to evaluate the impact of output distribution and RNN structure, alternative models with different probability distribution outputs and different RNN structures (LSTM and GRU), named GRU-N, GRU-S, GRU-G, LSTM-N, LSTM-S, LSTM-G, are established and compared in terms of their predictive performance on the testing dataset. In the “GRU-N” naming structure, the former denotes the RNN structure (including GRU and LSTM), while the latter denotes the model output distribution (including Normal distribution N, Student’s t-distribution S, and Gamma distribution G).

3. Results

Following the modeling framework, the results are presented in three sections: optimal probability distribution selection (Section 3.1), input configuration optimization (Section 3.2), and testing performance evaluation (Section 3.3).

3.1. Optimal Probability Distribution Selection

The distribution fitting results for both study areas over the data excluding the test dataset are presented in Figure 5, while the corresponding AIC and BIC values are listed in Table 2. The results reveals that the Gamma distribution provides superior statistical performance compared to the Normal and Student’s t-distributions. For the Upper WDDR area, the Gamma distribution achieves the lowest AIC (27,321.09) and BIC (27,337.03) values, substantially outperforming the Normal distribution (AIC: 28,574.36, BIC: 28,584.99) and Student’s t-distribution (AIC: 28,455.62, BIC: 28,471.57). Similarly, in the Upper SXR area, the Gamma distribution exhibits superior performance, with AIC (30,848.71) and BIC (30,864.66) values considerably lower than the alternative distributions. The probability density plots also illustrate that the Gamma distribution more accurately captures the right-skewed characteristics and tail behavior of the streamflow data in both study areas.

3.2. Input Configuration Optimization

The comparative analysis of different input configurations reveals consistent performance patterns across both study areas, which can be shown in Table 3. For the Upper WDDR area, the precipitation input scenario with temporal lags [0, 1, 2] achieves the lowest RMSE (1199.21 m3/s), indicating superior predictive accuracy when incorporating precipitation data from the current time step and two previous time steps. In contrast, the precipitation-only scenario with no temporal lag [0] shows the poorest performance with the highest RMSE (1277.07 m3/s). Similarly, in the Upper SXR area, the [0, 1, 2] precipitation configuration demonstrates optimal performance, with an RMSE value of 3481.18 m3/s, while the [0] configuration shows the highest RMSE (3634.35 m3/s). The performance gradient follows a consistent pattern across both areas, where increased temporal lag information progressively improves model accuracy.
Regarding streamflow input configurations, the comparison between temporal lags [1] and [1, 2] shows that incorporating additional historical streamflow information ([1, 2]) yields only marginal improvements, with RMSE decreasing modestly from 1245.33 to 1230.19 m3/s in the Upper WDDR area and from 3577.72 to 3577.31 m3/s in the Upper SXR area. This suggests that the contribution of additional streamflow lag information is relatively limited compared to the substantial performance gains observed with precipitation temporal lags.

3.3. Testing Performance Evaluation

The predictive performance metrics of the six models (i.e., GRU-N, GRU-S, GRU-G, LSTM-N, LSTM-S, and LSTM-G) are presented in Table 4, and the Taylor diagram are presented in Figure A1. The results consistently demonstrate that models utilizing a Gamma distribution output (GRU-G, LSTM-G) generally outperform their counterparts that use Gaussian or Student’s t-distributions. In the Upper WDDR area, LSTM-G demonstrates exceptional performance, achieving the lowest RMSE (1016.54 m3/s) and MAE (643.99 m3/s), and the highest NSE (0.89). For probabilistic forecasts, it also proves highly reliable, with a PICP of 93.15% and the lowest CRPS (473.26 m3/s). For the Upper SXR area, the superiority is shared between the Gamma-based models. The GRU-G model achieves the lowest MAE (2222.94 m3/s) and CRPS (1654.24 m3/s). Meanwhile, the LSTM-G model achieves the best RMSE (4047.15 m3/s) and the highest reliability in its probabilistic forecast, with an impressive PICP of 96.54%. The differences between GRU and LSTM architectures are relatively small and show no consistent patterns. In the Upper WDDR area, LSTM-G slightly outperforms GRU-G, while in the Upper SXR area, the differences are marginal. This suggests that RNN architecture choice has minimal impact compared to output distribution selection.

4. Discussion

The predictive performance of the six models (i.e., GRU-N, GRU-S, GRU-G, LSTM-N, LSTM-S, and LSTM-G) is discussed in this section. First, Section 4.1 discusses the deterministic prediction performance of models with different distribution outputs and different RNN structures across various forecast horizons. Then, Section 4.2 provides a comparative analysis of their probabilistic prediction performance. Finally, Section 4.3 demonstrates the overall predictive performance of the models across different forecast horizons.

4.1. Deterministic Prediction Performance of Different Models

The deterministic predictive performance of six models (i.e., GRU-N, GRU-S, GRU-G, LSTM-N, LSTM-S, and LSTM-G) across 18 forecast horizons (10-day periods) for two study areas is presented in Table 5 and Figure 6, with the results for RMSE shown as a representative example. It is evident that the forecasting accuracy is generally observed to decline as the forecast horizon increases, regardless of model architecture or output distribution. Among the three distributions, the Gamma distribution consistently results in lower RMSE values, particularly at longer lead times. This suggests a better capacity of the Gamma-based models to capture the positively skewed or heteroscedastic characteristics often found in hydrological data. In the Upper WDDR area, the LSTM-G model achieves the lowest RMSE across most forecast horizons, with values ranging from 859.8 m3/s at the 1st forecast period to 1096.7 m3/s at the 18th. A similar pattern is observed in the Upper SXR area, where LSTM-G yields the minimum RMSE of 3687.7 m3/s at the 1st period and 3998.1 m3/s at the 18th.
Differences between LSTM and GRU structures are also observed but are found to be less consistent and less influential than those resulting from output distribution selection. In the Upper WDDR area, GRU-N and GRU-S outperform their LSTM counterparts, while the advantage shifts to LSTM only when the Gamma distribution is employed. In the Upper SXR area, the superiority of LSTM-G is observed only at shorter forecast horizons, with GRU-G performing better as the forecast horizon increases. Across all configurations, no consistent advantage is associated with either architecture, suggesting that model structure plays a secondary role relative to the output distribution in determining predictive performance.
Notably, the magnitude of performance gains brought by the Gamma distribution varies across regions. In the Upper WDDR area, the use of Gamma distribution yields substantial improvements compared to Normal and Student’s t assumptions, particularly when paired with LSTM. In contrast, in the Upper SXR area, although Gamma-based models still outperform others, the improvement is relatively marginal. For instance, the reduction in RMSE from LSTM-N to LSTM-G at the 18th horizon is only modest (from 4203.0 m3/s to 4070.5 m3/s), whereas the gain is more pronounced in the Upper WDDR area (from 1559.5 m3/s to 1096.7 m3/s). These findings demonstrate that the benefit of applying flexible, non-Gaussian output distributions such as Gamma is region-dependent.
A primary reason for this discrepancy likely lies in the different statistical characteristics of the streamflow in the two basins, as detailed in Table 1. The streamflow in the Upper WDDR area exhibits a significantly higher degree of positive skewness (1.35) and kurtosis (1.23) compared to the Upper SXR area (skewness of 1.21, kurtosis of 0.98). This indicates that the Upper WDDR streamflow distribution has a longer right tail and is more “peaked,” deviating more substantially from a Normal distribution. Consequently, the flexible, skewed shape of the Gamma distribution provides a much better fit for the more asymmetric data of the Upper WDDR area, leading to more significant performance gains. Conversely, the streamflow in the Upper SXR area, while still positively skewed, is statistically closer to a symmetrical distribution. Therefore, the advantage of the Gamma distribution over simpler assumptions like the Gaussian distribution is less pronounced, resulting in only marginal improvements.
In summary, the results highlight the critical role of output distribution in deterministic streamflow forecasting, with the Gamma distribution consistently offering performance benefits, though to varying extents across regions. While model structure has some impact, it exerts less influence than distributional assumptions.

4.2. Probabilistic Prediction Performance of Different Models

The probabilistic predictive performance (i.e., CRPS, PICP, and MPIW) of six models (i.e., GRU-N, GRU-S, GRU-G, LSTM-N, LSTM-S, and LSTM-G) across 18 forecast horizons (10-day periods) for two study areas is presented in Figure 7. Similar to the deterministic results, probabilistic forecast accuracy generally declines with increasing lead time, as reflected by rising CRPS values across all model configurations. Among the output distributions, the Gamma distribution again provides the most substantial improvements in forecast skill. This consistency between deterministic and probabilistic evaluations reinforces the robustness of the Gamma assumption in capturing the inherent characteristics of streamflow data. In the Upper WDDR area, LSTM-G yields the lowest CRPS across nearly all horizons. In the Upper SXR area, LSTM-G yields the lowest CRPS during the first three horizons, while GRU-G shows superior accuracy at longer horizons. This pattern echoes deterministic results, where GRU-G also performed better in long-term forecasting under complex conditions.
Further analysis of the PICP and MPIW reveals deeper insights into the models’ reliability and sharpness (Figure 7c–f). As expected, the MPIW for all models generally increases with the forecast horizon, reflecting the growing uncertainty of longer-term predictions. However, the models based on the Gamma distribution (GRU-G and LSTM-G) consistently achieve higher PICP values, indicating that their prediction intervals more reliably encompass the observed values. Notably, these Gamma-based models manage to maintain relatively narrower interval widths (lower MPIW) for most lead times in the Upper WDDR area and for shorter lead times in the Upper SXR area. In contrast, while models using the Student’s t-distribution (GRU-S, LSTM-S) often produce the narrowest intervals (lowest MPIW), this comes at the cost of poor reliability, as their PICP values are frequently the lowest. This demonstrates their inability to effectively capture the true uncertainty of the streamflow.
Compared to output distribution selection, differences due to model structure are less consistent and generally less influential. While LSTM tends to perform better in the Upper WDDR area, GRU offers competitive or better results in the Upper SXR area when combined with the Gamma distribution. Regional differences are again observed, with CRPS values in the Upper SXR area consistently higher than those in the Upper WDDR area, reflecting greater predictive uncertainty. However, the relative benefit of using the Gamma distribution persists in both regions, though with different magnitudes.
In summary, probabilistic forecasting results confirm key findings from deterministic evaluations—particularly the consistent and robust advantage of using the Gamma distribution across regions and horizons. This advantage is not only reflected in superior accuracy (lower CRPS) but also in a better balance between forecast reliability (PICP) and sharpness (MPIW). However, nuanced differences are observed in the relative contributions of model structure, especially under probabilistic metrics. These findings suggest that while distributional assumptions remain the most critical component for improving streamflow forecast quality, model structure and regional hydrological characteristics jointly shape both deterministic and probabilistic forecasting performance.

4.3. Overall Predictive Performance

The predicted and observed streamflow at the first forecast horizon are shown in Figure 8. It is evident that all models produce narrow prediction intervals that closely follow the observations during non-flood seasons, when streamflow is low and uncertainty is limited. In contrast, flood seasons are characterized by increased variability, leading to wider forecast intervals and larger deviations, particularly around peak flows. The choice of output distribution also significantly affects forecast reliability. For example, models using the Normal distribution often generate overly wide intervals due to their poor fit to the skewed nature of streamflow, which is reflected in their generally higher MPIW values (Figure 7e,f). In contrast, Gamma-based models produce tighter and more stable intervals while maintaining high coverage probability, as evidenced by their superior balance between PICP and MPIW across most horizons (Figure 7c–f). While Gamma-based models generally provide better predictions in most cases, their advantage may not hold in every situation—for example, the GRU-G model performs worst in the Upper WDDR area. Nonetheless, their ability to constrain uncertainty remains evident, and such model-specific variability also underscores the importance of model fusion strategies for improving forecast robustness under diverse application scenarios [21,61,62,63,64].
The predicted and observed streamflow across 18 forecast periods for models using Gamma distribution in both the Upper WDDR and Upper SXR areas are presented in Figure 9. As the forecast horizon increases, streamflow uncertainty becomes more pronounced, and the width of the predictive intervals correspondingly expands, which is quantitatively confirmed by the rising MPIW values shown in Figure 7. This widening of intervals effectively reflects the growing uncertainty associated with longer lead times, particularly during high-flow periods. The use of the Gamma distribution enables the models to better capture the skewed nature of streamflow, resulting in predictive intervals that not only reflect the asymmetry of the data but also reliably encompass the observed hydrographs across seasons and forecast periods, which is validated by the high PICP values (Figure 7). This demonstrates the capability of using suitable output distribution to improve both the reliability and calibration in probabilistic streamflow prediction.

4.4. Limitations and Future Research Directions

Despite the promising performance of the proposed framework in the two selected study areas, there are still several limitations that need to be further studied. On one hand, given that different models exhibit varying performance under different scenarios, future work could explore model fusion or ensemble approaches. This is motivated by the finding that the LSTM-G model performs best at all lead times in the Upper WDDR area and at shorter lead times in the Upper SXR area, while the GRU-G model excels at longer lead times in the Upper SXR area. A model fusion method could leverage complementary model strengths to enhance overall forecast robustness.
On the other hand, the applicability of our modeling framework needs to be validated in basins with different hydrological regimes. The areas used in this study are humid, monsoon-influenced regions, for which precipitation is the primary input driver and the non-negative Gamma distribution is suitable. However, its effectiveness in other basins requires further investigation. For example, in semi-arid regions, streamflow is often intermittent, and evapotranspiration becomes a dominant influencing factor. Therefore, the model would need to incorporate evapotranspiration as an input, and the Gamma distribution, which cannot handle zero values, would likely need to be replaced with more suitable alternatives, such as Zero-Inflated or mixture distribution models. Similarly, in snowmelt-dominated regions, the framework would need to be adapted to include key input variables such as air temperature and snow water equivalent to capture the essential physics of snow accumulation and melt processes.

5. Conclusions

In this study, a DeepAR-based modeling framework is developed to generate probabilistic streamflow forecasts by integrating distribution selection with the DeepAR model. The framework is applied to two case studies to evaluate both deterministic and probabilistic forecasting performance. The influence of model structure and output distribution choice on prediction accuracy is also examined. The main conclusions are as follows.
(1) The proposed DeepAR-based modeling framework effectively identifies the most appropriate output distribution and input configuration, leading to superior deterministic and probabilistic forecasting performance in both case studies. By integrating distribution selection and input optimization, the framework ensures that the final models are well-calibrated to the characteristics of local streamflow dynamics.
(2) Multiple factors influence predictive performance, among which the selection of output distribution has the most significant impact. Notably, the advantage of the Gamma distribution is more pronounced in basins with higher skewness, providing a data-driven explanation for the observed regional performance differences. Meanwhile, differences between RNN model structures (i.e., GRU vs. LSTM) are relatively minor and less consistent.
(3) As forecast lead time increases and during flood seasons, streamflow uncertainty grows substantially. The developed probabilistic forecasting framework is capable of capturing this variation in uncertainty, and models using the Gamma distribution demonstrate superior performance by better representing the skewed nature of streamflow. This contributes to more reliable and better-calibrated forecasts across both typical and high-variability conditions.

Author Contributions

Conceptualization, S.X. and J.W.; methodology, S.X.; software, S.X. and K.S.; validation, S.X., D.W. and B.J.; formal analysis, S.X., C.Y. and H.C.; investigation, D.W. and C.Y.; resources, S.X. and K.S.; data curation, J.W. and B.J.; writing—original draft preparation, S.X., J.W. and B.J.; writing—review and editing, S.X., D.W., K.S. and H.C.; visualization, S.X., C.Y. and H.C.; supervision, D.W. and H.C.; project administration, S.X.; funding acquisition, D.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by The National Key Research and Development Programof China (grant numbers: 2023YFC3206002, 2024YFE0100100), the Natural Science Foundation of Hubei Province (grant numbers: 2023AFB039, 2022CFD027), the National Natural Science Foundation of China (grant number: U2340211), the Key Project of Chinese Water Resources Ministry (grant number: SKS-2022120), fundamental research project for central public welfare research institutes (CKSF20241021/SZ). The authors declare that this study received funding from China Yangtze Power Co., Ltd. (contract no. Z242302057 and project no. 2423020055). The funder was not involved in the study design, collection, analysis, interpretation of data, the writing of this article or the decision to submit it for publication. Shuai Xie is supported by a program of China Scholarship Council (No. 202303340001) during his visit to the University of Regina, where the research is conducted.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

Various Python 3.6.1 open-source frameworks were used in this study. We express our gratitude to all contributors. We also give special thanks to the anonymous reviewers and editors for their constructive comments.

Conflicts of Interest

Author Jin Wang, Keyan Shen, Benjun Jia and Hui Cao were employed by the company China Yangtze Power Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MLSPmid–long-term streamflow prediction
AIartificial intelligence
SVRsupport vector regression
ANNartificial neural network
LSTMlong short-term memory network
GRUgated recurrent unit neural network
RNNrecurrent neural network
WDDRWudongde Reservoir
SXRSanxia Reservoir
RMSEroot mean square error
CRPScontinuous ranked probability score

Appendix A

Table A1. The hyperparameters.
Table A1. The hyperparameters.
HyperparametersValue
Number of layers2
Number of cells40
Dropout rate0.2
Time featuresTrue
Batch size32
Epochs200
Learning rate0.001
Figure A1. Taylor diagram of six models: (a) Upper WDDR area; (b) Upper SXR area.
Figure A1. Taylor diagram of six models: (a) Upper WDDR area; (b) Upper SXR area.
Water 17 02506 g0a1

References

  1. Nguyen, Q.H.; Tran, V.N. Temporal Changes in Water and Sediment Discharges: Impacts of Climate Change and Human Activities in the Red River Basin (1958–2021) with Projections up to 2100. Water 2024, 16, 1155. [Google Scholar] [CrossRef]
  2. Jia, L.; Niu, Z.; Zhang, R.; Ma, Y. Sensitivity of Runoff to Climatic Factors and the Attribution of Runoff Variation in the Upper Shule River, North-West China. Water 2024, 16, 1272. [Google Scholar] [CrossRef]
  3. Xu, H.; Liu, L.; Wang, Y.; Wang, S.; Hao, Y.; Ma, J.; Jiang, T. Assessment of climate change impact and difference on the river runoff in four basins in China under 1.5 and 2.0 °C global warming. Hydrol. Earth Syst. Sci. 2019, 23, 4219–4231. [Google Scholar] [CrossRef]
  4. Zou, L.; Zhou, T. Near future (2016-40) summer precipitation changes over China as projected by a regional climate model (RCM) under the RCP8, 5 emissions scenario: Comparison between RCM downscaling and the driving GCM. Adv. Atmos. Sci. 2013, 30, 806–818. [Google Scholar] [CrossRef]
  5. Shukla, P.; Skeg, J.; Buendia, E.C.; Masson-Delmotte, V.; Pörtner, H.-O.; Roberts, D.; Zhai, P.; Slade, R.; Connors, S.; van Diemen, S. Climate Change and Land: An IPCC Special Report on Climate Change, Desertification, Land Degradation, Sustainable Land Management, Food Security, and Greenhouse Gas Fluxes in Terrestrial Ecosystems. 2019. Available online: https://www.ipcc.ch/srccl/ (accessed on 4 August 2025).
  6. Haj-Amor, Z.; Acharjee, T.K.; Dhaouadi, L.; Bouri, S. Impacts of climate change on irrigation water requirement of date palms under future salinity trend in coastal aquifer of Tunisian oasis. Agric. Water Manag. 2020, 228, 105843. [Google Scholar] [CrossRef]
  7. Piao, S.; Ciais, P.; Huang, Y.; Shen, Z.; Peng, S.; Li, J.; Zhou, L.; Liu, H.; Ma, Y.; Ding, Y. The impacts of climate change on water resources and agriculture in China. Nature 2010, 467, 43–51. [Google Scholar] [CrossRef] [PubMed]
  8. Larraz, B.; García-Rubio, N.; Gámez, M.; Sauvage, S.; Cakir, R.; Raimonet, M.; Pérez, J.M.S. Socio-Economic Indicators for Water Management in the South-West Europe Territory: Sectorial Water Productivity and Intensity in Employment. Water 2024, 16, 959. [Google Scholar] [CrossRef]
  9. Gong, G.; Wang, L.; Condon, L.; Shearman, A.; Lall, U. A simple framework for incorporating seasonal streamflow forecasts into existing water resource management practices 1. JAWRA J. Am. Water Resour. Assoc. 2010, 46, 574–585. [Google Scholar] [CrossRef]
  10. Sunday, R.; Masih, I.; Werner, M.; van der Zaag, P. Streamflow forecasting for operational water management in the Incomati River Basin, Southern Africa. Phys. Chem. Earth Parts A/B/C 2014, 72, 1–12. [Google Scholar] [CrossRef]
  11. Bărbulescu, A.; Zhen, L. Forecasting the River Water Discharge by Artificial Intelligence Methods. Water 2024, 16, 1248. [Google Scholar] [CrossRef]
  12. Chu, H.; Wei, J.; Wu, W. Streamflow prediction using LASSO-FCM-DBN approach based on hydro-meteorological condition classification. J. Hydrol. 2020, 580, 124253. [Google Scholar] [CrossRef]
  13. Feng, Z.-K.; Niu, W.-J.; Tang, Z.-Y.; Jiang, Z.-Q.; Xu, Y.; Liu, Y.; Zhang, H.-R. Monthly runoff time series prediction by variational mode decomposition and support vector machine based on quantum-behaved particle swarm optimization. J. Hydrol. 2020, 583, 124627. [Google Scholar] [CrossRef]
  14. Xie, S.; Xiang, Z.; Wang, Y.; Wu, B.; Shen, K.; Wang, J. An Index Used to Evaluate the Applicability of Mid-to-Long-Term Runoff Prediction in a Basin Based on Mutual Information. Water 2024, 16, 1619. [Google Scholar] [CrossRef]
  15. Xie, S.; Huang, Y.; Li, T.; Liu, Z.; Wang, J. Mid-long term runoff prediction based on a Lasso and SVR hybrid method. J. Basic Sci. Eng. 2018, 26, 709–722. [Google Scholar]
  16. Shamir, E. The value and skill of seasonal forecasts for water resources management in the Upper Santa Cruz River basin, southern Arizona. J. Arid. Environ. 2017, 137, 35–45. [Google Scholar] [CrossRef]
  17. Zhao, H.; Li, H.; Xuan, Y.; Bao, S.; Cidan, Y.; Liu, Y.; Li, C.; Yao, M. Investigating the critical influencing factors of snowmelt runoff and development of a mid-long term snowmelt runoff forecasting. J. Geogr. Sci. 2023, 33, 1313–1333. [Google Scholar] [CrossRef]
  18. Nguyen, D.H.; Elshorbagy, A.; Khaliq, M.N.; Shen, C.; Akhtar, M.K.; Moghairib, M.; Unduche, F.; Razavi, S.; Lamontagne, P. Advancing Sub-Seasonal to Seasonal Streamflow Forecasting in Canada: A Review of Conventional and Emerging Approaches for Operational Applications. Results Eng. 2025, 27, 106345. [Google Scholar] [CrossRef]
  19. He, C.; Chen, F.; Long, A.; Qian, Y.; Tang, H. Improving the precision of monthly runoff prediction using the combined non-stationary methods in an oasis irrigation area. Agric. Water Manag. 2023, 279, 108161. [Google Scholar] [CrossRef]
  20. Samsudin, R.; Saad, P.; Shabri, A. River flow time series using least squares support vector machines. Hydrol. Earth Syst. Sci. 2011, 15, 1835–1852. [Google Scholar] [CrossRef]
  21. Bennett, J.C.; Wang, Q.; Li, M.; Robertson, D.E.; Schepen, A. Reliable long-range ensemble streamflow forecasts: Combining calibrated climate forecasts with a conceptual runoff model and a staged error model. Water Resour. Res. 2016, 52, 8238–8259. [Google Scholar] [CrossRef]
  22. Crochemore, L.; Ramos, M.-H.; Pappenberger, F. Bias correcting precipitation forecasts to improve the skill of seasonal streamflow forecasts. Hydrol. Earth Syst. Sci. 2016, 20, 3601–3618. [Google Scholar] [CrossRef]
  23. Le, X.-H.; Ho, H.V.; Lee, G.; Jung, S. Application of long short-term memory (LSTM) neural network for flood forecasting. Water 2019, 11, 1387. [Google Scholar] [CrossRef]
  24. Choi, J.; Won, J.; Jang, S.; Kim, S. Learning Enhancement Method of Long Short-Term Memory Network and Its Applicability in Hydrological Time Series Prediction. Water 2022, 14, 2910. [Google Scholar] [CrossRef]
  25. Reichstein, M.; Camps-Valls, G.; Stevens, B.; Jung, M.; Denzler, J.; Carvalhais, N. Deep learning and process understanding for data-driven Earth system science. Nature. 2019, 566, 195–204. [Google Scholar] [CrossRef] [PubMed]
  26. Yaseen, Z.M.; El-Shafie, A.; Jaafar, O.; Afan, H.A.; Sayl, K.N. Artificial intelligence based models for stream-flow forecasting: 2000–2015. J. Hydrol. 2015, 530, 829–844. [Google Scholar] [CrossRef]
  27. Liu, Y.; Yin, Z.; Zhang, Y.; Wang, Q. Mid and long-term hydrological classification forecasting model based on KDE-BDA and its application research. IOP Conf. Ser. Earth Environ. Sci. 2019, 330, 032010. [Google Scholar] [CrossRef]
  28. Kratzert, F.; Klotz, D.; Shalev, G.; Klambauer, G.; Hochreiter, S.; Nearing, G. Towards learning universal, regional, and local hydrological behaviors via machine learning applied to large-sample datasets. Hydrol. Earth Syst. Sci. 2019, 23, 5089–5110. [Google Scholar] [CrossRef]
  29. Asefa, T.; Kemblowski, M.; Mckee, M.; Khalil, A. Multi-time scale stream flow predictions: The support vector machines approach. J. Hydrol. 2006, 318, 7–16. [Google Scholar] [CrossRef]
  30. Kalteh, A.M. Monthly river flow forecasting using artificial neural network and support vector regression models coupled with wavelet transform. Comput. Geosci. 2013, 54, 1–8. [Google Scholar] [CrossRef]
  31. Maity, R.; Bhagwat, P.P.; Bhatnagar, A. Potential of support vector regression for prediction of monthly streamflow using endogenous property. Hydrol. Process. 2010, 24, 917–923. [Google Scholar] [CrossRef]
  32. Noori, R.; Karbassi, A.R.; Moghaddamnia, A.; Han, D.; Zokaei-Ashtiani, M.H.; Farokhnia, A.; Gousheh, M.G. Assessment of input variables determination on the SVM model performance using PCA, Gamma test, and forward selection techniques for monthly stream flow prediction. J. Hydrol. 2011, 401, 177–189. [Google Scholar] [CrossRef]
  33. Lin, J.; Cheng, C.; Chau, K. Using support vector machines for long-term discharge prediction. Hydrol. Sci. J. 2006, 51, 599–612. [Google Scholar] [CrossRef]
  34. Tao, H.; Abba, S.I.; Al-Areeq, A.M.; Tangang, F.; Samantaray, S.; Sahoo, A.; Siqueira, H.V.; Maroufpoor, S.; Demir, V.; Dhanraj Bokde, N.; et al. Hybridized artificial intelligence models with nature-inspired algorithms for river flow modeling: A comprehensive review, assessment, and possible future research directions. Eng. Appl. Artif. Intell. 2024, 129, 107559. [Google Scholar] [CrossRef]
  35. Kelly, R.A.; Jakeman, A.J.; Barreteau, O.; Borsuk, M.E.; ElSawah, S.; Hamilton, S.H.; Henriksen, H.J.; Kuikka, S.; Maier, H.R.; Rizzoli, A.E. Selecting among five common modelling approaches for integrated environmental assessment and management. Environ. Model. Softw. 2013, 47, 159–181. [Google Scholar] [CrossRef]
  36. Mount, N.J.; Maier, H.R.; Toth, E.; Elshorbagy, A.; Solomatine, D.; Chang, F.-J.; Abrahart, R. Data-driven modelling approaches for socio-hydrology: Opportunities and challenges within the Panta Rhei Science Plan. Hydrol. Sci. J. 2016, 61, 1192–1208. [Google Scholar] [CrossRef]
  37. Roy, A.; Kasiviswanathan, K.; Patidar, S.; Adeloye, A.J.; Soundharajan, B.S.; Ojha, C.S.P. A novel physics-aware machine learning-based dynamic error correction model for improving streamflow forecast accuracy. Water Resour. Res. 2023, 59, e2022WR033318. [Google Scholar] [CrossRef]
  38. Sujay, R.N.; Deka, P.C. Support vector machine applications in the field of hydrology: A review. Appl. Soft. Comput. J. 2014, 19, 372–386. [Google Scholar]
  39. Liang, Z.; Li, Y.; Hu, Y.; Li, B.; Wang, J. A data-driven SVR model for long-term runoff prediction and uncertainty analysis based on the Bayesian framework. Theor. Appl. Climatol. 2018, 133, 137–149. [Google Scholar] [CrossRef]
  40. Wang, Q.; Zhao, T.; Yang, Q.; Robertson, D. A Seasonally Coherent Calibration (SCC) Model for Postprocessing Numerical Weather Predictions. Mon. Weather. Rev. 2019, 147, 3633–3647. [Google Scholar] [CrossRef]
  41. May, R.; Dandy, G.; Maier, H. Review of Input Variable Selection Methods for Artificial Neural Networks. InTech 2011, 10, 19–45. [Google Scholar]
  42. Mo, R.; Xu, B.; Zhong, P.-A.; Zhu, F.; Huang, X.; Liu, W.; Xu, S.; Wang, G.; Zhang, J. Dynamic long-term streamflow probabilistic forecasting model for a multisite system considering real-time forecast updating through spatio-temporal dependent error correction. J. Hydrol. 2021, 601, 126666. [Google Scholar] [CrossRef]
  43. Qian, X.; Wang, B.; Chen, J.; Fan, Y.; Mo, R.; Xu, C.; Liu, W.; Liu, J.; Zhong, P. An explainable ensemble deep learning model for long-term streamflow forecasting under multiple uncertainties. J. Hydrol. 2025, 662, 133968. [Google Scholar] [CrossRef]
  44. Mo, R.; Xu, B.; Zhong, P.-A.; Dong, Y.; Wang, H.; Yue, H.; Zhu, J.; Wang, H.; Wang, G.; Zhang, J. Long-term probabilistic streamflow forecast model with “inputs–structure–parameters” hierarchical optimization framework. J. Hydrol. 2023, 622, 129736. [Google Scholar] [CrossRef]
  45. Salinas, D.; Flunkert, V.; Gasthaus, J.; Januschowski, T. DeepAR: Probabilistic forecasting with autoregressive recurrent networks. Int. J. Forecast. 2020, 36, 1181–1191. [Google Scholar] [CrossRef]
  46. Salinas, D.; Bohlke-Schneider, M.; Callot, L.; Medico, R.; Gasthaus, J. High-Dimensional Multivariate Forecasting with Low-Rank Gaussian Copula Processes. arXiv 2019, arXiv:1910.03002. [Google Scholar]
  47. Schaduangrat, N.; Anuwongcharoen, N.; Charoenkwan, P.; Shoombuatong, W. DeepAR: A novel deep learning-based hybrid framework for the interpretable prediction of androgen receptor antagonists. J. Cheminform. 2023, 15, 50. [Google Scholar] [CrossRef]
  48. Li, J.; Chen, W.; Zhou, Z.; Yang, J.; Zeng, D. DeepAR-Attention probabilistic prediction for stock price series. Neural Comput. Appl. 2024, 36, 15389–15406. [Google Scholar] [CrossRef]
  49. Liao, Y.; Liang, C. A Temperature Time Series Forecasting Model Based on DeepAR. In Proceedings of the 2021 7th International Conference on Computer and Communications (ICCC), Chengdu, China, 10–13 December 2021. [Google Scholar]
  50. Chow, V. Applied Hydrology; McGraw-Hill: Columbus, OH, USA, 1971. [Google Scholar]
  51. England Jr, J.F.; Cohn, T.A.; Faber, B.A.; Stedinger, J.R.; Thomas, W.O., Jr.; Veilleux, A.G.; Kiang, J.E.; Mason, R.R., Jr. Guidelines for Determining Flood Flow Frequency—Bulletin 17C; US Geological Survey: Reston, VA, USA, 2018.
  52. Wu, W.; Dandy, G.C.; Maier, H.R. Protocol for developing ANN models and its application to the assessment of the quality of the ANN model development process in drinking water quality modelling. Environ. Model. Softw. 2014, 54, 108–127. [Google Scholar] [CrossRef]
  53. Xie, S.; Wu, W.; Mooser, S.; Wang, Q.; Nathan, R.; Huang, Y. Artificial neural network based hybrid modeling approach for flood inundation modeling. J. Hydrol. 2021, 592, 125605. [Google Scholar] [CrossRef]
  54. Maier, H.R.; Jain, A.; Dandy, G.C.; Sudheer, K.P. Methods used for the development of neural networks for the prediction of water resource variables in river systems: Current status and future directions. Environ. Model. Softw. 2010, 25, 891–909. [Google Scholar] [CrossRef]
  55. Demir, V. Evaluation of Solar Radiation Prediction Models Using AI: A Performance Comparison in the High-Potential Region of Konya, Türkiye. Atmosphere 2025, 16, 398. [Google Scholar] [CrossRef]
  56. Kingma, D.; Adam, J.B. A Method for Stochastic Optimization. Comput. Sci. 2014, 5, 6. [Google Scholar]
  57. De Faria, V.; De Queiroz, A.; Lima, L.; Lima, J.; Da Silva, B. An assessment of multi-layer perceptron networks for streamflow forecasting in large-scale interconnected hydrosystems. Int. J. Environ. Sci. Technol. 2022, 19, 5819–5838. [Google Scholar] [CrossRef]
  58. Humphrey, G.B.; Maier, H.R.; Wu, W.; Mount, N.J.; Dandy, G.C.; Abrahart, R.J.; Dawson, C.W. Improved validation framework and R-package for artificial neural network models. Environ. Model. Softw. 2017, 92, 82–106. [Google Scholar] [CrossRef]
  59. Gneiting, T.; Raftery, A.E.; Westveld, A.H., III; Goldman, T. Calibrated probabilistic forecasting using ensemble model output statistics and minimum CRPS estimation. Mon. Weather. Rev. 2005, 133, 1098–1118. [Google Scholar] [CrossRef]
  60. Renard, B.; Kavetski, D.; Kuczera, G.; Thyer, M.; Franks, S.W. Understanding predictive uncertainty in hydrologic modeling: The challenge of identifying input and structural errors. Water Resour. Res. 2010, 46, W05521. [Google Scholar] [CrossRef]
  61. See, L.; Abrahartand, R.J. Multi-model data fusion for hydrological forecasting. Comput. GeoSci. 2001, 27, 987–994. [Google Scholar] [CrossRef]
  62. Azmi, M.; Araghinejad, S.; Kholghi, M. Multi model data fusion for hydrological forecasting using k-nearest neighbour method. Iran. J. Sci. Technol. 2010, 34, 81. [Google Scholar]
  63. Wang, Q.; Schepen, A.; Robertson, D.E. Merging seasonal rainfall forecasts from multiple statistical models through Bayesian model averaging. J. Clim. 2012, 25, 5524–5537. [Google Scholar] [CrossRef]
  64. Schepen, A.; Wang, Q.; Everingham, Y. Calibration, bridging, and merging to improve GCM seasonal temperature forecasts in Australia. Mon. Weather. Rev. 2016, 144, 2421–2441. [Google Scholar] [CrossRef]
Figure 1. DeepAR-based modeling framework.
Figure 1. DeepAR-based modeling framework.
Water 17 02506 g001
Figure 2. Study area.
Figure 2. Study area.
Water 17 02506 g002
Figure 3. Observed 10-day streamflow, linear trend, and data splitting for two study areas: (a) Upper WDDR area; (b) Upper SXR area.
Figure 3. Observed 10-day streamflow, linear trend, and data splitting for two study areas: (a) Upper WDDR area; (b) Upper SXR area.
Water 17 02506 g003
Figure 4. The workflow diagram.
Figure 4. The workflow diagram.
Water 17 02506 g004
Figure 5. Distribution fitting figure: (a) Upper WDDR area; (b) Upper SXR area.
Figure 5. Distribution fitting figure: (a) Upper WDDR area; (b) Upper SXR area.
Water 17 02506 g005
Figure 6. RMSE of six models across various forecast horizons: (a) Upper WDDR area; (b) Upper SXR area.
Figure 6. RMSE of six models across various forecast horizons: (a) Upper WDDR area; (b) Upper SXR area.
Water 17 02506 g006
Figure 7. Probabilistic predictive performance metrics of six models across various forecast horizons. Subfigures (a,c,e) are CRPSs, PICPs and MPIWs of six models in the Upper WDDR, and Subfigures (b,d,f) are CRPSs, PICPs and MPIWs of six models in the Upper SXR.
Figure 7. Probabilistic predictive performance metrics of six models across various forecast horizons. Subfigures (a,c,e) are CRPSs, PICPs and MPIWs of six models in the Upper WDDR, and Subfigures (b,d,f) are CRPSs, PICPs and MPIWs of six models in the Upper SXR.
Water 17 02506 g007aWater 17 02506 g007b
Figure 8. Predicted streamflow by different models at the 1st forecast lead time: (a) Upper WDDR area; (b) Upper SXR area.
Figure 8. Predicted streamflow by different models at the 1st forecast lead time: (a) Upper WDDR area; (b) Upper SXR area.
Water 17 02506 g008
Figure 9. Predicted streamflow at different forecast lead times: (a) Upper WDDR area; (b) Upper SXR area.
Figure 9. Predicted streamflow at different forecast lead times: (a) Upper WDDR area; (b) Upper SXR area.
Water 17 02506 g009
Table 1. Available data.
Table 1. Available data.
Study AreaUpper WDDR Area Upper SXR Area
VariableNaturalized StreamflowAreal Mean PrecipitationNaturalized StreamflowAreal Mean Precipitation
Temporal coverage and scaleJanuary 1980 to September 2022
10-day time scale
Min557 m3/s0.00 mm2910 m3/s0.10 mm
Max18,600 m3/s90.83 mm56,500 m3/s91.41 mm
Average3819.97 m3/s17.59 mm13,432.22 m3/s22.69 mm
Standard deviation3247.06 m3/s18.98 mm10,016.31 m3/s20.36 mm
Coefficient of variation0.851.070.750.84
Skewness1.350.221.21−0.26
Kurtosis1.231.080.980.89
Table 2. AIC and BIC values of three distributions for two study areas.
Table 2. AIC and BIC values of three distributions for two study areas.
Study AreaDistributionAICBIC
Upper WDDR areaNormal25,335.3525,345.73
Student’s t25,223.8025,239.38
Gamma24,044.3624,059.95
Upper SXR areaNormal28,283.9528,294.34
Student’s t28,285.9528,301.54
Gamma27,308.5527,324.14
Note: The bold values indicate the minimum AIC/BIC.
Table 3. Average RMSE on the validation dataset of different models.
Table 3. Average RMSE on the validation dataset of different models.
Input ConfigurationAverage RMSE on Validation Dataset of Different Models for Upper WDDR Area (m3/s)Average RMSE on Validation Dataset of Different Models for Upper SXR Area (m3/s)
Lags of precipitation01277.073634.55
0, 11237.003616.81
0, 1, 21199.213481.18
Lags of streamflow11245.333577.72
1, 21230.193577.31
Table 4. Predictive performance on the testing dataset of six models with different RNN structures and different probability distribution outputs.
Table 4. Predictive performance on the testing dataset of six models with different RNN structures and different probability distribution outputs.
Study AreaEvaluation MetricsModel
GRU-NGRU-SGRU-GLSTM-NLSTM-SLSTM-G
Upper WDDR area RMSE (m3/s)1356.741282.091098.981407.771331.081016.54
MAE (m3/s)817.92775.23699.40844.78816.79643.99
NSE0.820.840.880.810.830.89
CRPS (m3/s)608.89578.95517.54637.95620.08473.26
PICP (%)90.0884.8192.9387.0976.8493.15
MPIW (m3/s)3980.523585.953330.883778.343228.713485.17
Upper SXR areaRMSE (m3/s)4217.124143.334057.334091.224296.164047.15
MAE (m3/s)2346.572289.882222.942370.062487.382312.61
NSE0.860.870.870.870.860.87
CRPS (m3/s)1776.921716.521654.241771.431870.941717.93
PICP (%)91.7791.4596.4693.1187.2796.54
MPIW (m3/s)10,015.7310,669.2011,048.1511,218.449156.8711,464.91
Note: The bold values indicate the best predictive performance in terms of the specific evaluation metric.
Table 5. RMSE of six models across various forecast horizons.
Table 5. RMSE of six models across various forecast horizons.
Forecast Horizon (10-Day Periods)Upper WDDR AreaUpper SXR Area
GRU-NGRU-SGRU-GLSTM-NLSTM-SLSTM-GGRU-NGRU-SGRU-GLSTM-NLSTM-SLSTM-G
1876.2878.3940.6892.9881.5859.83897.64024.84126.83863.24022.03687.2
21035.71007.61017.91083.11054.9958.94031.04091.64104.63880.44254.23841.5
31124.91085.5995.71164.51129.9913.64167.44109.54221.64045.04260.83938.4
41201.31157.91032.11235.91190.4924.83982.73978.03947.04011.74176.24011.2
51239.91159.21006.41264.51178.6902.74080.14002.44118.14106.44272.73977.0
61270.41215.31020.51312.11219.7912.34098.84148.24043.14024.74213.84068.3
71330.81247.11082.21351.11263.2986.24145.44102.04044.44104.74285.84020.6
81367.01303.71119.51406.91350.61051.44190.14031.13977.64065.04249.43936.1
91408.01340.61102.51467.41390.41093.24297.14161.74030.84063.24283.74023.0
101453.81350.41119.31492.21430.31089.74312.44183.74002.94155.74348.24117.6
111451.81362.31116.01518.11441.81068.94343.34163.74020.14116.14329.24064.6
121487.01394.61153.51532.01458.41073.34374.34184.04009.14158.24295.44095.5
131483.91394.11152.81545.51469.71048.04328.64229.54107.54140.84311.14103.9
141485.51394.91167.41559.41449.51055.84358.94277.44034.34176.34406.54258.0
151484.51390.31174.91574.51454.61071.84339.04217.54096.74170.84371.74235.7
161494.81391.51164.61565.61456.01069.74331.04219.14042.14119.74391.14180.6
171507.11407.01177.81574.51474.51068.04336.24245.04098.14216.84428.74178.2
181503.51430.71192.91559.51468.41096.74251.04194.83998.14203.04411.24070.5
Note: The bold values indicate the miminum RMSE value among six models.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xie, S.; Wang, D.; Wang, J.; Yang, C.; Shen, K.; Jia, B.; Cao, H. A DeepAR-Based Modeling Framework for Probabilistic Mid–Long-Term Streamflow Prediction. Water 2025, 17, 2506. https://doi.org/10.3390/w17172506

AMA Style

Xie S, Wang D, Wang J, Yang C, Shen K, Jia B, Cao H. A DeepAR-Based Modeling Framework for Probabilistic Mid–Long-Term Streamflow Prediction. Water. 2025; 17(17):2506. https://doi.org/10.3390/w17172506

Chicago/Turabian Style

Xie, Shuai, Dong Wang, Jin Wang, Chunhua Yang, Keyan Shen, Benjun Jia, and Hui Cao. 2025. "A DeepAR-Based Modeling Framework for Probabilistic Mid–Long-Term Streamflow Prediction" Water 17, no. 17: 2506. https://doi.org/10.3390/w17172506

APA Style

Xie, S., Wang, D., Wang, J., Yang, C., Shen, K., Jia, B., & Cao, H. (2025). A DeepAR-Based Modeling Framework for Probabilistic Mid–Long-Term Streamflow Prediction. Water, 17(17), 2506. https://doi.org/10.3390/w17172506

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop