Next Article in Journal
Effect of Effluent Recirculation on the Performance of an Anaerobic Baffled Reactor in Municipal Wastewater Treatment: A Modeling Approach
Previous Article in Journal
Looking Beyond Lithium for Breakthroughs in Magnesium-Ion Batteries as Sustainable Solutions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine-Learning-Based Ensemble Prediction of the Snow Water Equivalent in the Upper Yalong River Basin

Department of Water Resources, China Institute of Water Resources and Hydropower Research, Beijing 100038, China
*
Author to whom correspondence should be addressed.
Sustainability 2025, 17(9), 3779; https://doi.org/10.3390/su17093779
Submission received: 16 December 2024 / Revised: 4 April 2025 / Accepted: 16 April 2025 / Published: 22 April 2025

Abstract

:
The snow water equivalent (SWE) in high-altitude regions is crucial for water resource management and disaster risk reduction, yet accurate predictions remain challenging due to complex snowmelt processes, nonlinear meteorological factors, and time-lag effects. This study used snow remote sensing products from the Advanced Microwave Scanning Radiometer (AMSR) as the predictand for evaluating SWE predictions. It applied nine machine learning models—linear regression (LR), decision trees (DT), support vector regression (SVR), random forest (RF), artificial neural networks (ANNs), AdaBoost, XGBoost, gradient boosting decision trees (GBDT), and CatBoost. For each machine learning model, submodels were constructed to predict the SWE for the next 1 to 30 days. The 30 submodels of each machine learning model formed the prediction model for the snow water equivalent over the next 30 days. Through an accuracy evaluation and ensemble forecasting, the snow water equivalent prediction for the next 30 days in the Yalong River above the Ganzi Basin was finally achieved. The results showed that for all models, the average Nash–Sutcliffe Efficiency (NSE) rate was greater than 0.8, the average root mean square error (RMSE) was under 8 mm, and the average relative error (RE) was below 7% across three lead time periods (1–10, 11–20, and 21–30 days). The ensemble average model, combining ANNs, GBDT, and CatBoost, demonstrated superior accuracy, with NSE values exceeding 0.85 and RMSE values under 6 mm. A sensitivity analysis using the Shapley Additive Explanations (SHAP) model revealed that temperature variables (average, minimum, and maximum temperatures) were the most influential factors, while relative humidity (Rhu) significantly affected the SWE by reducing evaporation. These findings provide insights for improving SWE prediction accuracy and support water resource management in high-altitude regions.

1. Introduction

Snow, as an important component of the cryosphere, accounts for about 98% of seasonal snow, with the largest area reaching 4.7 × 107 km2 in the Northern Hemisphere [1,2,3]. It is highly sensitive to climate change and significantly impacts the climate, water cycle processes, and the structure and function of ecosystems [4,5,6]. Snowmelt water can alleviate seasonal water shortages, providing stable water resources during dry seasons. It is a crucial freshwater source for about one-sixth of the world’s population [7]. Spring snowmelt runoff typically accounts for 10–15% of the global annual runoff [8,9,10], playing an important role in global runoff. Accurate snow data can enhance the effective utilization of watershed water resources and can be used to assess the potential impact of climate change on hydrological processes in cold regions, providing important observational indicators for climate change monitoring and hydrological model construction [11,12]. Therefore, the acquisition and prediction of snow data are important for water resource management, ecosystem health, and early disaster warnings [13,14,15,16,17].
Currently, snow data acquisition mainly relies on ground-based observations, remote sensing inversion, and model simulations [18,19]. Ground-based observations can provide high-precision local data, including on the snow depth, density, and SWE. However, due to the sparse distribution of observation stations, it is difficult for ground-based observations to meet the needs of large-scale and fine-grained applications. Satellite remote sensing technology has distinct advantages in acquiring large-scale snow data, enabling snow data collection in data-scarce areas of cold regions. For instance, MODIS data from optical remote sensing and AMSR data from passive microwave remote sensing have been widely used for snow data acquisition [20,21,22,23,24]. Traditional optical remote sensing data are limited by cloud cover, resulting in significant errors. Although microwave remote sensing data can penetrate clouds, they are still influenced by surface features, and their transmission exhibits latency. Remote sensing data still face limitations in inversion accuracy and real-time acquisition, and there is a significant gap in future snow data predictions [25,26].
To further explore snow evolution mechanisms and predict future snow changes, many scholars have turned to physical or artificial intelligence models. These models use meteorological and surface data for the refined simulation and prediction of snow processes [27,28]. In current physical models, snowmelt mechanisms mainly include degree-day factors and energy balance models [29,30]. The degree-day factor model requires fewer data and can relatively simply simulate snowmelt runoff, although due to its simple mechanism, it is difficult to accurately predict future snow changes. Energy balance models, on the other hand, have complex parameters and high requirements for initial data, limiting their application in data-scarce high-altitude regions. For example, Bi et al. [31] used the upper Lancang River as the study area to compare the snowmelt runoff simulation processes under both the degree-day factor and energy balance methods, which partly reflected their simulation’s effectiveness for snow evolution. Grusson et al. [32] applied the SWAT model to simulate snow in the upper Shule River watershed. Zhao et al. [33] combined WRF model forecast data with the DHSVM model to achieve a 24 h snowmelt runoff forecast. However, the WRF model relies on high-quality initial and boundary conditions, which requires high data accuracy, and there are limitations in simulating long-term forecasts and complex meteorological phenomena. Therefore, due to the limitations of meteorological and hydrological data, there is still room for further research in predicting long-term snowmelt scenarios.
The limitations of the current methods have led to the application of machine learning in snowmelt inversion and snowmelt runoff simulations. Compared with traditional hydrological models, machine learning models do not require consideration of the complex and variable physical processes within the watershed. By using past meteorological datasets to explore the complex nonlinear relationships between input factors and target values, more accurate simulation results can be obtained. With the development of remote sensing and machine learning technologies, data-driven methods based on machine learning have gradually become some of the mainstream approaches for snow retrieval [34,35]. For example, Steel et al. [11] applied traditional machine learning algorithms such as random forests to capture the complex nonlinear relationship between environmental factors and the SWE for snowmelt inversion using remote sensing data. Moradizadeh et al. [36] used machine learning algorithms such as SVM and CNN models for the spatial downscaling of snow data in high-altitude areas. Wang et al. [37] used ANN and RF models to simulate runoff and capture snow change trends in the Xiying River basin in western Qilian. In addition to the machine learning algorithms commonly used in snowmelt assessments, such as ANN, random forest, and SVM models, existing models such as linear regression [38], decision trees [39], XGBoost [40], GBDT [41], AdaBoost [42], and CatBoost [43] have also been effectively applied in data inversion and analysis processes. However, machine learning algorithms depend on training data, which still leads to uncertainty in the model’s adaptability and prediction performance across domains. Moreover, differences in the sensitivity of various models to data features result in varying performance under specific conditions, increasing the difficulty of model selection and parameter optimization [44].
This study addresses this issue by using seven meteorological factors, including precipitation, temperature, wind speed, and sunshine hours, as driving data. We constructed 30 daily snow water equivalent prediction models for the next 1 to 30 days, and combined these 30 daily prediction models to form the snow water equivalent prediction model for the next 30 days. To comprehensively select machine learning models with better accuracy, our experiments compared the performance of nine common machine learning models—linear regression, decision trees, random forest, support vector machine, neural networks, AdaBoost, XGBoost, GBDT, and CatBoost—in snow data prediction. The top-performing models, CatBoost, ANNs, and GBDT, were then selected for ensemble forecasting to obtain the snow water equivalent prediction values for the next 30 days. This study selected the upper Yalong River Basin above the Ganzi Station as the typical research area, aiming to explore the complex relationships between meteorological factors and the SWE in a more comprehensive manner and conduct a comparative analysis of the simulation results. The SWE prediction model developed in this study can provide a reference for hydrological forecasting in cold regions. The experimental results show that different models perform differently under various time scales and meteorological conditions, providing a solid foundation for accurate SWE prediction. This study not only demonstrates the potential of machine learning in SWE prediction but also provides strong support for further optimizing snow water equivalent forecasting in the future, ultimately contributing to the sustainable utilization of water resources.

2. Study Area and Data

2.1. Study Area

The Yalong River originates from the southern slopes of the Bayan Har Mountains in Yushu Prefecture, Qinghai Province. The main stream stretches 1571 km, with a basin area of 136,000 km2 and a natural elevation drop of 3830 m, with maximum topographic relief exceeding 5000 m. Among the key hydrological stations in the upper basin, the Ganzi Hydrological Station controls a drainage area of 32,500 km2, with an average altitude of over 4500 m (Figure 1a). The region experiences a high-altitude, cold plateau climate characterized by intense sunlight and long winters, with significant spatiotemporal variations in snow cover and seasonal permafrost. Spring snowmelt and seasonal permafrost thaw play a critical role in runoff generation, confluence mechanisms, and hydrological processes during both flood and dry seasons [45].
From autumn to winter, the temperatures in the upper basin above Ganzi gradually drop below freezing, with snow accumulation starting in October and November. The snow cover peaks in March of the following year, and the snowmelt period primarily occurs in April and May. By early June, the snow is almost completely melted, with the annual average snow cover exceeding 50% of the basin (Figure 1b). The seasonal distribution of transient and seasonal permafrost follows a similar pattern to the snow cover. The permafrost typically begins to form around October, reaching its maximum depth between January and February of the following year. In spring, the permafrost thaws rapidly during April and May, with near-complete thawing by early June.
Under the influence of global climate change, the spatiotemporal variability of the snowmelt and seasonal permafrost in the basin has increased in recent years, leading to greater instability in the spring runoff [46,47]. This variability poses challenges to the scientific management of the cascade reservoir system and the efficient development and utilization of hydropower resources in the basin.

2.2. Data

2.2.1. Meteorological Data

The meteorological data used in this study were sourced from the CN05.1 [48,49] daily meteorological grid dataset released by the China Meteorological Administration. The dataset has a temporal resolution of 1 day and a spatial resolution of 0.25°. The CN05.1 dataset applies observational data from over 2400 stations within China, using the anomaly approximation method. The climate field and anomaly field are interpolated separately and then superimposed to obtain the data. The dataset includes daily precipitation (Pre), average temperature (Tm), maximum temperature (Tmax), minimum temperature (Tmin), wind speed (Win), sunshine duration (Ssd), and relative humidity (Rhu) data. The data are of good quality and are widely used in daily research. Area-averaged daily data for Pre, Tm, Tmax, Tmin, Win, Ssd, and Rhu from 2013 to 2022 for the Ganzi Basin were collected for this analysis and were used as driving data for the subsequent snow water equivalent prediction model. Figure 2 shows the spatial distribution map of CN05.1 precipitation, mean temperature, maximum temperature, minimum temperature, sunshine duration, relative humidity, and wind speed data.

2.2.2. Remote Sensing Snow Water Equivalent Data

The snow water equivalent (SWE) data were obtained from the AMSR2 daily snow water equivalent grid dataset. This dataset is based on satellite observations using passive microwave remote sensing technology to monitor the water content of surface snow. The SWE refers to the depth of water that would result from the melting of the snow on the surface, typically measured in millimeters, and is a key parameter for assessing snow water resources and predicting snowmelt runoff. The dataset, based on satellite remote sensing observations, has a temporal resolution of 1 day and a spatial resolution of 0.25°. To ensure the continuity and reliability of the data, after collecting and extracting the daily SWE raster data for the area above Ganzi, two steps were performed: removing outliers and imputing missing values. First, it was set that if the difference between the value of a grid on a given day and the values of the adjacent two days was greater than three times the value, the value was considered an outlier. The identified outlier grid values were then replaced with the average of the values from the preceding and following days. Similarly, when there were missing values for a grid, they were filled with the average of the values from the preceding and following days. After removing outliers and imputing missing values in the raster data, the area-averaged values for the basin above Ganzi were extracted. Daily snow water equivalent data for the period from 2013 to 2022 for the basin above Ganzi were collected and used for preliminary simulations and a forecast accuracy evaluation of the snow water equivalent prediction.

3. Methods

In recent years, machine learning models have been widely applied in fields such as snow remote sensing data inversion and snowmelt runoff numerical simulation. Different models have their own advantages under different simulation scenarios. To comprehensively compare the performance of various machine learning algorithms in simulation and prediction, we selected nine commonly used machine learning algorithms in snow accumulation and melting, and compared their accuracy in snow water equivalent prediction. The following is a brief introduction to these nine algorithms.

3.1. Machine Learning Algorithms

(1)
Linear Regression (LR)
Linear regression (LR) is a supervised machine learning model [50,51] suitable for numerical prediction tasks. Compared to other machine learning algorithms, linear regression has a short training time and performs well in datasets with few features and a strong linear relationship. The main calculation formula is as follows:
y = β 0 + β 1 x 1 + β 2 x 2 + + β n x n + ε
where y is the predicted value; β i is the predicted value; x i are the features, i [1, n]; ε is the error term.
(2)
Decision Trees (DT)
Decision trees represent a supervised machine learning model that uses a tree structure to recursively split data [52,53]. At each node, the model splits the dataset based on feature selection, typically by minimizing the impurity index to choose the best splitting feature. During training, for regression tasks, the model selects the optimal features for splitting based on the mean square error (MSE) criterion. The main calculation formula is as follows:
M S E = 1 n i = 1 n ( y i y ^ i ) 2
where n is the number of samples; y i is the actual value of the i-th sample; y ^ i is the predicted value of the sample.
(3)
Random Forest (RF)
The random forest (RF) model is a supervised machine learning model [54,55]. Compared with other machine learning algorithms, it has a short training time and high accuracy. The main calculation formula is as follows:
y ^ = 1 T t 1 T   h t ( x )
where ŷ is the output of the random forest; h t ( x ) is the prediction of the t-th decision tree; T is the number of decision trees.
(4)
Support Vector Machine (SVM)
A support vector machine (SVM) is a supervised machine learning model aimed at optimizing classification results by maximizing the margin of the hyperplane that separates the classes [56,57]. The SVM model performs well on data with a few outliers and high-dimensional data. By introducing a kernel function, the model can handle nonlinear data. The main formula is as follows:
Decision   Function : f x = s i g n w x + b
Optimization   Objective : m i n 1 2 w 2
Constraints : w T x i + b 1 ,   y i = + 1
w T x i + b 1 ,   y i = 1
where f(x) is the decision function, representing the classification result, which outputs +1 or −1, corresponding to the positive and negative classes, respectively; w is the normal vector of the hyperplane; x is the feature vector; b is the bias term, which controls the position of the decision boundary and allows it to flexibly adapt to the data; y i is the label of the i-th sample, with values +1 or −1; x i is the feature vector of the i-th sample.
(5)
Artificial Neural Network (ANN)
Artificial neural networks (ANNs) are machine learning models inspired by biological neural networks, capable of modeling complex nonlinear relationships [58,59] They excel in handling large-scale data and complex relationships between features, but the training process can be relatively time-consuming. The main calculation formula is as follows:
y = f i = 1 n   w i x i + b
where y is the output of the neuron, also known as the predicted value; f is the activation function, which introduces nonlinearity, enabling the neural network to model complex nonlinear mappings; w i is the weight associated with the i-th input feature, indicating the importance of the input feature in the output; x i is the value of the i-th input feature, representing the input data; b is the bias term, used to adjust the output of the model, making it more flexible; n is the total number of input features.
(6)
AdaBoost
AdaBoost (adaptive boosting) is an ensemble learning algorithm that improves the accuracy of a classifier by combining multiple weak classifiers (e.g., decision trees) [60,61,62]. During training, the model progressively focuses on misclassified samples, increasing the weights of these samples to improve the model’s performance on them. The main calculation formula is as follows:
H ( x ) = s i g n t 1 T   α t h t ( x )
where H(x) is the final strong classifier’s prediction for sample x, which is the weighted sum of the predictions from multiple weak classifiers; T is the total number of weak classifiers; α t is the weight of the t-th weak classifier, indicating its importance in the final decision, typically inversely proportional to the classifier’s error rate; h t ( x ) is the prediction of the t-th weak classifier for sample x.
(7)
XGBoost
XGBoost is an optimized version of the gradient boosting algorithm that incorporates a regularization term to control the model complexity, thereby reducing overfitting. XGBoost is known for its speed and high performance, making it suitable for handling large-scale data [63,64]. The main calculation formula is as follows:
Obj = i = 1 n   l y i , y ˆ i + k = 1 K   Ω f k
where Obj is the overall objective function; n is the number of training samples; l y i , y ˆ i is the loss function, which measures the error between the predicted value y ˆ i and the true value y i ; K is the total number of decision trees; Ω f k is the regularization term for the k-th tree, which controls the complexity of the tree.
(8)
Gradient Boosting Decision Tree (GBDT)
A gradient boosting decision tree (GBDT) is an ensemble model that iteratively trains multiple decision trees, with each tree optimizing the residuals from the previous one. The GBDT method is suitable for both regression and classification tasks [65,66]. It effectively captures nonlinear relationships between features but requires longer training times. The main calculation formula is as follows:
r i ( m ) = y i F m 1 x i
where x i is represents the feature vector of the i-th sample; r i ( m ) is the residual for the i-th sample in the m-th iteration, indicating the model’s improvement direction for the target in this iteration; y i is the true value of the i-th sample; F ( m 1 ) x i is the prediction of the i-th sample from the model in the (m − 1)-th iteration.
(9)
CatBoost
CatBoost is an improved gradient boosting algorithm with a similar formula to XGBoost [67,68]. The main advantage of CatBoost lies in its unique handling of categorical features; by using ordered target encoding, it effectively prevents information leakage. Additionally, CatBoost employs a symmetric tree structure, enhancing both the model’s stability and training efficiency.

3.2. Sensitivity Analysis with the SHAP Model

The Shapley Additive Explanations (SHAP) model can measure the interaction of different features on the final prediction by separating the marginal contribution of each feature to the predicted value [69,70]. Let the i-th sample be xi, the j-th feature of the i-th sample be xij, the model’s prediction for the sample be yi, and the average value of the sample variables be ybase. The SHAP value is calculated as follows:
y i j = y b a s e + f x i 1 + f x i 2 + f x i 3 + + f x i j
In this equation, f x i j represents the SHAP value of xij, which is the contribution of the j-th feature of the i-th sample to the final prediction yi. The absolute value of the SHAP value reflects the impact of the predictor variable on the model’s prediction; the larger the absolute value, the more significant the influence of that predictor variable on the feature variable. When f x i j > 0, this indicates that the feature increases the predicted value and has a positive effect on the model’s prediction, meaning an increase in the factor value promotes an increase in the SWE value. Conversely, when f x i j < 0, this indicates that the feature decreases the predicted value and has a negative effect on the model’s prediction, meaning an increase in the factor value leads to a decrease in the SWE value. This study analyzed the marginal contribution of each feature in the snow water equivalent prediction model to the predicted values, obtaining the sensitivity levels of different feature values.

3.3. Model Construction

The input features for machine learning models are crucial for the accuracy of the model’s training results. To comprehensively simulate the correlation between various factors and snow, it is necessary to consider the basic elements across the watershed. In this study, machine learning algorithms were used to model the daily SWE in the upper Yalong River Basin above Ganzi. The algorithm takes into account not only the conventional meteorological elements but also the characteristics of snow in the watershed. The primary input features include the total precipitation, average temperature at 2 m, maximum temperature, minimum temperature, sunshine duration, wind speed, and relative humidity. Among the input features, the total precipitation is the daily cumulative value, while the average temperature, sunshine duration, wind speed, and relative humidity represent daily averages.
Since the SWE is related to the cumulative effect of meteorological factors such as precipitation and temperature, if the forecast start date is 1 January 2022, with a lead time of 1–30 days, the changes in the SWE during the lead times are related to the cumulative changes in meteorological factors (precipitation, temperature, etc.) over 1–30 days. Therefore, the input data for the model include the SWE on the start date, along with the cumulative values of the total precipitation, average temperature at 2 m, maximum temperature, minimum temperature, sunshine duration, wind speed, and relative humidity during the lead time.
This study applied nine machine learning models, with each model predicting the SWE for every day from the 1st day up to the 30th day, resulting in 30 SWE prediction models. Each model was applied for a different lead time ranging from 1 to 30 days, and the 30 models were combined into the final daily prediction model for the next 30 days. The expression is as follows:
y t , j = f j S W E 0 ;   i = 1 t P i t ;   i = 1 t T m i t ;   i = 1 t T m a x i t i = 1 t T m i n i t ;   i = 1 t R h u i t i = 1 t S s d i t ;   i = 1 t W i n i t , t 1 ,   30 , j [ 1 ,   9 ]
where t is the lead time, t 1 ,   30 ; j is the model number, j 1 ,   9 ; f j is the j-th model; y t , j is the predicted SWE for the j-th model on the t-th day; S W E 0 is the SWE from the day before the forecast start date; P i is the precipitation on the i-th day; T m i is the average temperature on the i-th day; T m a x i is the maximum temperature on the i-th day; T m i n i is the minimum temperature on the i-th day; R h u i is the relative humidity on the i-th day; S s d i is the sunshine duration on the i-th day; W i n i is the wind speed on the i-th day.
The SWE prediction models were trained with the objective of maximizing the Nash–Sutcliffe Efficiency (NSE).

3.4. Ensemble Mean (EM) Model

The ensemble mean (EM) model is a statistical method that improves the overall prediction performance by combining the prediction results from multiple models [71]. In hydrological and meteorological forecasting, the EM method is widely used to reduce random errors and systematic biases that may exist in individual models. The core idea is to integrate the predictions of multiple models through simple averaging or weighted averaging, thereby generating more robust and accurate results.
The mathematical expression of the EM method is as follows:
E M t = 1 N j = 1 N   y t , j
where E M t is the prediction result of the ensemble mean model; N is the number of models participating in the ensemble; y t , j is the prediction of the j-th model for the t-th day.

3.5. Snow Water Equivalent Prediction Model

This study used the period from 2013 to 2019 as the model calibration period for training and hyperparameter optimization, and the period from 2020 to 2022 as the validation period to assess the model’s generalization ability. During the calibration period, K-fold cross-validation combined with grid search cross-validation [72] was employed for hyperparameter tuning. Nine machine learning models were used to establish 30 individual snow water equivalent prediction models, forecasting the snow water equivalent for the next 1 to 30 days. These models were then combined to form a set of daily prediction models for the next 30 days. Finally, based on the combined NSE, RMSE, and RE rankings, the top three models were selected for ensemble forecasting, ultimately obtaining the snow water equivalent prediction values for the next 30 days (Figure 3).

3.6. Evaluation Metrics

Three evaluation metrics were used to assess the accuracy of the results: the Nash–Sutcliffe Efficiency (NSE), root mean square error (RMSE), and relative error (RE) [73].
The Nash–Sutcliffe Efficiency (NSE) is a commonly used parameter for evaluating the quality of hydrological models, with its value ranging from negative infinity to 1 [74]. The closer the value is to 1, the better the model quality and reliability. Values closer to 0 indicate that the model is close to the mean of the observed values, suggesting that while the overall results are reliable, significant errors exist in the simulation process; values far less than 0 indicate that the model is unreasonable. The specific mathematical formula for the NSE is as follows:
N S E = 1 t = 1 T X 0 t X m t 2 t = 1 T X 0 t X ¯ 0 2
where T is the number of samples; X 0 t is the observed value; X m t is the predicted value; X ¯ 0 is the mean of the observed values.
The root mean square error (RMSE) is a method for assessing the goodness of fit of a regression model to a dataset, displaying the average distance between the model’s predictions and the actual values in the dataset. The lower the RMSE, the better the model “fits” the dataset. A value closer to 0 is preferred. The formula for the RMSE is:
R M S E = 1 n i = 1 n ( y i y ^ i ) 2
where n is the number of samples; y i is the observed value; y ^ i is the predicted value.
The relative error (RE) refers to the ratio of the absolute error of the measurement to the true value (or accepted value) of the quantity being measured, multiplied by 100%, and is expressed as a percentage. Generally, the relative error provides a better reflection of the reliability of the measurement. The unit of the RE is %, and the closer the RE value is to 0, the better [75].
R E = | y i y ^ i | y i × 100 %
where y i is the observed value; y ^ i is the predicted value.

4. Results

4.1. Comparison of Machine Learning Model Performance During the Testing Period

The performance of nine machine learning models was evaluated, and the results showed that the Nash–Sutcliffe Efficiency (NSE) for all models during the training period was greater than 0.87, indicating that the models had high capability to capture trends and changes in observed values. In hydrological forecasting, predictions are typically categorized based on lead time into short-term (0–3 days), medium-term (3–10 days), and long-term (over 10 days) forecasts. Based on historical training data, the SWE for the testing period from 1 January 2020 to 31 December 2022 was predicted for each of the 1st to 30th days. The lead time was further divided into three sub-periods: days 1–10, days 11–20, and days 21–30. A comparative analysis of the prediction performance for these nine models across each period is shown in Figure 4.
During the 1–10 day lead time, the average NSE for all nine machine learning models was greater than 0.9, indicating that the models fit the short-term SWE change trends with high accuracy (Figure 4a). Additionally, the average RMSE was less than 5.1 mm. Except for the SVR model, which exhibited larger fluctuations in RMSE, the other models had RMSE fluctuations of around 3 mm (Figure 4d). Furthermore, the average RE also showed a small error range for the models during the short-term lead time (Figure 4g). Therefore, for short-term SWE forecasting, models such as the ANN, GBDT, and RF demonstrated high accuracy and stability.
During the medium-term lead time (days 11–20), the average NSE for all nine models was greater than 0.85, showing that the models still maintained high accuracy in predictions over a longer period (Figure 4b). The average RMSE was less than 6.3 mm, with the decision tree model exhibiting relatively large fluctuations in RMSE (around 0.4 mm), while the RMSE fluctuations for the other models were all below 0.1 mm (Figure 4e). The average RE for all models was less than 5.5%, with small fluctuations (Figure 4h). In this period, models such as CatBoost, ANN, and GBDT performed particularly well, effectively capturing the SWE variation trends over the medium term.
In the long-term lead time (days 21–30), the average NSE for all models was greater than 0.8, with fluctuations in NSE being less than 0.2, demonstrating a certain level of stability (Figure 4c). The average RMSE was less than 6.8 mm, showing that the models maintained good forecasting capabilities over a longer time scale (Figure 4f). The average RE was below 6%, with minimal fluctuation (Figure 4i). A comprehensive analysis found that during the long-term forecast, models such as CatBoost, ANN, and GBDT continued to show relatively high prediction accuracy and stability, providing valuable insights for future long-term SWE forecasting.
To comprehensively evaluate the prediction performance of machine learning models under different SWE scenarios, this study selected three models with the best overall performance: ANN, GBDT, and CatBoost. Figure 5a–f presents the prediction results for the future 30 days of the SWE for several example initialized forecast dates, corresponding to high SWE values, low SWE values, and rapid accumulation and melt periods, in order to validate the model’s applicability under diverse scenarios.
The results show that the EM model exhibited stable and accurate prediction performance during the aforementioned typical periods. The RMSE for all periods was below 6 mm, reflecting high accuracy, especially in cases where the initial snow conditions were stable, such as on 1 January 2020 (Figure 5a) and 15 January 2022 (Figure 5d), which showed the lowest RMSE values, demonstrating the model’s good adaptability to initial snow conditions. The NSE exceeded 0.85 for all periods, indicating that the model could accurately capture the dynamic trends of SWE changes. Notably, during the rapid snowmelt phases on 26 March 2020 (Figure 5c) and 1 March 2022 (Figure 5e), the NSE approached 0.9, showing the model’s sensitivity to rapid changes. At the same time, for the snowmelt process during the period of 21 March 2020 (Figure 5b), the snow water equivalent monthly melt value was greater than 50 mm, and the NSE value of 0.87 indicated a good forecast of the melting trend. The RE values fluctuated within 10%, with the lowest RE values observed on 1 January 2020 (Figure 5a) and 25 April 2022 (Figure 5f), further confirming the robustness and reliability of the ensemble mean model for future SWE predictions.
Furthermore, the uncertainty range of the model predictions (represented by the pink shaded area) covered almost all observed true values, especially during the high SWE and rapid melt phases, where the prediction intervals effectively captured the trend changes. For instance, in the forecast initialized on 1 January 2020, during the middle period, the ensemble mean model overestimated the SWE by approximately 3–4 mm. However, considering the uncertainty, the model provided a range from the three models, which included the observed true value. These results further validate the applicability and advantages of the ensemble mean model in predicting the SWE in the dynamic and complex high-altitude regions.

4.2. Snow Sensitivity Analysis

Given the artificial neural network (ANN) model performs best in predicting SWE, this study further conducted a sensitivity analysis on the key influencing factors for the ANN using the SHAP method. The SHAP plots provide deep insights into the importance and directional influence of each meteorological factor on SWE prediction, offering an intuitive basis for understanding the model’s prediction logic.
The SHAP values reflect the contribution of each feature to the model’s prediction outcome. According to the SHAP analysis results (Figure 6), aside from the initial SWE value on the forecast start date, Tm, Tmin, and Tmax had SHAP values of 6.2, 5, and 3.3, respectively, making them the most influential meteorological factors with the highest sensitivity to the SWE. Following these, Rhu and Pre had SHAP values of 2 and 1.4, respectively, and also had considerable impacts on the SWE. In contrast, Ssd and Win had SHAP values of 0.6 and 0.3, indicating relatively lower sensitivity. Specifically (Figure 7), higher SHAP values (light yellow) contribute strongly in a positive direction to the prediction outcome, while lower values (dark green) exert a negative influence. The minimum temperature, relative humidity, and sunshine duration have a positive effect on the SWE, meaning their increase helps to increase the SWE. In contrast, the average temperature, maximum temperature, precipitation, and wind speed show a negative effect, where their increase leads to a decrease in SWE.
From the analysis of Figure 6 and Figure 8a–f, it can be further observed that the SWE exhibits a positive change when the following meteorological conditions are met: average temperature below −2.5 °C, minimum temperature below −12 °C, maximum temperature below 7.5 °C, precipitation less than 1 mm, relative humidity greater than 55%, sunshine duration exceeding 6 h, and wind speed below 2.5 m/s. These conditions collectively indicate the accumulation and retention process of the SWE. Low temperatures reduce the rate of snowmelt, high humidity minimizes evaporation losses, and longer sunshine durations may indirectly affect surface temperature and precipitation patterns in cold environments.

5. Discussion

5.1. Model Accuracy Evaluation

This study focused on the upper Yalong River Basin above Ganzi, an area with an average elevation exceeding 4500 m. The high altitude and high solar radiation characteristics make the snowmelt process highly sensitive to meteorological variables (such as temperature and humidity), and the relationship between meteorological variables and the SWE is relatively clear. These characteristics provide favorable conditions for machine learning models to capture the trends of SWE variations.
The CatBoost model [76], through optimizing the decision tree training process, can fully utilize mixed inputs of categorical and numerical features, significantly enhancing the model’s ability to model complex variable interactions. Its high sensitivity to outliers makes it particularly effective under the complex climate conditions of the study area. The ANN model automatically extracts nonlinear patterns of input features through hidden layers and can dynamically adjust feature weights. It is particularly sensitive to the impacts of key variables, such as temperature and humidity, meaning the simulation results closely align with actual trends. The GBDT model, by iteratively fitting residuals, can precisely model nonlinear relationships between variables, especially for complex but nonexplicit variable interactions within the study area. Furthermore, the GBDT model [65,66] shows greater robustness in handling outliers and missing values, with excellent performance in predicting extreme values.
An analysis of the EM model [77], which combines the predictions from CatBoost, ANN, and GBDT methods, shows that integrating these models’ results effectively reduces the bias of individual models, significantly improving the prediction accuracy and robustness of SWE forecasts for the next 30 days. This fusion method has shown good adaptability and universality under the high-altitude and complex climatic conditions of the study area. This indicates that the ensemble mean model not only provides high-precision daily forecasts but also offers a scientific basis for risk assessment and emergency management, helping decision-makers make more accurate and robust judgments when addressing snowmelt risks and water resource allocation needs.
From January to May, the SWE prediction accuracy is highest in January, followed by March and February, with the largest errors in May. In January, the snow accumulation peaks and SWE changes are stable, making predictions easier. In March, the snow remains abundant and melts slowly, keeping the impact of meteorological factors stable. In February, as melting begins and temperatures rise, the SWE becomes more sensitive to fluctuating weather, increasing the prediction difficulty. By April and May, accelerated snowmelt and more volatile conditions lead to greater prediction errors.

5.2. Sensitivity Analysis

Figure 6, Figure 7 and Figure 8 display the sensitivity analysis of the SWE, showcasing the sensitivity rankings and positive (negative) correlations of different meteorological factors, providing important insights for further discussion.
From Figure 6, it can be seen that Rhu is ranked higher in terms of sensitivity compared to the Pre. High-humidity environments reduce evaporation losses and help maintain snow cover, thereby increasing the SWE [78]. Conversely, low humidity accelerates the sublimation and evaporation of snow, leading to a significant reduction in the SWE. Particularly in high-altitude or cold regions, high humidity is typically accompanied by lower radiative evaporation demands, a phenomenon that is especially pronounced. Additionally, the relative humidity not only directly impacts the SWE by reducing evaporation but may also enhance its importance through interactions with other variables, such as temperature and radiation. In contrast, although precipitation has a direct contribution to the SWE, its sensitivity ranking may be lower due to the data distribution, time scale, or complexity of variable interactions in the model.
According to Figure 7 and Figure 8e, the Ssd has a significant positive impact on the SWE [79]. From a physical mechanism perspective, a longer sunshine durations mean that the snow surface absorbs more solar radiation energy, accelerating snowmelt. This melting process increases the surface SWE in a short time. Additionally, under low-temperature conditions, an increased sunshine duration may raise both surface and atmospheric temperatures, altering precipitation forms, such as converting solid precipitation (snow) into liquid precipitation (rain), further increasing the SWE. The interaction between the sunshine duration and Tm also significantly influences SWE. Under lower temperature conditions, increased sunshine duration has a more pronounced effect on snowmelt, while under higher temperature conditions, its effect is more apparent in changes in precipitation forms. These interaction effects are clearly presented in the SHAP dependence plots, showing the variation of SHAP values under different sunshine durations.
In conclusion, the relative humidity and sunshine duration not only have direct effects on the SWE but also amplify their influence through interactions with other key factors, such as temperature. These results provide important insights for understanding the mechanisms behind SWE variations and improving forecasting models.

5.3. Advantages Compared to Other Snow Water Equivalent Models

The snow water equivalent prediction method proposed in this study has the following three main advantages compared to some of the existing models [80,81]. First, nine machine learning models were selected and comprehensively evaluated, providing multiple model options and offering valuable references for related research [82]. Second, based on the prediction results of these models, an ensemble forecasting method was applied [71,77]. By combining the predictions of multiple models, a more reliable forecast range for snow water equivalent can be provided, better supporting early warning and prevention measures under uncertainties. Finally, the study introduced a rolling prediction method, where the snow water equivalent value at the initial time was assimilated into the model with each prediction, thereby continuously improving the accuracy of the model [83].

5.4. Model Future Improvement

This study used CN05.1 meteorological data and AMSR remote sensing SWE data for model training. In real-time forecasting, future SWE predictions for the next 30 days could be made by combining weather forecast data from climate forecasting agencies, such as the European Centre for Medium-Range Weather Forecasts (ECMWF) [84], which includes variables such as the temperature, precipitation, relative humidity, sunshine duration, and wind speed.
Although the model demonstrated high accuracy during lead times, the simulation accuracy still needs to be improved during periods of rapid SWE changes. Traditional machine learning models lack the ability to deeply model temporal dependencies in time series data. Additionally, some key surface factors, such as terrain variations, land use, and soil moisture, were not included in the models, which may have contributed to the prediction errors. Future research studies could be improved in the following ways:
(1)
Introduce time series modeling methods (e.g., LSTM or transformer) [85] to better capture the temporal dynamics of snowmelt processes;
(2)
Combine multi-model ensemble techniques to integrate the advantages of linear and nonlinear models to enhance the prediction accuracy and robustness;
(3)
Further expand the input features, such as the initial snow conditions, surface evaporation, and soil moisture, to improve the model’s explanatory power for SWE variation mechanisms.
These improvements could provide a more solid theoretical foundation for accurately forecasting the SWE during the snowmelt period in high-altitude regions.

6. Conclusions

This study focused on the variation characteristics of the SWE in high-altitude regions, and predicted and analyzed the future 30-day SWE using nine machine learning models. Based on a model performance evaluation and sensitivity analysis, the following conclusions were drawn:
(1)
Nine machine learning models were selected for predicting the future 30-day SWE: linear regression, decision trees, random forest, SVR, ANN, AdaBoost, XGBoost, GBDT, and CatBoost. From the results of the single-day predictions, all nine models demonstrated average NSE values greater than 0.8, average RMSE values less than 8 mm, and average RE values less than 7% during the 1–10 day, 11–20 day, and 21–30 day lead times. Among these, the CatBoost, ANN, and GBDT models performed well across the three lead times (1–10 days, 11–20 days, 21–30 days) and three evaluation metrics (RMSE, NSE, RE), showing excellent trend capture ability and low error values.
(2)
The results showed that the ensemble mean model (a fusion of the CatBoost, ANN, and GBDT models) was able to capture the SWE trend effectively for each forecast start date, especially during key periods (such as the spring snowmelt season), demonstrating strong trend simulation capabilities. Compared to the individual models, the ensemble mean model significantly reduced the error impacts of individual models, producing more robust and accurate predictions. This fusion method is adept at handling the nonlinear characteristics of climate variations in high-altitude regions, providing stable predictions for continuous SWE data over the next 30 days.
(3)
The sensitivity analysis revealed that the variation in the SWE is highly sensitive to meteorological factors. Among these, Tm, Tmin, and Tmax are the most significant drivers of the SWE, with a negative impact on the SWE. On the other hand, Rhu has a positive regulating effect on the SWE; high humidity reduces snow evaporation losses, thereby increasing SWE. Furthermore, Ssd and Win have lower sensitivity, although under specific conditions (such as high-radiation or low-temperature environments) they may still influence SWE. The interactions among these factors were well reflected in the model predictions, providing important insights into the mechanisms driving SWE variation.

Author Contributions

Conceptualization, writing—original draft preparation, J.Z.; writing—review and editing, N.D.; visualization, writing—review and editing, Y.W.; project administration, writing—review and editing, M.Y. All authors have read and agreed to the published version of the manuscript.

Funding

The research was funded by the National Key Research and Development Program of China (2023YFC3081000), the National Natural Science Foundation of China (42401053), the Free Exploration Project of the State Key Laboratory of Watershed Hydrological Cycle Simulation and Regulation (WR110146B0062024, SKL2024YJZD02), the Fundamental Research Funds for the Central Universities project of the China Institute of Water Resources and Hydropower Research (WR110145B0112024), the Research Programme of the Kunming Engineering Corporation Limited (No. DJ-HXGG-2021–04), and the Key Research and Development Programme of Yunnan Province (No. 202203AA080010).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data can be obtained from the author.

Conflicts of Interest

The authors declare that this study received funding from Kunming Engineering Corporation Limited. The funder was not involved in the study design, collection, analysis, interpretation of data, the writing of this article or the decision to submit it for publication.

References

  1. Zhu, L.; Ma, G.; Zhang, Y.; Wang, J.; Tian, W.; Kan, X. Accelerated Decline of Snow Cover in China from 1979 to 2018 Observed from Space. Sci. Total Environ. 2022, 814, 152491. [Google Scholar] [CrossRef] [PubMed]
  2. Yu, X.; Hu, X.; Wang, G.; Wang, K.; Chen, X. Machine Grid Search Cross-Validation Learning Estimation of Snow Depth in 2021 Texas Statewide Winter Storm Using SAR Imagery. Geophys. Res. Lett. 2022, 49, e2022GL099119. [Google Scholar] [CrossRef]
  3. Yang, J.; Chen, Y.; Wilson, J.P.; Chun, Y.; Chen, Y.; Su, H. A Zero-Inflated Spatiotemporal Analysis for Snowpack Variations and Influence of Environmental Factors in the Northern Hemisphere. J. Hydrol. 2023, 616, 128760. [Google Scholar] [CrossRef]
  4. Brown, R.D.; Brasnett, B.; Robinson, D. Gridded North American Monthly Snow Depth and Snow Water Equivalent for GCM Evaluation. Atmos. Ocean 2003, 41, 1–14. [Google Scholar] [CrossRef]
  5. Fontrodona-Bach, A.; Schaefli, B.; Woods, R.; Teuling, A.J.; Larsen, J.R. NH-SWE: Northern Hemisphere Snow Water Equivalent Dataset Based on In-Situ Snow Depth Time Series. Earth Syst. Sci. Data Discuss. 2023, 15, 2577–2599. [Google Scholar] [CrossRef]
  6. Wu, X.J.; Zhu, R.; Long, Y.P.; Zhang, W. Spatial Trend and Impact of Snowmelt Rate in Spring across China’s Three Main Stable Snow Cover Regions over the Past 40 Years Based on Remote Sensing. Remote Sens. 2022, 14, 4176. [Google Scholar] [CrossRef]
  7. Liu, Z.; Cuo, L.; Sun, N. Tracking Snowmelt During Hydrological Surface Processes Using a Distributed Hydrological Model in a Mesoscale Basin on the Tibetan Plateau. J. Hydrol. 2023, 616, 128796. [Google Scholar] [CrossRef]
  8. Abudu, S.; Cui, C.; Muattar, S.; King, J. Application of Snowmelt Runoff Model (SRM) in Mountainous Watersheds: A Review. Water Sci. Eng. 2012, 5, 123–136. [Google Scholar] [CrossRef]
  9. Tian, F.; Lin, Z.; Zhang, H.; Yang, G.C. Interannual Variation of Spring Streamflow and Its Relationship with Snowpack in the Three-River Headwater Region, China. Clim. Environ. Res. 2024, 29, 588–604. [Google Scholar] [CrossRef]
  10. Zhao, Q.; Tan, X.; Zeng, Q.; Hang, Z.; Wu, J.W. Combined Effects of Temperature and Precipitation on the Spring Runoff Generation Process in a Seasonal Freezing Agricultural Watershed. Environ. Earth Sci. 2021, 80, 490. [Google Scholar] [CrossRef]
  11. Steele, H.; Small, E.E.; Raleigh, M.S. Demonstrating a Hybrid Machine Learning Approach for Snow Characteristic Estimation Throughout the Western United States. Water Resour. Res. 2024, 60, e2023WR035805. [Google Scholar] [CrossRef]
  12. Dietz, A.J.; Kuenzer, C.; Gessner, U.; Dech, S. Remote Sensing of Snow–A Review of Available Methods. Int. J. Remote Sens. 2011, 33, 4094–4134. [Google Scholar] [CrossRef]
  13. Li, G.Y. In-Depth Implementation of the Spirit of the 20th National Congress of the Communist Party of China and Solid Promotion of High-Quality Development of Water Resources in the New Era—Speech at the 2023 National Water Resources Work Conference. Water Resour. Dev. Res. 2023, 23, 1–11. [Google Scholar] [CrossRef]
  14. Li, G.Y. Providing Strong Water Security Guarantee for the Comprehensive Advancement of National Power Construction and the Great Cause of National Rejuvenation with Chinese-Style Modernization—Speech at the 2024 National Water Resources Work Conference. Water Resour. Dev. Res. 2024, 24, 1–10. [Google Scholar] [CrossRef]
  15. Zuo, Q.T.; Wang, Z.Y.; Ma, J.X. Research Hotspots and Development Prospects of Modern Water Management in China. Water Resour. Dev. Res. 2024, 24, 13–19. [Google Scholar] [CrossRef]
  16. Li, Z.Z.; Zhang, W.; Liu, L. Vigorously Developing New Productive Forces to Promote High-Quality Development of Water Resources—Understanding and Reflections on the Development of New Productive Forces in the Water Industry. Water Resour. Dev. Res. 2024, 24, 1–6. [Google Scholar] [CrossRef]
  17. Xue, X.S.; Zhang, Y.X. From Carbon Peaking to Water Peaking: Observations on China’s Water Strategy and Innovative Development. Water Resour. Dev. Res. 2024, 24, 17–23. [Google Scholar] [CrossRef]
  18. Yu, L.-x.; Zhang, S.-w.; Bu, K.; Yang, J.-c.; Yan, F.-q.; Chang, L.-p. A Review on Snow Data Sets. Sci. Geogr. Sin. 2013, 33, 878–883. [Google Scholar]
  19. Dai, L.Y.; Xiao, L.; Wang, J.; Che, T. China Snow Cover Ground Survey Dataset (2017–2020). China Sci. Data 2022, 7, 9–23. [Google Scholar]
  20. Xiao, L.; Che, T.; Dai, L. Evaluation of Remote Sensing and Reanalysis Snow Depth Datasets over the Northern Hemisphere During 1980–2016. Remote Sens. 2020, 12, 3253. [Google Scholar] [CrossRef]
  21. Yang, K.; Musselman, K.N.; Rittger, K.; Margulis, S.A.; Painter, T.H.; Molotch, N.P. Combining Ground-Based and Remotely Sensed Snow Data in a Linear Regression Model for Real-Time Estimation of Snow Water Equivalent. Adv. Water Resour. 2022, 160, 104075. [Google Scholar] [CrossRef]
  22. Hall, D.K.; Riggs, G.A. Accuracy Assessment of the MODIS Snow Products. Hydrol. Process. 2007, 21, 1534–1547. [Google Scholar] [CrossRef]
  23. Goa, Y.; Xie, H.J.; Lu, N.; Yao, T.D.; Liang, T.G. Toward advanced daily cloud-free snow cover and snow water equivalent products from Terra-Aqua MODIS and Aqua AMSR-E measurements. J. Hydrol. 2010, 385, 23–35. [Google Scholar] [CrossRef]
  24. Fletcher, S.J.; Liston, G.E.; Hiemstra, C.A.; Miller, S.D. Assimilating MODIS and AMSR-E Snow Observations in a Snow Evolution Model. J. Hydrometeor. 2012, 13, 1475–1492. [Google Scholar] [CrossRef]
  25. Xu, G.; Liu, Q.; Chen, L.; Liu, L.Y. Remote Sensing and Sustainable Development in China: Opportunities and Challenges. J. Remote Sens. 2016, 20, 679–688. [Google Scholar]
  26. Shao, D.; Li, H.; Wang, J.; Hao, X.; Wang, R.; Ma, Y. Study on Snow Albedo Inversion Based on Multi-Source Remote Sensing Data. Remote Sens. Technol. Appl. 2017, 32, 71–77, 139. [Google Scholar]
  27. Jiang, Y.; Chen, F.; Gao, Y.; He, C.; Barlage, M.; Huang, W. Assessment of Uncertainty Sources in Snow Cover Simulation in the Tibetan Plateau. J. Geophys. Res. Atmos. 2020, 125, e2020JD032674. [Google Scholar] [CrossRef]
  28. Sharma, V.; Mishra, V.D.; Joshi, P.K. Snow Cover Variation and Streamflow Simulation in a Snow-Fed River Basin of the Northwest Himalaya. J. Mt. Sci. 2012, 9, 853–868. [Google Scholar] [CrossRef]
  29. Kustas, W.P.; Rango, A.; Uijlenhoet, R. A simple energy budget algorithm for the snowmelt runoff model. Water Resour. Res. 1994, 30, 1515–1527. [Google Scholar] [CrossRef]
  30. Zhou, G.; Cui, M.; Wan, J.; Zhang, S. A Review on Snowmelt Models: Progress and Prospect. Sustainability 2021, 13, 11485. [Google Scholar] [CrossRef]
  31. Bi, Y.J.; Zhao, Y.; Zhou, Z.M.; Zhai, J.Q. Application of improved double-layer snow model in the upper Lancang River. China Rural. Water Hydropower 2011, 08, 49–52. [Google Scholar]
  32. Grusson, Y.; Sun, X.; Gascoin, S.; Sauvage, S.; Raghavan, S.; Anctil, F.; Sánchez-Pérez, J.-M. Assessing the Capability of the SWAT Model to Simulate Snow, Snow Melt, and Streamflow Dynamics over an Alpine Watershed. J. Hydrol. 2015, 531, 574–588. [Google Scholar] [CrossRef]
  33. Zhao, Q.; Liu, Z.; Ye, B.; Qin, Y.; Wei, Z.; Fang, S. A Snowmelt Runoff Forecasting Model Coupling WRF and DHSVM. Hydrol. Earth Syst. Sci. 2009, 13, 1897–1906. [Google Scholar] [CrossRef]
  34. Mohammadi, B. A Review on the Applications of Machine Learning for Runoff Modeling. Sustain. Water Resour. Manag. 2021, 7, 98. [Google Scholar] [CrossRef]
  35. Wang, Y.; Huang, X.; Wang, J.; Zhou, M.; Liang, T. AMSR2 Snow Depth Downscaling Algorithm Based on a Multifactor Approach over the Tibetan Plateau, China. Remote Sens. Environ. 2019, 231, 111268. [Google Scholar] [CrossRef]
  36. Moradizadeh, M.; Alijanian, M.; Moeini, R. Spatial Downscaling of Snow Water Equivalent Using Machine Learning Methods over the Zayandehroud River Basin, Iran. PFG 2023, 91, 391–404. [Google Scholar] [CrossRef]
  37. Wang, G.; Hao, X.; Yao, X.; Wang, J.; Li, H.; Chen, R.; Liu, Z. Simulations of Snowmelt Runoff in a High-Altitude Mountainous Area Based on Big Data and Machine Learning Models: Taking the Xiying River Basin as an Example. Remote Sens. 2023, 15, 1118. [Google Scholar] [CrossRef]
  38. Zheng, L.K.; Wan, L.P.; Li, Z.Q. Linear Regression Analysis of the Impact of Study Pressure on Students’ Physical and Mental Health. Chin. Sch. Health 2001, 223, 2. [Google Scholar] [CrossRef]
  39. Yu, J.; Asche, C.V.; Fairchild, C.J. The Economic Burden of Dry Eye Disease in the United States: A Decision Tree Analysis. Cornea 2011, 30, 379–387. [Google Scholar] [CrossRef] [PubMed]
  40. Tan, J.; Ding, J.; Han, L.; Ge, X.; Wang, X.; Wang, J.; Wang, R.; Qin, S.; Zhang, Z.; Li, Y. Exploring PlanetScope Satellite Capabilities for Soil Salinity Estimation and Mapping in Arid Regions Oases. Remote Sens. 2023, 15, 1066. [Google Scholar] [CrossRef]
  41. Bo, W.; Zhao, K.; Cheng, G.; Wang, Y.; Zhang, J.; Cheng, M.; Yang, C.; Da, W. Study on Transportation Carbon Emissions in Tibet: Measurement, Prediction Model Development, and Analysis. Sustainability 2024, 16, 8419. [Google Scholar] [CrossRef]
  42. Morra, J.H.; Tu, Z.; Apostolova, L.G.; Green, A.E.; Thompson, P.M. Comparison of AdaBoost and Support Vector Machines for Detecting Alzheimer’s Disease Through Automated Hippocampal Segmentation. IEEE Trans. Med. Imaging 2009, 29, 30–43. [Google Scholar] [CrossRef]
  43. Muhammad, Z.; Faisal, H.M.; Muhammad, I.; Iqra, A.; Irfan, U.; Tufail, A.; He, Z.B. Factors Affecting Injury Severity in Motorcycle Crashes: Different Age Groups Analysis Using CatBoost and SHAP Techniques. Traffic Inj. Prev. 2024, 25, 472–481. [Google Scholar]
  44. Hou, J.; Huang, C.; Chen, W.; Zhang, Y. Improving Snow Estimates Through Assimilation of MODIS Fractional Snow Cover Data Using Machine Learning Algorithms and the Common Land Model. Water Resour. Res. 2021, 57, e2020WR029010. [Google Scholar] [CrossRef]
  45. Dong, N.P.; Wang, H.; Yang, M.X.; Zhang, J.J.; Xu, S.Q. New Xin’anjiang Model Considering Snowmelt and Soil Freezing/Thawing and Its Application—A Case Study of Runoff Simulation in the Upper Yalong River. Adv. Water Sci. 2024, 35, 530–542. [Google Scholar] [CrossRef]
  46. Ding, Y.J.; Zhang, S.Q.; Chen, R.S.; Qin, J.; Zhao, Q.; Liu, J.F.; Yang, Y.; He, X.B.; Chang, Y.P.; Shangguan, D.H.; et al. A Review on the Impact of Climate Change on the Cryospheric Hydrology. Adv. Clim. Change Res. 2025, 21, 1–21. [Google Scholar]
  47. Lv, S.; Tang, Y.; Tang, Q.; Li, H.; Xiao, H.; Xie, D. The Impact of Climate Change on Annual and Seasonal Runoff in the Lena River Basin, Arctic. Geogr. Res. 2024, 79, 2811–2829. [Google Scholar]
  48. Xu, Y.; Gao, X.J.; Shen, Y.; Xu, C.H.; Shi, Y.; Giorgi, F. A Daily Temperature Dataset Over China and Its Application in Validating a RCM Simulation. Adv. Atmos. Sci. 2009, 26, 763–772. [Google Scholar] [CrossRef]
  49. Wu, J.; Gao, X.J. A Gridded Daily Observation Dataset Over China Region and Comparison with the Other Datasets. Chin. J. Geophys. 2013, 56, 1102–1111. [Google Scholar]
  50. James, G.M.; Wang, J.; Zhu, J. Functional Linear Regression That’s Interpretable. Ann. Stat. 2009, 37, 2083–2108. [Google Scholar] [CrossRef]
  51. Seber, G.A.F.; Lee, A.J. Linear Regression Analysis, 2nd ed.; Wiley: New York, NY, USA, 2012. [Google Scholar] [CrossRef]
  52. Dietterich, T.G.; Kong, E.B. Machine Learning Bias, Statistical Bias, and Statistical Variance of Decision Tree Algorithms; Technical Report; Department of Computer Science, Oregon State University: Corvallis, OR, USA, 1995. [Google Scholar]
  53. Myles, A.J.; Feudale, R.N.; Liu, Y.; Woody, N.A.; Brown, S.D. An Introduction to Decision Tree Modeling. J. Chemom. 2010, 18, 275–285. [Google Scholar] [CrossRef]
  54. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  55. Daliakopoulos, I.N.; Tsanis, I.K. Comparison of an Artificial Neural Network and a Conceptual Rainfall–Runoff Model in the Simulation of Ephemeral Streamflow. Hydrol. Sci. J. 2016, 61, 2763–2774. [Google Scholar] [CrossRef]
  56. Alkheder, S.; Alrukaibi, F.; Aiash, A. Support Vector Machine (SVM), Random Forest (RF), Artificial Neural Network (ANN), and Bayesian Network for Prediction and Analysis of GCC Traffic Accidents. J. Ambient. Intell. Humaniz. Comput. 2022, 14, 7331–7339. [Google Scholar] [CrossRef]
  57. Gurung, S. Support Vector Machine (SVM). 2020. Available online: https://www.researchgate.net/publication/344747383_Support_Vector_Machine_SVM (accessed on 8 February 2025).
  58. Zemnazi, O.; El Filali, S.; Ouahabi, S. Weather Forecasting Using Artificial Neural Network (ANN): A Review. Procedia Comput. Sci. 2024, 241, 618–623. [Google Scholar] [CrossRef]
  59. Tan, Q.F.; Lei, X.H.; Wang, X.; Wang, H.; Wen, X.; Ji, Y.; Kang, A.Q. An Adaptive Middle and Long-Term Runoff Forecast Model Using EEMD-ANN Hybrid Approach. J. Hydrol. 2018, 567, 767–780. [Google Scholar] [CrossRef]
  60. Collins, M.; Schapire, R.E.; Singer, Y. Logistic Regression, AdaBoost, and Bregman Distances. Mach. Learn. 2002, 48, 253–285. [Google Scholar] [CrossRef]
  61. Wang, W.; Shan, D.; Li, D.; Liu, S.; Song, M.; Xiao, S.; Zhang, H. A Random Feature Mapping Method Based on the AdaBoost Algorithm and Results Fusion for Enhancing Classification Performance. Expert Syst. Appl. 2024, 256, 124902. [Google Scholar] [CrossRef]
  62. Freund, Y.; Schapire, R.E. A Decision-Theoretic Generalization of Online Learning and an Application to Boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef]
  63. Kumar, V.R.; Rahul, M.B.; Srinivasa, R.K. Machine Learning Algorithms for Streamflow Forecasting of Lower Godavari Basin. H2Open J. 2022, 5, 670–685. [Google Scholar] [CrossRef]
  64. Robert, S. Daily Streamflow Forecasting in Mountainous Catchment Using XGBoost, LightGBM, and CatBoost. Hydrology 2022, 9, 226. [Google Scholar] [CrossRef]
  65. Li, Z.; Lu, T.; Yu, K.; Wang, J. Interpolation of GNSS Position Time Series Using GBDT, XGBoost, and RF Machine Learning Algorithms and Models Error Analysis. Remote Sens. 2023, 15, 4374. [Google Scholar] [CrossRef]
  66. Liao, Z.; Huang, Y.; Yue, X.; Lu, H.; Ju, Y. In Silico Prediction of Gamma-Aminobutyric Acid Type-A Receptors Using Novel Machine-Learning-Based SVM and GBDT Approaches. BioMed Res. Int. 2016, 2016, 1–12. [Google Scholar] [CrossRef]
  67. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased Boosting with Categorical Features. arXiv 2017, arXiv:1706.09516. [Google Scholar] [CrossRef]
  68. Zhang, L.; Janosik, D. Enhanced Short-Term Load Forecasting with Hybrid Machine Learning Models: CatBoost and XGBoost Approaches. Expert Syst. Appl. 2024, 241, 122. [Google Scholar] [CrossRef]
  69. Xu, Y.; Lin, K.; Hu, C.; Wang, S.; Wu, Q.; Zhang, J.; Xiao, M.; Luo, Y. Interpretable Machine Learning on Large Samples for Supporting Runoff Estimation in Ungauged Basins. J. Hydrol. 2024, 639, 131598. [Google Scholar] [CrossRef]
  70. Bai, T.; Wang, X.-S.; Han, P.-F. Controls of Groundwater-Dependent Vegetation Coverage in the Yellow River Basin, China: Insights from Interpretable Machine Learning. J. Hydrol. 2024, 631, 130747. [Google Scholar] [CrossRef]
  71. Zaherpour, J.; Mount, N.; Gosling, S.N.; Dankers, R.; Eisner, S.; Gerten, D. Exploring the Value of Machine Learning for Weighted Multi-Model Combination of an Ensemble of Global Hydrological Models. Environ. Model. Softw. 2019, 114, 112–128. [Google Scholar] [CrossRef]
  72. Tharwat, A.; Gabel, T. Parameters Optimization of Support Vector Machines for Imbalanced Data Using Social Ski Driver Algorithm. Neural Comput. Appl. 2020, 32, 6925–6938. [Google Scholar] [CrossRef]
  73. Dong, N.; Wei, J.; Yang, M.; Yan, D.; Yang, C.; Gao, H.; Arnault, J.; Laux, P.; Zhang, X.; Liu, Y.; et al. Model estimates of China’s terrestrial water storage variation due to reservoir operation. Water Resour. Res. 2022, 58, e2021WR031787. [Google Scholar] [CrossRef]
  74. Dong, N.; Yang, M.; Wei, J.; Arnault, J.; Laux, P.; Xu, S.; Wang, H.; Yu, Z.; Kunstmann, H. Toward improved parameterizations of reservoir operation in ungauged basins: A synergistic framework coupling satellite remote sensing, hydrologic modeling, and conceptual operation schemes. Water Resour. Res. 2023, 59, e2022WR033026. [Google Scholar] [CrossRef]
  75. Hao, H.; Dong, N.; Yang, M.; Wei, J.; Zhang, X.; Xu, S.; Yan, D.; Ren, L.; Leng, G.; Chen, L.; et al. The changing hydrology of an irrigated and dammed Yangtze River: Streamflow, extremes, and lake hydrodynamics. Water Resour. Res. 2024, 60, e2024WR037841. [Google Scholar] [CrossRef]
  76. Ibrahim, A.A.; Ridwan, R.L.; Muhammed, M.M.; Abdulaziz, R.O.; Saheed, G.A. Comparison of the CatBoost Classifier with Other Machine Learning Methods. Int. J. Adv. Comput. Sci. Appl. 2020, 11, 11. [Google Scholar] [CrossRef]
  77. Lu, D.; Peng, Y.; Xu, W.; Zhou, H.C. Application of Meteorological Ensemble Forecasts in Hydrological Field: A Review. South–North Water Transf. Water Sci. Technol. 2014, 2, 116–119, 124. [Google Scholar] [CrossRef]
  78. Sun, N.; Yan, H.; Wigmosta, M.; Leung, L.; Skaggs, R.; Hou, Z. Regional Snow Parameters Estimation for Large-Domain Hydrological Applications in the Western United States. J. Geophys. Res. Atmos. 2019, 124, 5296–5313. [Google Scholar] [CrossRef]
  79. Zhou, Y.; Xu, W.; Bai, A.; Zhang, J.; Liu, X.; Ouyang, J. Analysis of Dynamic Snowmelt Process and Its Relationship with Temperature in the Tuotuo River Region of the Tibetan Plateau. Plateau Meteorol. 2017, 36, 24–32. [Google Scholar]
  80. Bair, E.H.; Calfa, A.A.; Rittger, K.; Dozier, J. Using Machine Learning for Real-Time Estimates of Snow Water Equivalent in the Watersheds of Afghanistan. Cryosphere 2017, 12, 1579–1594. [Google Scholar] [CrossRef]
  81. Vafakhah, M.; Mohseni, S.M.; Mahdavi, M.; Alavipanah, S.K. Snowmelt Runoff Prediction by Using Artificial Neural Network and Adaptive Neuro-Fuzzy Inference System in Taleghan Watershed. Iranian J. Watershed Manag. Sci. Eng. 2011, 5, 23–35. [Google Scholar]
  82. Vafakhah, M.; Nasiri Khiavi, A.; Janizadeh, S.; Ganjkhanlo, H. Evaluating Different Machine Learning Algorithms for Snow Water Equivalent Prediction. Earth Sci. Inform. 2022, 15, 2431–2445. [Google Scholar] [CrossRef]
  83. Pan, Y.m.; Pan, Y.m.; Zhang, X.; Xue, P. A New Method of Rolling Prediction for Gas Emission Based on Wavelet Neural Network. Adv. Mater. Res. 2012, 433–440, 2288–2294. [Google Scholar] [CrossRef]
  84. Ugbah, P. Evaluation of European Centre for Medium-Range Weather Forecast; Scholars’ Press: London, UK, 2018. [Google Scholar]
  85. Song, Y.; Tsai, W.; Gluck, J.; Rhoades, A.; Zarzycki, C.; McCrary, R.; Lawson, K.; Shen, C. LSTM-Based Data Integration to Improve Snow Water Equivalent Prediction and Diagnose Error Sources. J. Hydrometeor. 2024, 25, 223–237. [Google Scholar] [CrossRef]
Figure 1. Elevation and snow distribution map of the Yalong River Basin above Ganzi: (a) the upper basin above Ganzi and its location within the Yalong River Basin; (b) snow cover distribution in the Upper Basin above Ganzi.
Figure 1. Elevation and snow distribution map of the Yalong River Basin above Ganzi: (a) the upper basin above Ganzi and its location within the Yalong River Basin; (b) snow cover distribution in the Upper Basin above Ganzi.
Sustainability 17 03779 g001
Figure 2. Data distribution in the Upper Basin above Ganzi. (a) Pre; (b) Tm (c) Tmax (d) Tmin (e) Rhu (f) Ssd (g) Win distribution in the Upper Basin above Ganzi.
Figure 2. Data distribution in the Upper Basin above Ganzi. (a) Pre; (b) Tm (c) Tmax (d) Tmin (e) Rhu (f) Ssd (g) Win distribution in the Upper Basin above Ganzi.
Sustainability 17 03779 g002aSustainability 17 03779 g002b
Figure 3. Flowchart of the snow water equivalent prediction model (n ∈ [1, 30]).
Figure 3. Flowchart of the snow water equivalent prediction model (n ∈ [1, 30]).
Sustainability 17 03779 g003
Figure 4. Comparison of evaluation metrics for the nine machine learning models during the testing period across different forecasting horizons. (a) Comparison of NSE of different machine learning in the next 1 to 10 days. (b) Comparison of NSE of different machine learning in the next 11 to 20 days. (c) Comparison of NSE of different machine learning in the next 21 to 30 days. (d) Comparison of RMSE of different machine learning in the next 1 to 10 days. (e) Comparison of RMSE of different machine learning in the next 11 to 20 days. (f) Comparison of RMSE of different machine learning in the next 21 to 30 days. (g) Comparison of RE of different machine learning in the next 1 to 10 days. (h) Comparison of RE of different machine learning in the next 11 to 20 days. (i) Comparison of RE of different machine learning in the next 21 to 30 days.
Figure 4. Comparison of evaluation metrics for the nine machine learning models during the testing period across different forecasting horizons. (a) Comparison of NSE of different machine learning in the next 1 to 10 days. (b) Comparison of NSE of different machine learning in the next 11 to 20 days. (c) Comparison of NSE of different machine learning in the next 21 to 30 days. (d) Comparison of RMSE of different machine learning in the next 1 to 10 days. (e) Comparison of RMSE of different machine learning in the next 11 to 20 days. (f) Comparison of RMSE of different machine learning in the next 21 to 30 days. (g) Comparison of RE of different machine learning in the next 1 to 10 days. (h) Comparison of RE of different machine learning in the next 11 to 20 days. (i) Comparison of RE of different machine learning in the next 21 to 30 days.
Sustainability 17 03779 g004
Figure 5. Prediction performance of the future 30-day snow water equivalent for typical periods during the testing period. AMSR is the observed snow water equivalent value during the test period. EM is the ensemble mean of three models: CatBoost, ANN, and GBDT. Ensemble refers to the prediction range of the three models: CatBoost, artificial neural network, and GBDT. Snow water equivalent forecast for the next 30 days with a reporting date of (a) 1 January 2020.; (b) 21 March 2020. (c) 26 March 2020. (d) 15 January 2022. (e) 1 March 2022. (f) 25 April 2022.
Figure 5. Prediction performance of the future 30-day snow water equivalent for typical periods during the testing period. AMSR is the observed snow water equivalent value during the test period. EM is the ensemble mean of three models: CatBoost, ANN, and GBDT. Ensemble refers to the prediction range of the three models: CatBoost, artificial neural network, and GBDT. Snow water equivalent forecast for the next 30 days with a reporting date of (a) 1 January 2020.; (b) 21 March 2020. (c) 26 March 2020. (d) 15 January 2022. (e) 1 March 2022. (f) 25 April 2022.
Sustainability 17 03779 g005aSustainability 17 03779 g005b
Figure 6. Distribution of influential factors on the snow water equivalent. SHAP value reflects the importance of different elements. AMSRt represents the snow water equivalent data for the forecast start date in the model input data.
Figure 6. Distribution of influential factors on the snow water equivalent. SHAP value reflects the importance of different elements. AMSRt represents the snow water equivalent data for the forecast start date in the model input data.
Sustainability 17 03779 g006
Figure 7. SHAP value distribution for SWE sensitivity.
Figure 7. SHAP value distribution for SWE sensitivity.
Sustainability 17 03779 g007
Figure 8. SHAP values of the influence of different climatic factors on the SWE: (a) average temperature; (b) minimum temperature; (c) maximum temperature; (d) relative humidity; (e) sunshine hours; (f) wind speed.
Figure 8. SHAP values of the influence of different climatic factors on the SWE: (a) average temperature; (b) minimum temperature; (c) maximum temperature; (d) relative humidity; (e) sunshine hours; (f) wind speed.
Sustainability 17 03779 g008aSustainability 17 03779 g008b
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, J.; Yang, M.; Dong, N.; Wang, Y. Machine-Learning-Based Ensemble Prediction of the Snow Water Equivalent in the Upper Yalong River Basin. Sustainability 2025, 17, 3779. https://doi.org/10.3390/su17093779

AMA Style

Zhang J, Yang M, Dong N, Wang Y. Machine-Learning-Based Ensemble Prediction of the Snow Water Equivalent in the Upper Yalong River Basin. Sustainability. 2025; 17(9):3779. https://doi.org/10.3390/su17093779

Chicago/Turabian Style

Zhang, Jujia, Mingxiang Yang, Ningpeng Dong, and Yicheng Wang. 2025. "Machine-Learning-Based Ensemble Prediction of the Snow Water Equivalent in the Upper Yalong River Basin" Sustainability 17, no. 9: 3779. https://doi.org/10.3390/su17093779

APA Style

Zhang, J., Yang, M., Dong, N., & Wang, Y. (2025). Machine-Learning-Based Ensemble Prediction of the Snow Water Equivalent in the Upper Yalong River Basin. Sustainability, 17(9), 3779. https://doi.org/10.3390/su17093779

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop