Next Article in Journal
Research on Soil Inorganic Nitrogen Detection Technology Based on Dielectric Response
Previous Article in Journal
Balancing Poverty Alleviation and Ecosystem Vulnerability Reduction: Implication from China’s Targeted Interventions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Natural Gas Futures Price Prediction Based on Variational Mode Decomposition–Gated Recurrent Unit/Autoencoder/Multilayer Perceptron–Random Forest Hybrid Model

School of Mathematics and Statistics, Ludong University, Yantai 264001, China
*
Author to whom correspondence should be addressed.
Sustainability 2025, 17(6), 2492; https://doi.org/10.3390/su17062492
Submission received: 3 February 2025 / Revised: 3 March 2025 / Accepted: 10 March 2025 / Published: 12 March 2025
(This article belongs to the Section Energy Sustainability)

Abstract

:
Forecasting natural gas futures prices can help to promote sustainable global energy development, as the efficient use of natural gas as a clean energy source has become key to the growing global demand for sustainable development. This study proposes a new hybrid model for the prediction of natural gas futures prices. Firstly, the original price series is decomposed, and the subsequences, along with influencing factors, are used as input variables. Secondly, the input variables are grouped based on their correlations with the output variable, and different models are employed to forecast each group. A gated recurrent unit (GRU) captures the long-term dependence, an autoencoder (AE) downscales and extracts the features, and a multilayer perceptron (MLP) maps the complex relationships. Subsequently, random forest (RF) integrates the results of the different models to obtain the final prediction. The experimental results show that the model has a mean absolute error (MAE) of 0.32427, a mean absolute percentage error (MAPE) of 10.17428%, a mean squared error (MSE) of 0.46626, a root mean squared error (RMSE) of 0.68283, an R-squared (R²) of 93.10734%, and an accuracy rate (AR) of 89.82572%. The results demonstrate that the proposed decomposition–selection–prediction–integration framework reduces prediction errors, enhances the stability through multiple experiments, improves the prediction efficiency and accuracy, and provides new insights for forecasting.

1. Introduction

By accurately forecasting natural gas futures prices, it is possible not only to provide decision-making support for energy market participants [1,2] but also to enhance the market competitiveness of natural gas as a clean energy source, thus promoting the sustainable optimization of the energy structure and the development of a low-carbon economy [3]. However, the prediction of natural gas futures prices presents various challenges. To effectively tackle challenges such as non-stationarity and noise in data, scholars have introduced the concept of “decomposition–integration” [4]. This approach decomposes complex price time series into multiple simpler and more manageable subsequences [5]. Each subsequence is then forecast independently [6], and their predictions are integrated to obtain the final forecast value [7]. Specifically, techniques like the completely adaptive ensemble empirical modal decomposition of noise [8], the improved ensemble empirical modal decomposition (IEEMD) of noise [9], and variational modal decomposition (VMD) [10], among others, can decompose natural gas futures price time series into modal components with different frequency characteristics. The advantages of the decomposition–integration approach include the following: decomposing complex problems to enhance the prediction performance and reduce error accumulation [11]; handling complex data issues such as nonlinearity, randomness, and high noise levels [12]; improving the computational efficiency by processing in frequency bands to enhance the interpretability [13]; and enhancing models’ accuracy and robustness [14].
When constructing forecasting models [15], various factors influencing natural gas futures prices are thoroughly considered [16,17], including but not limited to historical price data [18], global economic indicators [19], policy changes [20], unforeseen events, and energy consumption trends. Related indicators such as volatility [21] and market sentiment indicators [22,23] are also incorporated. These factors are appropriately processed to serve as input features for the model [24], thereby enhancing its predictive ability and robustness [25]. The advantages of incorporating influencing factors include gaining a deeper understanding of the data relationships [26]; supporting multivariate analysis, applicable to multi-input single-output model forms [27]; providing rich information, where how each influencing factor addresses different forecasting needs [28,29]; and integrating multiple methodologies [30,31].
Hybrid models combine advanced algorithms and tools such as long short-term memory (LSTM) [32], autoencoders (AEs), autoregressive integrated moving average (ARIMA) [33,34], gated recurrent units (GRUs) [35], convolutional neural networks (CNNs) [36], generalized autoregressive conditional heteroskedasticity (GARCH) [37], multilayer perceptron (MLP), random forest (RF), and others for final prediction purposes [38,39]. The advantages of hybrid models include enhancing models’ robustness and stability against noise [40], offering flexible combinations with multiple choices [41], conducting multidimensional analyses to overcome the limitations of single models, and being widely applicable to address complex issues [42].
To address frequent unexpected events in the energy market, such as global pandemics, geopolitical conflicts [43], and natural disasters, models incorporate dynamic adjustment mechanisms and anomaly detection algorithms to capture the real-time impacts of these events on prices and make corresponding adjustments, thereby enhancing the model’s predictive capabilities under extreme conditions [44]. To comprehensively reflect the dynamic changes in natural gas futures prices, scholars typically categorize the input features based on their varying degrees of influence on the prices into input feature sets. These include, but are not limited to, historical prices [45], trading volumes, open interest [46], macroeconomic indicators [47], seasonal factors, and market sentiment indices extracted through natural language processing [48]. During model training and testing [49], factors such as the prediction periods, time steps [50], and sequence length segmentation ratios also influence the prediction outcomes. Adjusting these periods or steps can aid in the study of the impact of different prediction windows on model performance to find optimal prediction strategies [51].
To objectively assess model performance, new models are typically compared with traditional models in terms of predictive performance [52]. A comparative analysis using evaluation metrics highlights the significant advantages of the new model in terms of prediction accuracy, stability, and generalization capabilities [53]. Comparing different combinations of models and experimental results helps to validate the accuracy of the proposed model, identifies deficiencies in the control group models [54], optimizes innovative settings [55], and promotes interdisciplinary integration.
Table 1 compares the different methods used in the studies cited in this paper for the forecasting of futures prices; these are divided into decomposition and modeling categories, with the decomposition category mainly used to split the price series and enrich the data features; the modeling category is mainly applicable when learning the patterns of price changes and is used for forecasting.
The VMD-GRU/AE/MLP-RF hybrid model proposed in this study is significantly innovative in time series forecasting. The research framework is shown in Figure 1. The model adopts the “decomposition–selection–forecasting–integration” framework, which not only includes the traditional decomposition and forecasting steps but also introduces a grouping strategy based on the correlation coefficients and model integration to systematically deal with the input variables. Through VMD decomposition, the time series is split into multiple subseries with different frequencies to reduce the complexity. The correlation coefficient grouping strategy quickly filters important variables in the preprocessing stage, reduces the input dimensions, and is more efficient in computation compared to the attention mechanism and SHAP values, making it suitable for large-scale data. In terms of model combination, the pairing of the GRU, AE, and MLP gives full play to the advantages of each model: the GRU efficiently captures the long-term dependence of the time series and has a concise structure; the AE reduces the dimensionality and extracts the key features to remove noise; and the MLP performs nonlinear combination to enhance the expressive ability. This combination ensures a balance between computational efficiency and performance, satisfying the requirement for low computational complexity and small data volumes, being suitable for long-series prediction. Finally, the prediction results are integrated via random forest to further improve the model’s robustness and accuracy. The experiments verify the effectiveness of the model and prove that it has significant advantages in terms of accuracy and robustness [56].
This article is structured as follows. Section 2 introduces the theoretical framework and the proposed models used in the study. Section 3 covers the data preprocessing, introduces the evaluation metrics, and discusses the grouping of the input variables. Section 4 presents the experimental correlations and compares the results. Section 5 concludes the study.

2. Methodology

2.1. Variational Mode Decomposition

Variational mode decomposition is a signal processing method. It decomposes complex signals into multiple intrinsic mode functions (IMFs), each with specific frequency characteristics and stability. The formula is given by (1), where A k ( t ) represents the instantaneous amplitude and ϕ k ( t ) represents the instantaneous phase.
s k ( t ) = A k ( t ) cos ( ϕ k ( t ) ) .

2.2. Pearson Correlation Coefficient

The Pearson correlation coefficient is a widely used statistical measure in statistics, assessing the linear correlation between two variables X and Y. It is represented by (2). The coefficient ranges within [−1, 1], where a value closer to 1 indicates a strong positive correlation between the variables, a value closer to 0 indicates a weak correlation, and a value closer to −1 indicates a strong negative correlation.
ρ X , Y = cov ( X , Y ) σ X σ Y .

2.3. Gated Recurrent Unit

The gated recurrent unit (GRU) is a popular type of recurrent neural network architecture. It retains much of LSTM’s effectiveness while reducing the model complexity and computational burden. The GRU consists of two gate units, the update gate and the reset gate, defined by (3) and (4), respectively. Here, W and b denote the weights and biases of the gates, σ represents the sigmoid activation function, h t 1 is the previous hidden state, x t is the current input, and · , · denotes vector concatenation.
z t = σ ( W z · h t 1 , x t + b z ) .
r t = σ ( W r h t 1 , x t + b r ) .

2.4. Autoencoder

An autoencoder (AE) is a type of neural network model; it is used primarily for unsupervised or semi-supervised learning tasks such as data dimensionality reduction, compression, and feature extraction. The AE consists of two main components: the encoder and the decoder. The encoder maps the input data to a lower-dimensional encoded space, as shown in (5), while the decoder attempts to reconstruct the original data from the encoded space, as shown in (6). Here, f and g represent nonlinear transformation functions in the encoder and decoder, respectively.
z = f ( x ) .
x = g ( z ) .

2.5. Multilayer Perceptron

A multilayer perceptron (MLP) is a type of feedforward artificial neural network model that consists of at least three layers of nodes: an input layer, one or more hidden layers, and an output layer. In an MLP, input signals pass through the input layer, are weighted and transformed nonlinearly by neurons in the hidden layers, and finally reach the output layer to produce prediction results. The equations for the MLP are shown in (7) and (8), where u denotes the weighted sum and y represents the result.
u = i = 1 n w i x i + b .
y = f ( u ) .

2.6. Random Forest

Random forest (RF) is a type of machine learning model based on the ensemble learning principle. It enhances the prediction accuracy and stability by constructing multiple decision trees and combining their predictions. The formula is shown in (9), where w i represents individuals and h i represents the weights corresponding to predictions.
H ( x ) = 1 T i = 1 T w i h i ( x ) .

2.7. VMD-GRU/AE/MLP-RF Hybrid Model

This study constructs a hybrid model named VMD-GRU/AE/MLP-RF. Figure 2 illustrates the framework of the VMD-GRU/AE/MLP-RF hybrid model, with the following procedural steps.
Step 1. Data Collection. Collect raw data including daily natural gas futures: open, close, high, low prices, trading volume, amplitude, production, import volume, export volume, and storage volume. Output variable is the median price between the open and close prices.
Step 2. Decomposition. Use the VMD model to decompose price data sequences, determining the optimal number of modes. This produces 5 intrinsic mode functions (IMFs) and 1 residual (Res), with the 5 IMFs serving as new input variables.
Step 3. Feature Selection. Employ Pearson correlation coefficients to filter variables. Sort input variables by their correlation coefficients with the price in descending order.
Step 4. Grouping. Divide variables into groups based on correlation coefficient intervals. Strong correlation group: open, close, high, low prices, IMF5. Moderate correlation group: import volume, IMF4, IMF3. Weak correlation group: IMF2, IMF1, amplitude. Negative correlation group: trading volume, production, export volume, storage volume
Step 5. Prediction. Predict results for different correlation groups using different models. Use GRU, AE, MLP to predict strong, moderate, and weak correlation groups based on evaluation metrics. Generate 3 sets of prediction sub-results.
Step 6. Integration. Integrate sub-results using RF to obtain the final prediction result. Train decision trees with bootstrap sampling from the three sets of sub-results. Weight integration of sub-results is performed with weights of a, b, c as defined in (10), aimed at reducing randomness and incorporating the data characteristics of the sub-results.
Y = a y 1 + b y 2 + c y 3

3. Data and Preliminary Analysis

3.1. Data Source

The data for this study are sourced from the official website of the U.S. Energy Information Administration (EIA), including variables such as the opening, closing, high, and low prices; trading volume; amplitude; production; import volume; export volume; and storage volume. For simplicity, the output variable, the original price sequence, is taken as the median between the open and close prices. The dataset spans from 4 April 1990 to 28 May 2024. The opening price, closing price, high price, low price, trading volume, and amplitude are daily frequency data, which are recorded as a daily time unit. Meanwhile, production, imports, exports, and storage are monthly frequency data, which are recorded in each month as a time unit. In order to unify the time frequencies, we split the monthly frequency data into daily frequency data: this is achieved by copying the monthly frequency data of a particular month to each day of the month. In this way, the monthly frequency data are converted into daily frequency data, resulting in a total of 8706 data points.

3.2. Data Preprocessing

3.2.1. Handling Missing Values

The collected dataset includes 11 variables over 8706 time points with daily frequency data. Missing values are handled using linear interpolation, defined by (11).
y = y 0 + x x 0 y 1 y 0 x 1 x 0 .

3.2.2. Normalization

Min–max normalization is applied in this study. Due to the vast differences in the ranges of price-related variables (very small) and quantity-related variables (very large), normalization helps to simplify calculations and improve the computational efficiency. Normalized data are mapped to the range [0, 1], as shown in (12).
x = x x min x max x min .

3.2.3. Data Splitting

The dataset is split into training and testing sets in an 85%/15% ratio. The training set consists of 7400 data points, covering the period from 4 April 1990 to 18 June 2019. The testing set comprises 1306 data points, spanning from 19 June 2019 to 28 May 2024.

3.3. Evaluation Metrics

This study evaluates the predictive performance of the model using several metrics: the mean absolute error (MAE), mean absolute percentage error (MAPE), mean squared error (MSE), root mean squared error (RMSE), coefficient of determination ( R 2 ), and accuracy rate (AR). Among these, smaller values of the MAE, MAPE, MSE, and RMSE indicate the stronger predictive capabilities of the model. A value of R 2 and AR closer to 1 signifies the better fit of the model to the data. The formulas for these evaluation metrics are given in (13)–(18).
M A E = 1 n t = 1 n y ^ t y t .
M A P E = 100 % n t = 1 n y ^ t y t .
M S E = 1 n t = 1 n y ^ t y t 2 .
R M S E = 1 n t = 1 n y ^ t y t 2 .
R 2 = 1 t = 1 n y t y ^ t 2 t = 1 n y t y ¯ t 2 .
A R = 1 100 % n t = 1 n y ^ t y t y t .

3.4. Decomposition

This study employs the VMD model to decompose the original price sequence into five sub-series and a residue based on the stability of the central frequencies, as illustrated in Figure 3.
Table 2 shows the center frequency distribution of VMD decomposition under different modal numbers. From the table data, it can be seen that, when the number of IMFs is increased to 5, the center frequency shows a more stable distribution, and the change in the center frequency is not significant when the number of IMFs is further increased, which suggests that the decomposed modal components can effectively separate out the different frequency components in the signal without modal overlapping or over-decomposition when K = 5. Too many modal components will increase the computational complexity, while too few modal components may not be able to fully capture the details of the signal. Therefore, we choose to decompose the original price series into 5 IMFs and 1 residual.
In Figure 3, the original price sequence is represented by the black line, the sub-series by blue lines, and the residue by the red line. This decomposition separates the characteristics of the original price sequence, enabling multiscale analysis. The frequency decreases from IMF1 to IMF5: high-frequency sub-series such as IMF1 and IMF2 reflect short-term rapid fluctuations in the price sequence; medium-frequency sub-series like IMF3 contain transitional information from high to low frequencies; and low-frequency sub-series like IMF4 and IMF5 capture long-term trends and major features of the original sequence.

3.5. Variable Selection Based on Correlation

Figure 4 illustrates the correlation coefficients between variables obtained using the Pearson correlation method. The coefficients range within [−0.2, 1], where a deeper red indicates a stronger positive correlation and a deeper blue indicates a stronger negative correlation between the variables. Lighter colors indicate weaker correlations.
From Figure 4, it is observed that the variables with strong correlations include the open price, close price, high price, low price, IMF5, and the price itself, all exhibiting correlation coefficients above 0.9 due to their reflection of price data changes. Additionally, the import volume shows a correlation coefficient of 0.66 with price-related variables. Strong correlations are also evident among the production, storage volume, and trading volume. Most variables exhibit weak correlations, with a few variables displaying negative correlations, such as the production and import volume, reflecting an inverse relationship where increased production enhances the supply and may weaken the import volume.
We translate the last row or column data from the correlation matrix into groups to generate Table 3, displaying the correlation levels between different variables and prices. The variables are grouped based on their correlation strength with prices into the strong correlation, moderate correlation, weak correlation, and negative correlation groups. The strong correlation group’s correlation coefficients range within (0.7, 1]; it contains variables that are highly correlated with the target variable and can be considered almost linear. The medium correlation group’s correlation coefficients range within (0.1, 0.7]; it contains variables that have some correlation but not enough to be considered strongly correlated. This interval was chosen to capture variables that may have a potential effect on the target variable but are not significant. The weak correlation group’s range of correlation coefficients is (0, 0.1], which indicates that the variable is correlated with the target variable but at a low level. For the negative correlation group, the correlation coefficients range within [−1, 0].The correlation coefficient of 0.17683 for IMF3 indicates a weak correlation but is still above the threshold for weak correlations and therefore it cannot be ignored completely. Its categorization into the medium correlation group is intended to retain this part of the potential impact in the model and avoid losing important information due to over-screening.

4. Results

4.1. Model Selection Experiment

Due to the different characteristics of the variable data in the different correlation groups, the optimal predictive model may vary. Therefore, this section applies GRU, MLP, RNN, CNN, AE, LSTM, and MLP models to predict the three correlation groups. By comparing the prediction errors, the most suitable models for each correlation group are selected. Table 4, Table 5 and Table 6 present the evaluation metrics for prediction with various models for each correlation group. The upper part shows numerical values, and the lower part ranks the models based on the prediction errors under each evaluation metric.
Table 4 displays the prediction errors and rankings for the strong correlation group using different models. Considering all factors, the GRU model is selected for the prediction of data in the strong correlation group.
Table 5 evaluates various models for the prediction of variables in the medium correlation group. The AE model is chosen based on comprehensive consideration for the prediction of data in the medium correlation group.
Table 6 evaluates various models for the prediction of variables in the weak correlation group. The MLP model is selected based on comprehensive consideration for the prediction of data in the weak correlation group.
After an internal comparison within Table 4, Table 5 and Table 6 to select models suitable for the different correlation groups, vertical comparisons across these tables reveal that the prediction effectiveness is the highest for the strong correlation group, achieving an accuracy rate of over 90%, while the other two groups have accuracies of only 50% to 60%.

4.2. Robustness Test

The robustness of the samples is tested by dividing them into in-sample training and out-of-sample prediction groups. In-sample training involves using the model to train on 85% of the training set data and predict the results. Out-of-sample prediction involves using the model to predict on the remaining 15% test set data. To eliminate randomness, each experiment is conducted 10 times, and the average values are shown in Table 7.
Table 7 compares the results of in-sample training and out-of-sample prediction. They are close, but in-sample training involves optimizing the model by adjusting the parameters and structure, while out-of-sample prediction tests the model’s performance on entirely new data, thus being slightly inferior to in-sample training. Figure 5 illustrates a comparison of 10 predictions each for in-sample training and out-of-sample prediction. It shows that the in-sample training results closely align with the actual values, especially when the data undergo sudden changes, whereas the out-of-sample prediction results show greater deviations from the actual values.
Table 8 demonstrates the in-sample and out-of-sample variances and confidence intervals for the 10 experiments. The in-sample variance ranges from 0.3753 to 0.4537 with small fluctuations, with Pred6 being the most stable and Pred3 the most volatile. The confidence intervals are narrower, such as for Pred6, which has a small and precise prediction error, and Pred9, which is also narrower but close to zero, indicating a concentrated distribution of errors. The out-of-sample variance ranges from 0.3504 to 0.4494, with slightly larger fluctuations than in the in-sample case. The confidence intervals are generally wide and mostly negative, such as [−0.1141, −0.0412] for Pred1, indicating a wide margin of error and the possible underestimation of the actual value. Overall, the degree of volatility is similar inside and outside the sample, but the stability outside the sample is slightly lower.

4.3. Correlation Comparison Experiment

The models selected in Section 4.1 are used to predict the strong, medium, and weak correlation groups separately. Specifically, the GRU is used to predict the strong correlation group’s test set data, the AE for the medium correlation group’s test set data, and the MLP for the weak correlation group’s test set data. The prediction results are shown in Figure 6 and Table 9.
Figure 6 depicts the prediction results for the different correlation groups. The black line represents the actual natural gas futures prices, the blue line represents the predictions for the strong correlation group using the GRU model, the red line represents the predictions for the medium correlation group using the AE model, and the yellow line represents the predictions for the weak correlation group using the MLP model. It can be observed that the blue line closely overlaps with the black line, with slight variations during changes. The red line reflects the long-term trend of the black line’s variations, while the yellow line reflects the short-term fluctuations in the black line’s variations. Thus, the predictions for the strong correlation group closely match the actual data, the predictions for the medium correlation group reflect the long-term characteristics of the actual data, and the predictions for the weak correlation group reflect short-term data fluctuations.
Table 9 presents the error results for the three correlation groups. The first four evaluation metrics (MAE, MAPE, MSE, RMSE) should be minimized, while the last two ( R 2 , AR) should approach 1 for better performance. Comparing the MAE, MAPE, MSE, and RMSE among the three groups reveals that the prediction errors for the medium and weak correlation groups are approximately 10 times higher than those for the strong correlation group. The MAE, MAPE, MSE, and RMSE for the medium and weak correlation groups are quite similar, indicating significantly reduced prediction accuracy. Comparing the R 2 and AR values, the prediction error for the strong correlation group approaches 1, indicating very high accuracy and a negligible difference from the actual data. In contrast, the prediction errors for the medium and weak correlation groups are approximately half those for the strong correlation group, indicating similar results with significantly reduced prediction accuracy. Based on the comparison of the prediction error values, the prediction accuracy for the strong correlation group is ten times higher than that for the other two groups, and the accuracy for the medium and weak correlation groups is reduced by half compared to the strong correlation group.
Figure 7 shows the convergence curve of the GRU model with 20 iterations and a loss function of MSE. The strong correlation group dataset is used, of which 85% is used as the training set and then 20% is separated from it as the validation set. The loss values for both the training set (blue line) and the validation set (red line) decrease with the number of iterations, decreasing rapidly at the beginning and slowing down at the end, indicating that the model gradually converges. The loss of the validation set is lower than that of the training set, showing that the model is not overfitted, has a strong generalization ability, learns data features effectively, and performs well on new data.
According to the test set evaluation error metrics in Table 9, the errors of the medium correlation group and the weak correlation group in the first four metrics are about 10 times those of the strong correlation group, while the errors of the medium correlation group and the weak correlation group in the last two metrics are only half those of the strong correlation group. This suggests that the strong correlation group has significantly higher prediction accuracy for most indicators, especially for key indicators, where the errors are much lower than those of the other groups. Therefore, the strong correlation group is assigned an 80% weight to ensure that it plays a dominant role in the overall prediction. The medium correlation group and the weak correlation group also contribute to some indicators, but their overall performance is not as strong as that of the strong correlation group, and their error values are not very different, so each group is assigned a 10% weight. The final weighting ratio is 8:1:1, which highlights the advantages of the strong correlation group and retains the contributions of the other groups to ensure that the prediction results are mainly driven by the strong correlation group, while taking into account the overall prediction ability of the model. The formula for the determination of the weights is given in (19).
Y = 0.8 y 1 + 0.1 y 2 + 0.1 y 3

4.4. Ablation Experiment

To demonstrate the effectiveness of this model, four sets of experiments were designed with each group, removing one of the correlation groups to assess its overall impact. The purpose of the “strong + medium” group, “strong + weak” group, and “medium + weak” group was to test the effects of the weak, medium, and strong correlation groups, respectively, on the prediction. The “full model” group included input variables from all correlation groups, aiming to evaluate the optimization level of the model.
Figure 8 displays the prediction results for the first three experimental groups. The black line represents the actual values, the blue line represents the predictions using the strong + medium group variables, the red line represents the predictions using the strong + weak group variables, and the green line represents the predictions using the medium + weak group variables. From Figure 8, it can be observed that the blue line deviates the least from the black line, indicating the highest accuracy. The green line shows the greatest deviation, especially at price inflection points, while the red line falls between the other two. This suggests that the predictions are the least accurate with the medium + weak group variables, followed by the strong + weak group, and they are the most accurate with the strong + medium group variables. This also confirms that the GRU predictions for the strong correlation group contribute the most to the predictions, whereas the predictions using the AE for the medium correlation group and the MLP for the weak correlation group are less effective and similar in performance.
Table 10 presents the error evaluation metrics for the four experimental groups in the sensitivity analysis. Comparing the MAE, MAPE, MSE, and RMSE, the medium + weak group exhibits the highest errors, ranging from 2 to 10 times higher than the other experiments. The errors for the strong + weak, strong + medium, and full model groups are relatively close, but, overall, the full model outperforms the strong + weak group, which in turn outperforms the strong + medium group. Comparing the R 2 and AR across the four groups, the predictions for the medium + weak group are significantly less accurate than those of the other three groups, while the remaining three groups show similar performance, with the full model being superior to the strong + medium group, which is superior to the strong + weak group.
The ablation experiment demonstrates that the full model group achieves the highest prediction accuracy and best overall performance. Within the predictive model structure, the GRU predictions for the strong correlation group are identified as the most valuable, followed by the predictions using the AE for the medium correlation group and the MLP for the weak correlation group.

4.5. Step Length Experiment

This section explores the impact of the time step length on the prediction results. Three experimental groups were set up with sliding time windows of 3, 5, 7, 9, 11, and 13 time steps for the input sequences, while keeping the other settings constant.
In Figure 9, the black line represents the actual data, and the blue solid line, red solid line, light blue solid line, green solid line, purple solid line, and yellow solid line represent the sliding time windows of 3, 5, 7, 9, 11, and 13 steps, respectively. The light blue line follows the black line, indicating that the predicted value for the seven-step window is closest to the actual value. The largest deviations from the black line are the blue and yellow lines, indicating that the 3- and 13-step prediction windows are the least accurate, and that a too long or too short window affects the prediction results. Combined with Table 11 (which lists the errors in detail), it can be seen that the prediction error decreases and then increases with the increase in the time step, which indicates that a time window with a certain length will allow the prediction to generate the optimal results, an increase in the optimal step will lead to redundancy in the amount of data, and the model will not be able to find the key points in learning the data. Moreover, a decrease in the optimal step will lead to a reduction in the amount of data and reduce the feature richness of the data, and the model will have limited access to information, leading to a decrease in the prediction performance. From the results in Table 11, the optimal step size range is around 7.

4.6. Experiment with Addition of Negative Variable

This section explores the predictive effects of variables negatively correlated with prices. A control experiment is set up where one group uses the predictions from this model, while another group adds the negatively correlated variables as input variables to this model. Section 4.1 compares the prediction errors of the negatively correlated group using the MLP model for the prediction of negatively correlated variables.
Figure 10 presents the prediction results for the experimental and control groups. The black line represents the actual results, the red line represents the predictions from the experimental group using this model, and the blue line represents the predictions from the control group after adding negatively correlated group variables. From Figure 10, it can be seen that both experimental groups have varying degrees of error compared to the actual results. The vertical colored lines in Figure 10 indicate the magnitude of the errors, with the addition of negatively correlated group variables resulting in decreased prediction accuracy, demonstrating that factors that are negatively correlated with the output variable reduce the prediction accuracy.
Table 12 shows the prediction errors for the two experimental groups considered in this section. Comparing these results reveals that adding negatively correlated group variables significantly increases the prediction errors compared to the full model experimental group without these variables, thus validating the hypothesis.

5. Conclusions and Discussion

This paper proposes a VMD-GRU-AE-MLP-RF hybrid model for the forecasting of natural gas futures. The approach involves decomposing the price original series using VMD and utilizing the decomposed sub-series along with influencing factors as the input variables. By conducting a correlation analysis, the input variables are categorized into groups based on their degrees of correlation. Different prediction models—the GRU, AE, and MLP—are then applied to forecast each group according to their respective characteristics. Finally, an RF model integrates the predictions from the three groups of sub-results to generate the final forecast, effectively enhancing the prediction accuracy and reducing forecasting errors.
The innovation of this study lies in enriching the input variable information by incorporating sub-series with different frequencies and influencing factors. Furthermore, the adaptation of different prediction models to different input dimensions enhances the model flexibility. The use of the RF model to integrate diverse prediction outputs is another innovative aspect, contributing to improved forecasting performance. Additionally, we paper conducted numerous experiments to empirically analyze and validate the proposed model, and, according to the evaluation indices, all of them achieve good prediction results. For example, in the case of the RMSE, the best result in the model selection experiment was 0.1695; the best result in the stability test was 0.6328; the best result in the correlation experiment was 0.1695; the best result in the ablation experiment was 0.7792; the best result in the time span experiment was 0.6503; and the best result in the all-variable experiment was 0.6828. These experimental results effectively prove the feasibility and validity of the model.
In contrast to previous research methods, the model proposed in this study no longer relies solely on a single advantage, such as enriching data features only through serial decomposition methods, adding only factors that affect prices to aid in forecasting, or relying only on advanced models to improve the forecasting accuracy. Instead, this model employs a variety of advanced algorithms to optimize every aspect of the analytical framework.
In terms of findings, the model screening test employs eight different models to predict three groups of input variables with different degrees of correlation and selects the model that best matches the input variables through the prediction error. This method is more flexible and adaptable than the use of a single model to predict the correlation groups of the input variables. Robustness experiments were also used to comprehensively examine both in-sample and out-of-sample data, and the results showed that this multilink model prediction structure was able to output more stable results, further proving its superiority. The correlation and ablation experiments deeply analyzed the prediction results of different models for the corresponding correlation groups, and the results showed that the high correlation group had the best prediction effect due to its close correlation with the price; the prediction results for the medium correlation group mainly presented the long-term trends of the data; and the prediction results of the low correlation group reflected the short-term trends of the data. By presenting the different information contained in the prediction results in groups, the present study provides a superior perspective for an understanding of the internal logic of the prediction results. In addition, the step size experiment explored the effect of the length of the sliding time window on the prediction results, and it was found that the longer the sliding time window, the richer the historical information included, but too long a window will instead make it difficult for the model to recognize important information. This type of experiment is less common in previous studies, and it provides a new perspective for the design of the prediction model. Finally, a negative variable experiment was conducted by adding negative correlation group data to the input variables, and the results showed that deleting the negative correlation group data could significantly improve the prediction accuracy. This indicates that eliminating inappropriate input variables can optimize the prediction effect. Most of the previous research has “added” the input variables through subsequence decomposition or adding related factors, while this study further optimized the quality of the input variables through “subtracting” the input variables, which provides a direction for the improvement of prediction models.
The hybrid VMD-GRU-AE-MLP-RF model proposed in this paper shows high accuracy and flexibility in natural gas futures forecasting, but there are some limitations. First, the number of subsequences generated in the price decomposition part affects the input variable grouping. Second, changes in the weights assigned to the final forecasts of each correlation group of the input variables can also affect the final results, but the current model does not yet provide a mechanism to dynamically adjust the weights.
The future research directions include the following: firstly, relevant algorithmic mechanisms can be introduced to adapt to different data sets by assigning weights according to the prediction error; secondly, the data characteristics can be enriched by introducing more influencing factors.
Natural gas futures prediction modeling not only aids in market decision-making but also provides support for sustainable development goals. By accurately predicting the natural gas price trend, the model can provide data support for policymakers to optimize the energy structure and reduce the dependence on traditional fossil fuels. In addition, the optimization directions of the model, such as dynamic weight adjustment and multimodal data fusion, are in line with the requirements of sustainable development in terms of flexibility and adaptability. Future research could further explore the application of the model in energy transition and ecosystem protection and promote the deep integration of natural gas futures forecasting with sustainable development goals.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/su17062492/s1.

Author Contributions

Conceptualization, H.Y. and S.S.; Methodology, H.Y. and S.S.; Software, S.S.; Formal analysis, H.Y. and S.S.; Investigation, H.Y. and S.S.; Resources, H.Y.; Writing—original draft, S.S.; Writing—review and editing, H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

The authors would like to acknowledge the support of the Social Science Planning Project of Shandong Province (22CSDJ13); the Shandong Province Undergraduate Teaching Reform Research Key Project entitled Learning and Doing, Fusing and Creating, Knowing and Doing: Reform and Practice of Cultivating Talents’ Applied Innovation Ability of Statistics Majors in Local Colleges and Universities in the Era of Big Data (Z2021274); and a project to study the effect of digital economy and eco-efficiency correlation in the urban agglomeration of the Yellow River Basin (IPGS2024-044).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Hariram, N.; Mekha, K.; Suganthan, V.; Sudhakar, K. Sustainalism: An integrated socio-economic-environmental model to address sustainable development and sustainability. Sustainability 2023, 15, 10682. [Google Scholar] [CrossRef]
  2. Kuhlman, T.; Farrington, J. What is sustainability? Sustainability 2010, 2, 3436–3448. [Google Scholar] [CrossRef]
  3. Holechek, J.L.; Geli, H.M.; Sawalhah, M.N.; Valdez, R. A global assessment: Can renewable energy replace fossil fuels by 2050? Sustainability 2022, 14, 4792. [Google Scholar] [CrossRef]
  4. Wu, H.; Li, P.; Cao, J.; Xu, Z. Forecasting the Chinese crude oil futures volatility using jump intensity and Markov-regime switching model. Energy Econ. 2024, 134, 107588. [Google Scholar] [CrossRef]
  5. Wang, J.; Liu, Y.; Li, Y. A parallel differential learning ensemble framework based on enhanced feature extraction and anti-information leakage mechanism for ultra-short-term wind speed forecast. Appl. Energy 2024, 361, 122909. [Google Scholar] [CrossRef]
  6. Lu, B.; Ma, F.; Wang, J.; Ding, H.; Wahab, M.I.M. Harnessing the decomposed realized measures for volatility forecasting: Evidence from the US stock market. Int. Rev. Econ. Financ. 2021, 72, 672–689. [Google Scholar] [CrossRef]
  7. Hong, Y.; Qu, S. Beyond Boundaries: The AHP-DEA Model for Holistic Cross-Banking Operational Risk Assessment. Mathematics 2024, 12, 968. [Google Scholar] [CrossRef]
  8. Wang, J.; Sun, X.; Cheng, Q.; Cui, Q. An innovative random forest-based nonlinear ensemble paradigm of improved feature extraction and deep learning for carbon price forecasting. Sci. Total Environ. 2021, 762, 143099. [Google Scholar] [CrossRef]
  9. Yang, K.; Zhang, X.; Luo, H.; Hou, X.; Lin, Y.; Wu, J.; Yu, L. Predicting energy prices based on a novel hybrid machine learning: Comprehensive study of multi-step price forecasting. Energy 2024, 298, 131321. [Google Scholar] [CrossRef]
  10. Huang, Y.; Dai, X.; Wang, Q.; Zhou, D. A hybrid model for carbon price forecasting using GARCH and long short-term memory network. Appl. Energy 2021, 285, 116485. [Google Scholar] [CrossRef]
  11. Du, P.; Wang, J.; Hao, Y.; Niu, T.; Yang, W. A novel hybrid model based on multi-objective Harris hawks optimization algorithm for daily PM2.5 and PM10 forecasting. Appl. Soft Comput. 2020, 96, 106620. [Google Scholar] [CrossRef]
  12. Chen, L.; Zhao, X. A multiscale and multivariable differentiated learning for carbon price forecasting. Energy Econ. 2024, 131, 107353. [Google Scholar] [CrossRef]
  13. Yan, Z.; Tian, F.; Sun, Y.; Wang, S. A time-frequency-based interval decomposition ensemble method for forecasting gasoil prices under the trend of low-carbon development. Energy Econ. 2024, 134, 107609. [Google Scholar] [CrossRef]
  14. Liu, S.; Xie, G.; Wang, Z.; Wang, S. A secondary decomposition-ensemble framework for interval carbon price forecasting. Appl. Energy 2024, 359, 122613. [Google Scholar] [CrossRef]
  15. Yoo, J.; Son, Y.; Yoon, M.; Choi, S. A Wind Power Scenario Generation Method Based on Copula Functions and Forecast Errors. Sustainability 2023, 15, 16536. [Google Scholar] [CrossRef]
  16. Ji, Q.; Zhang, H.Y.; Geng, J.B. What drives natural gas prices in the United States? – A directed acyclic graph approach. Energy Econ. 2018, 69, 79–88. [Google Scholar] [CrossRef]
  17. Zou, M.; Han, L.; Yang, Z. Price discovery of the Chinese crude oil options and futures markets. Financ. Res. Lett. 2024, 60, 104809. [Google Scholar] [CrossRef]
  18. Yıldırım, F.; Ünlü, U.; Kuloğlu, A.; Çıtak, Ö. Modeling and Analysis of the Impact of Quality Growth and Financial Development on Environmental Sustainability: Evidence from EU Countries. Sustainability 2025, 17, 774. [Google Scholar] [CrossRef]
  19. Zhang, H.; Qian, S.; Ma, Z. An analysis of the market efficiency of the Chinese copper futures based on intertemporal and intermarket arbitrages. Int. Rev. Financ. Anal. 2024, 94, 103243. [Google Scholar] [CrossRef]
  20. Chen, Y.; Zhou, X.; Chen, S.; Mi, J.J. LNG freight rate and LNG price, carbon price, geopolitical risk: A dynamic connectedness analysis. Energy 2024, 302, 131517. [Google Scholar] [CrossRef]
  21. Geng, Q.; Wang, Y. Forecasting the volatility of crude oil basis: Univariate models versus multivariate models. Energy 2024, 295, 130969. [Google Scholar] [CrossRef]
  22. Fu, T.; Huang, D.; Feng, L.; Tang, X. More is better? The impact of predictor choice on the INE oil futures volatility forecasting. Energy Econ. 2024, 134, 107540. [Google Scholar] [CrossRef]
  23. Lu, X.; Qiu, J.; Yang, Y.; Zhang, C.; Lin, J.; An, S. Large Language Model-based Bidding Behavior Agent and Market Sentiment Agent-Assisted Electricity Price Prediction. IEEE Trans. Energy Mark. Policy Regul. 2024, 1–13. [Google Scholar] [CrossRef]
  24. Su, J.; Wang, W.; Bai, Y.; Zhou, P. Measuring the natural gas price features of the Asia-Pacific market from a complex network perspective. Energy 2025, 314, 134133. [Google Scholar] [CrossRef]
  25. Li, N.; Li, J.; Wang, Q.; Yan, D.; Wang, L.; Jia, M. A novel copper price forecasting ensemble method using adversarial interpretive structural model and sparrow search algorithm. Resour. Policy 2024, 91, 104892. [Google Scholar] [CrossRef]
  26. Zhang, Z.; Wang, Y.; Zhang, Y.; Wang, Q. Forecasting carbon prices under diversified attention: A dynamic model averaging approach with common factors. Energy Econ. 2024, 133, 107537. [Google Scholar] [CrossRef]
  27. Li, Y. Market- and future-level sentiment and futures returns in Chinese agricultural futures markets. Borsa Istanb. Rev. 2024, 24, 869–885. [Google Scholar] [CrossRef]
  28. Wu, W.; Xu, M.; Su, R.; Ullah, K. Modeling crude oil volatility using economic sentiment analysis and opinion mining of investors via deep learning and machine learning models. Energy 2024, 289, 130017. [Google Scholar] [CrossRef]
  29. Wang, S.; Ji, T.; Wang, L.; Sun, Y.; Liu, S.C.; Kumar, A.; Lu, C.T. Stocktime: A time series specialized large language model architecture for stock price prediction. arXiv 2024, arXiv:2409.08281. [Google Scholar]
  30. Al-Khayyat, A.T.K.; Nuri Ucan, O. A Multi-Branched Hybrid Perceptron Network for DDoS Attack Detection Using Dynamic Feature Adaptation and Multi-Instance Learning. IEEE Access 2024, 12, 192618–192638. [Google Scholar] [CrossRef]
  31. Veličković, M.; Stojanović, Đ.; Ilin, V.; Mirčetić, D. A Combined Multi-Criteria Decision-Making and Social Cost–Benefit Analysis Approach for Evaluating Sustainable City Logistics Initiatives. Sustainability 2025, 17, 884. [Google Scholar] [CrossRef]
  32. Zhang, J.; Liu, H.; Bai, W.; Li, X. A hybrid approach of wavelet transform, ARIMA and LSTM model for the share price index futures forecasting. North Am. J. Econ. Financ. 2024, 69, 102022. [Google Scholar] [CrossRef]
  33. Zhang, C.; Zhou, X. Forecasting value-at-risk of crude oil futures using a hybrid ARIMA-SVR-POT model. Heliyon 2024, 10, e23358. [Google Scholar] [CrossRef]
  34. Fang, W.; Zhang, T.; Wang, Z.; Ding, J. A multi-modal context-aware sequence stage validation for human-centric AR assembly. Comput. Ind. Eng. 2024, 194, 110355. [Google Scholar] [CrossRef]
  35. Fang, W.; Zhang, S.; Xu, C. Improving prediction efficiency of Chinese stock index futures intraday price by VIX-Lasso-GRU Model. Expert Syst. Appl. 2024, 238, 121968. [Google Scholar] [CrossRef]
  36. Zhu, J.; Ren, Z.; Li, Y.; Liu, X.; Wu, Q.; Li, Y.; Wang, Q. Machine Learning-Enhanced Model-Based Optical Proximity Correction by Using Convolutional Neural Network-Based Variable Threshold Method. IEEE Access 2024, 12, 191517–191526. [Google Scholar] [CrossRef]
  37. Wang, L.; Ma, F.; Liu, G. Forecasting stock volatility in the presence of extreme shocks: Short-term and long-term effects. J. Forecast. 2020, 39, 797–810. [Google Scholar] [CrossRef]
  38. Zhu, Q.; Zhang, F.; Liu, S.; Wu, Y.; Wang, L. A hybrid VMD–BiGRU model for rubber futures time series forecasting. Appl. Soft Comput. 2019, 84, 105739. [Google Scholar] [CrossRef]
  39. Cao, J.H.; Xie, C.; Zhou, Y.; Wang, G.J.; Zhu, Y. Forecasting carbon price: A novel multi-factor spatial-temporal GNN framework integrating graph WaveNet and self-attention mechanism. Energy Econ. 2025, 144, 108318. [Google Scholar] [CrossRef]
  40. Loizidis, S.; Kyprianou, A.; Georghiou, G.E. Electricity market price forecasting using ELM and Bootstrap analysis: A case study of the German and Finnish Day-Ahead markets. Appl. Energy 2024, 363, 123058. [Google Scholar] [CrossRef]
  41. Jha, N.; Kumar Tanneru, H.; Palla, S.; Hussain Mafat, I. Multivariate analysis and forecasting of the crude oil prices: Part I – Classical machine learning approaches. Energy 2024, 296, 131185. [Google Scholar] [CrossRef]
  42. Wu, H.; Liang, Y.; Heng, J.N.; Ma, C.X.; Gao, X.Z. MSV-net: Multi-scale visual-inspired network for short-term electricity price forecasting. Energy 2024, 291, 130350. [Google Scholar] [CrossRef]
  43. Chang, C.L. Extreme events, economic uncertainty and speculation on occurrences of price bubbles in crude oil futures. Energy Econ. 2024, 130, 107318. [Google Scholar] [CrossRef]
  44. Ma, F.; Lu, X.; Wang, L.; Chevallier, J. Global economic policy uncertainty and gold futures market volatility: Evidence from Markov regime-switching GARCH-MIDAS models. J. Forecast. 2021, 40, 1070–1085. [Google Scholar] [CrossRef]
  45. Huang, W.; Zhao, J.; Wang, X. Model-driven multimodal LSTM-CNN for unbiased structural forecasting of European Union allowances open-high-low-close price. Energy Econ. 2024, 132, 107459. [Google Scholar] [CrossRef]
  46. Pang, X.; Zhou, Y.; Wang, P.; Lin, W.; Chang, V. An innovative neural network approach for stock market prediction. J. Supercomput. 2018, 76, 2098–2118. [Google Scholar] [CrossRef]
  47. Cheng, Z.; Li, M.; Sun, Y.; Hong, Y.; Wang, S. Climate change and crude oil prices: An interval forecast model with interval-valued textual data. Energy Econ. 2024, 134, 107612. [Google Scholar] [CrossRef]
  48. Sai, W.; Pan, Z.; Liu, S.; Jiao, Z.; Zhong, Z.; Miao, B.; Chan, S.H. Event-driven forecasting of wholesale electricity price and frequency regulation price using machine learning algorithms. Appl. Energy 2023, 352, 121989. [Google Scholar] [CrossRef]
  49. Wang, L.; Wu, J.; Cao, Y.; Hong, Y. Forecasting renewable energy stock volatility using short and long-term Markov switching GARCH-MIDAS models: Either, neither or both? Energy Econ. 2022, 111, 106056. [Google Scholar] [CrossRef]
  50. Guo, Y.; He, F.; Liang, C.; Ma, F. Oil price volatility predictability: New evidence from a scaled PCA approach. Energy Econ. 2022, 105, 105714. [Google Scholar] [CrossRef]
  51. Ehsani, B.; Pineau, P.O.; Charlin, L. Price forecasting in the Ontario electricity market via TriConvGRU hybrid model: Univariate vs. multivariate frameworks. Appl. Energy 2024, 359, 122649. [Google Scholar] [CrossRef]
  52. Qin, C.; Qin, D.; Jiang, Q.; Zhu, B. Forecasting carbon price with attention mechanism and bidirectional long short-term memory network. Energy 2024, 299, 131410. [Google Scholar] [CrossRef]
  53. Liu, L.; Zhou, S.; Jie, Q.; Du, P.; Xu, Y.; Wang, J. A robust time-varying weight combined model for crude oil price forecasting. Energy 2024, 299, 131352. [Google Scholar] [CrossRef]
  54. Esangbedo, M.O.; Taiwo, B.O.; Abbas, H.H.; Hosseini, S.; Sazid, M.; Fissha, Y. Enhancing the exploitation of natural resources for green energy: An application of LSTM-based meta-model for aluminum prices forecasting. Resour. Policy 2024, 92, 105014. [Google Scholar] [CrossRef]
  55. Menéndez-García, L.A.; García-Nieto, P.J.; García-Gonzalo, E.; Sánchez Lasheras, F. Time series analysis for COMEX platinum spot price forecasting using SVM, MARS, MLP, VARMA and ARIMA models: A case study. Resour. Policy 2024, 95, 105148. [Google Scholar] [CrossRef]
  56. Li, L.; Liu, Z. Sustainable Evolution of China’s Provincial New Quality Productivity Based on Three Dimensions of Multi-Period Development and Combination Weights. Sustainability 2024, 16, 11259. [Google Scholar] [CrossRef]
Figure 1. Research framework.
Figure 1. Research framework.
Sustainability 17 02492 g001
Figure 2. Model flow.
Figure 2. Model flow.
Sustainability 17 02492 g002
Figure 3. VMD decomposition diagram.
Figure 3. VMD decomposition diagram.
Sustainability 17 02492 g003
Figure 4. Correlation coefficient heatmap.
Figure 4. Correlation coefficient heatmap.
Sustainability 17 02492 g004
Figure 5. Results of robustness experiments.
Figure 5. Results of robustness experiments.
Sustainability 17 02492 g005
Figure 6. Comparison of correlation prediction results.
Figure 6. Comparison of correlation prediction results.
Sustainability 17 02492 g006
Figure 7. Model convergence curves.
Figure 7. Model convergence curves.
Sustainability 17 02492 g007
Figure 8. Comparison of ablation prediction results.
Figure 8. Comparison of ablation prediction results.
Sustainability 17 02492 g008
Figure 9. Comparison of time step prediction results.
Figure 9. Comparison of time step prediction results.
Sustainability 17 02492 g009
Figure 10. Comparison of predicted results after adding negative variables.
Figure 10. Comparison of predicted results after adding negative variables.
Sustainability 17 02492 g010
Table 1. Reference methodologies.
Table 1. Reference methodologies.
CategorizationReference NumberMethod
Decomposition[8]EEMD
[9]IEEMD
[10]VMD
Model[32]LSTM
[33,34]ARIMA
[35]GRU
[36]CNN
[37]GARCH
Table 2. Center frequency.
Table 2. Center frequency.
IMF1IMF2IMF3IMF4IMF5IMF6IMF7IMF8IMF9IMF10
K = 13.7512000000000
K = 2 9.0 × 10 5 3.751100000000
K = 3 2.0 × 10 6 1.0 × 10 4 3.75100000000
K = 4 6.0 × 10 7 5.0 × 10 6 2.0 × 10 4 3.7510000000
K = 5 3.0 × 10 7 2.0 × 10 6 3.0 × 10 5 5.0 × 10 4 3.750700000
K = 6 2.0 × 10 7 7.0 × 10 7 3.0 × 10 6 3.0 × 10 5 5.0 × 10 4 3.75070000
K = 7 2.0 × 10 7 5.0 × 10 7 2.0 × 10 6 8.0 × 10 6 4.0 × 10 5 5.0 × 10 4 3.7506000
K = 8 4.0 × 10 8 1.0 × 10 7 5.0 × 10 7 2.0 × 10 6 8.0 × 10 6 4.0 × 10 5 5.0 × 10 4 3.750600
K = 9 4.0 × 10 8 1.0 × 10 7 2.0 × 10 7 6.0 × 10 7 2.0 × 10 6 9.0 × 10 6 4.0 × 10 5 5.0 × 10 4 3.75060
K = 10 5.0 × 10 8 1.0 × 10 7 2.0 × 10 7 4.0 × 10 7 9.0 × 10 7 2.0 × 10 6 9.0 × 10 6 4.0 × 10 5 5.0 × 10 4 3.7506
Table 3. Correlation table.
Table 3. Correlation table.
Correlation GroupVariableCorrelation Coefficient
Strong correlation groupLow price0.99955
High price0.99946
Open price0.99944
Close price0.99943
IMF50.94435
Moderate correlation groupImport volume0.66065
IMF40.43602
IMF30.17683
Weak correlation groupIMF20.07291
IMF10.03973
Amplitude0.00880
Negative correlation groupStorage volume−0.00547
Trading volume−0.01164
Export volume−0.02157
Production−0.07833
Table 4. Strong correlation model errors and rankings.
Table 4. Strong correlation model errors and rankings.
ModelMAEMAPEMSERMSE R 2 AR
GRU0.11273.07%0.02870.169599.62%96.93%
BP0.11443.26%0.03010.173499.57%96.74%
RNN0.15434.16%0.05450.233499.23%95.84%
CNN0.15504.25%0.05370.231899.25%95.75%
LSTM0.17904.82%0.07110.266699.00%95.18%
AE0.17704.85%0.06790.260599.05%95.15%
MLP0.22305.77%0.11320.336499.01%94.23%
RankMAEMAPEMSERMSE R 2 AR
GRU111111
BP222222
RNN334333
AE565556
LSTM656675
MLP777767
Table 5. Medium correlation model errors and rankings.
Table 5. Medium correlation model errors and rankings.
ModelMAEMAPEMSERMSE R 2 AR
AE1.158137.76%2.11021.452768.75%62.24%
MLP1.194142.00%2.21141.487171.69%58.00%
LSTM1.205140.55%2.23681.495668.14%59.45%
BP1.239543.26%2.28201.510668.44%56.74%
GRU1.247644.15%2.34081.530070.07%55.85%
RNN1.266146.03%2.38801.545371.80%53.97%
CNN1.274646.64%2.34621.531772.30%53.36%
RankMAEMAPEMSERMSE R 2 AR
AE111171
MLP232263
LSTM323322
BP444454
GRU555515
RNN666746
CNN777637
Table 6. Weak correlation model errors and rankings.
Table 6. Weak correlation model errors and rankings.
ModelMAEMAPEMSERMSE R 2 AR
MLP1.257944.79%2.35461.534564.96%55.21%
LSTM1.296441.35%2.80631.675244.92%58.65%
CNN1.302348.14%2.38951.545870.49%51.86%
BP1.327042.80%2.82661.681255.17%57.20%
RNN1.473350.64%3.17591.782139.56%49.36%
GRU1.547653.72%3.41181.847137.68%46.28%
AE2.020848.27%7.56992.751451.00%51.73%
RankMAEMAPEMSERMSE R 2 AR
MLP131123
LSTM213351
CNN342214
BP424432
RNN565566
GRU676677
AE757745
Table 7. Robustness test.
Table 7. Robustness test.
Experimental GroupMAEMAPEMSERMSE R 2 AR
In-sample0.25877.29%0.41190.641495.78%92.71%
Out-of-sample0.29619.62%0.40120.632894.18%90.38%
Table 8. Variance and confidence intervals.
Table 8. Variance and confidence intervals.
In-SampleOut-of-Sample
PredVarianceConfidence IntervalVarianceConfidence Interval
Pred_10.4122[−0.0027, 0.0266]0.4484[−0.1141, −0.0412]
Pred_20.4405[0.0083, 0.0386]0.3642[−0.1136, −0.0479]
Pred_30.4537[0.0096, 0.0404]0.4321[−0.1223, −0.0508]
Pred_40.4046[−0.0047, 0.0243]0.4438[−0.0934, −0.0209]
Pred_50.4091[−0.0044, 0.0248]0.3675[−0.1055, −0.0395]
Pred_60.3753[−0.0095, 0.0184]0.415[−0.1122, −0.0421]
Pred_70.4063[−0.0058, 0.0233]0.3504[−0.1044, −0.04]
Pred_80.4488[0.0076, 0.0382]0.3655[−0.1135, −0.0477]
Pred_90.3855[−0.003, 0.0253]0.4494[−0.134, −0.061]
Pred_100.3809[−0.0083, 0.0198]0.4004[−0.1326, −0.0638]
Table 9. Comparison of experimental errors in correlation.
Table 9. Comparison of experimental errors in correlation.
Correlation GroupMAEMAPEMSERMSE R 2 AR
Strong0.11273.07%0.02870.169599.62%96.93%
Medium1.158137.76%2.11021.452768.75%62.24%
Weak1.257944.79%2.35461.534564.96%55.21%
Table 10. Comparison of errors in ablation experiments.
Table 10. Comparison of errors in ablation experiments.
Experimental GroupMAEMAPEMSERMSE R 2 AR
Strong + Medium0.436113.91%0.66540.815790.25%86.09%
Strong + Weak0.423114.75%0.60710.779291.25%85.25%
Medium + Weak1.201541.06%2.18921.479667.42%58.94%
Full model0.324310.17%0.46630.682893.11%89.83%
Table 11. Comparison of time window errors.
Table 11. Comparison of time window errors.
Experimental GroupMAEMAPEMSERMSE R 2 AR
3-step0.420513.43%0.56770.753591.72%86.57%
5-step0.324510.18%0.46770.683993.12%89.82%
7-step0.29419.54%0.42480.651893.85%90.46%
9-step0.376010.28%0.64870.805490.34%89.72%
11-step0.412212.31%0.48980.699992.77%87.69%
13-step0.451413.58%0.55210.743191.85%86.42%
Table 12. Comparison of errors with negative variables added.
Table 12. Comparison of errors with negative variables added.
Experimental GroupMAEMAPEMSERMSE R 2 AR
Full Model0.324310.17%0.46630.682893.11%89.83%
Add negative correlation0.461514.12%0.83830.915687.21%85.88%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, H.; Song, S. Natural Gas Futures Price Prediction Based on Variational Mode Decomposition–Gated Recurrent Unit/Autoencoder/Multilayer Perceptron–Random Forest Hybrid Model. Sustainability 2025, 17, 2492. https://doi.org/10.3390/su17062492

AMA Style

Yu H, Song S. Natural Gas Futures Price Prediction Based on Variational Mode Decomposition–Gated Recurrent Unit/Autoencoder/Multilayer Perceptron–Random Forest Hybrid Model. Sustainability. 2025; 17(6):2492. https://doi.org/10.3390/su17062492

Chicago/Turabian Style

Yu, Haisheng, and Shenhui Song. 2025. "Natural Gas Futures Price Prediction Based on Variational Mode Decomposition–Gated Recurrent Unit/Autoencoder/Multilayer Perceptron–Random Forest Hybrid Model" Sustainability 17, no. 6: 2492. https://doi.org/10.3390/su17062492

APA Style

Yu, H., & Song, S. (2025). Natural Gas Futures Price Prediction Based on Variational Mode Decomposition–Gated Recurrent Unit/Autoencoder/Multilayer Perceptron–Random Forest Hybrid Model. Sustainability, 17(6), 2492. https://doi.org/10.3390/su17062492

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop