Next Article in Journal
On the Interplay Between Behavior Dynamics, Environmental Impacts, and Fairness in the Digitalized Circular Economy with Associated Business Models and Supply Chain Management
Previous Article in Journal
Organic Acids in Food Preservation: Exploring Synergies, Molecular Insights, and Sustainable Applications
Previous Article in Special Issue
Research on the Temporal and Spatial Distribution of Marginal Abatement Costs of Carbon Emissions in the Logistics Industry and Its Influencing Factors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale Temporal Integration for Enhanced Greenhouse Gas Forecasting: Advancing Climate Sustainability

1
School of Business, Soochow University, Suzhou 215021, China
2
School of Computer Science &Technology, Soochow University, Suzhou 215021, China
*
Author to whom correspondence should be addressed.
Sustainability 2025, 17(8), 3436; https://doi.org/10.3390/su17083436
Submission received: 26 February 2025 / Revised: 7 April 2025 / Accepted: 9 April 2025 / Published: 12 April 2025
(This article belongs to the Collection Air Pollution Control and Sustainable Development)

Abstract

:
Greenhouse gases (GHGs) significantly shape global climate systems by driving temperature rises, disrupting weather patterns, and intensifying environmental imbalances, with direct consequences for human life, including rising sea levels, extreme weather, and threats to food security. Accurate forecasting of GHG concentrations is crucial for crafting effective climate policies, curbing carbon emissions, and fostering sustainable development. However, current models often struggle to capture multi-scale temporal patterns and demand substantial computational resources, limiting their practicality. This study presents MST-GHF (Multi-Scale Temporal Greenhouse Gas Forecasting), an innovative framework that integrates daily and monthly CO2 data through a multi-encoder architecture to address these challenges. It leverages an Input Attention encoder to manage short-term daily fluctuations, an Autoformer encoder to capture long-term monthly trends, and a Temporal Attention mechanism to ensure stability across scales. Evaluated on a fifty-year NOAA dataset from Mauna Loa, Barrow, American Samoa, and Antarctica, MST-GHF surpasses 14 baseline models, achieving a Test_R2 of 0.9627 and a Test_MAPE of 1.47%, with notable stability in long-term forecasting. By providing precise GHG predictions, MST-GHF empowers policymakers with reliable data for crafting targeted climate policies and conducting scenario simulations enabling proactive adjustments to emission reduction strategies and enhancing sustainability by aligning interventions with long-term environmental goals. Its optimized computational efficiency, reducing resource demands compared to Transformer-based models, further strengthens sustainability in climate modeling, making it deployable in resource-limited settings. Ultimately, MST-GHF serves as a robust tool to mitigate GHG impacts on climate and human life, advancing sustainability across environmental and societal domains.

1. Introduction

Climate change is a global challenge impacting all of humanity and sustainability, and greenhouse gas emissions data play a crucial role in assessing the effectiveness of environmental policies and understanding the dynamic changes in the global climate system [1,2,3]. As a key parameter for quantifying climate system states, greenhouse gas concentration reflects the imbalance in the Earth’s carbon cycle, far surpassing the significance of regional or national emissions data [4,5,6]. By integrating diverse data sources—atmospheric monitoring stations, satellite remote sensing, and ground-based observations—GHG concentration data provides accurate insights into human activity, ecosystem carbon sink capacity, and the effectiveness of climate policies [7,8,9]. Notably, CO2, as a primary GHG, plays a dual role in sustainability—it is both a key driver of climate change and a crucial target for mitigation efforts. At the same time, it can be effectively harnessed through carbon capture and utilization technologies, transforming emissions into valuable resources and promoting sustainable development [10,11]. Consequently, accurately forecasting GHG concentrations, particularly CO2, is essential for predicting global climate trends, supporting international climate negotiations like the Paris Agreement, and aiding governments in optimizing emission reduction strategies, identifying critical points in the carbon peak process, and enhancing climate governance [12,13].
Accurately capturing the balance between carbon sources (e.g., industrial emissions, deforestation) and sinks (e.g., forests, oceans) is essential for reliable GHG concentration predictions. Multi-dimensional modeling approaches are employed to assess natural systems, human activities, policy interventions, and technological innovations, offering a holistic view of these interactions [14,15]. Historical data analysis, including trend decomposition and anomaly detection, unveils the mechanisms driving carbon cycle dynamics, particularly near critical thresholds such as permafrost thawing or ocean acidification [16,17]. However, policy analysis and scenario simulations, while vital for evaluating international agreements, carbon pricing reforms, and clean energy promotion, are fraught with uncertainty due to their dependence on political will, economic transitions, and unpredictable factors like technology diffusion [18,19]. In contrast, data-driven analysis, rooted in observational data and scientific principles, provides a more stable and repeatable basis for prediction [20].
Traditional climate models, such as Global Circulation Models (GCMs) and Regional Climate Models (RCMs), have been instrumental in simulating atmospheric dynamics. Smith et al., for example, developed a GCM that improved decadal predictions, marking a notable advancement [21]. However, GCMs have several limitations. They often underestimate the periods and amplitudes of major atmospheric oscillations—for instance, the SINTEX coupled model simulates ENSO cycles that are shorter than observed (3–4 years vs. 4–7 years) [22]. Additionally, limited resolution and parameterization deficiencies lead to an underestimation of AMO variability, affecting long-term climate sensitivity projections [23]. Furthermore, GCMs frequently decouple the climate system from the carbon cycle, excluding critical land and ocean carbon uptake feedbacks and thereby underestimating the impacts of carbon–climate interactions [24]. These shortcomings highlight the need for more adaptable and precise modeling approaches. Consequently, these traditional methods struggle to tackle today’s sustainability challenges, as they fail to fully connect greenhouse gas effects on climate and carbon emissions with human life and ecosystem dynamics, underscoring the demand for innovative solutions to achieve effective sustainable development.
As the demand for precise climate predictions grows, data-driven methods have gained traction, leveraging vast datasets to uncover patterns traditional models overlook. Machine learning techniques like Random Forests [25] and Gradient Boosting Decision Trees [26] excel at handling complex high-dimensional data, identifying nonlinear relationships, and pinpointing regional emission hotspots. For instance, Zhang Jianxun et al. applied XGBoost to predict carbon emissions in expanding megacities, demonstrating its utility for urban sustainability planning [27]. However, the complexity of climate systems arises from the nonlinear interactions among thousands of variables, making it difficult for traditional machine learning models to capture multi-scale coupling features [28]. Additionally, purely data-driven models, such as MLP and SVR, may violate fundamental physical principles like energy conservation, leading to systematic errors. For instance, some models misclassify wet months as dry when simulating seasonal precipitation cycles, underscoring the limitations of purely statistical approaches [29]. Consequently, these methods fall short in addressing current sustainability challenges because they cannot deliver multi-scale, long-term precision in forecasting greenhouse gas effects on climate and carbon emissions critical for human life and ecosystems, necessitating more advanced frameworks to ensure sustainable progress [30,31,32].
Deep learning has further transformed GHG forecasting by capturing spatio-temporal dependencies. Long Short-Term Memory (LSTM) networks [33] and Spatio-Temporal Graph Convolutional Networks (STGCNs) have proven effective at modeling atmospheric transport and regional carbon dynamics. Zhang Lei et al. combined CNNs with LSTM to forecast soil organic carbon, showcasing regional applicability [34], while Panja et al.’s E-STGCN model achieved robust air quality predictions in Delhi across seasons [35]. Despite their strengths, these models falter in data-scarce regions and may produce physically inconsistent extrapolations, limiting their generalizability. The Transformer architecture [36], with its self-attention mechanism, has opened new avenues for modeling global atmospheric processes. Wu Xingping et al. developed a temporal graph Transformer-based neural network, achieving an 89.5% accuracy rate in carbon emission predictions [37]. Yet, despite their high precision, Transformers suffer from significant drawbacks: their vast parameter sizes and computational demands make them challenging to deploy in resource-constrained settings, while their lack of multi-scale temporal modeling hinders their ability to capture both short- and long-term GHG dynamics effectively [38,39]. Consequently, these limitations render them inadequate for tackling current sustainability challenges, as they struggle to provide the comprehensive, multi-scale forecasts of greenhouse gas impacts on climate and carbon emissions needed to support human life and ecosystems, highlighting the need for more practical and versatile solutions to advance sustainability. Emerging models like Temporal Convolutional Networks (TCN) [40], DA-RNN [41], Autoformer [42], and Informer [43] have achieved notable success in certain time series forecasting tasks. However, their limitations are evident: they lack multi-scale capabilities, restricting their ability to integrate short- and long-term dynamics, and they have not been applied to address climate sustainability development, hindering their effectiveness for comprehensive GHG forecasting.
To address these gaps, this study introduces the MST-GHF framework, a novel multi-encoder fusion approach for cross-scale greenhouse gas (GHG) concentration forecasting. The true innovation lies in its multi-time-resolution design, which integrates daily and monthly CO2 data, and its hybrid use of specialized encoders, akin to ensemble machine learning where diverse models combine their strengths. Optimized Input Attention and Autoformer Encoders, paired with LSTM units, extract features across temporal scales: the Input Attention Encoder excels at capturing short-term fluctuations (e.g., daily weather impacts), while the Autoformer Encoder models long-term patterns (e.g., seasonal or policy-driven trends). A Temporal Attention mechanism drives multi-step forecasting, ensuring robust accuracy and stability across datasets. By harmonizing high- and low-frequency dynamics, MST-GHF surpasses the limitations of single-resolution models like TCN, DA-RNN, and Autoformer. This hybrid architecture boosts predictive accuracy—outperforming baselines such as LSTM and Transformers—while enhancing computational efficiency for real-time use, balancing interpretability and power through its temporal context integration.
Through the MST-GHF framework, this study seeks to improve the precision of long-term greenhouse gas (GHG) forecasting while reducing the training and inference costs associated with high-precision deep learning models applied to long-term GHG prediction tasks, with the aim of promoting climate sustainability development and thereby advancing overall sustainability. Experimental results indicate that this framework not only enhances prediction accuracy—surpassing all baseline models in accuracy tests with a substantial margin—but also improves computational efficiency.
This research delivers more accurate predictive data to climate policymakers, providing a reliable basis for precise decision-making as GHG emissions near critical thresholds, while advancing sustainability science research. By facilitating the evaluation of climate interventions—such as carbon taxes and ecosystem restoration—it enhances our understanding of sustainable practices. Consequently, the MST-GHF framework optimizes global carbon management, accelerates the transition to a low-carbon economy, and drives progress toward climate sustainability.

2. Materials and Methods

2.1. Data Collection

The dataset consists of daily averages of Carbon Dioxide (CO2) in situ measurements (Atmospheric Carbon Dioxide Dry Air Mole Fraction) from the Barrow Atmospheric Baseline Observatory in the United States, collected by the National Oceanic and Atmospheric Administration (NOAA) [44]. The time span of the dataset is from 24 July 1973 to 30 April 2024, comprising 16,597 daily data points.
This dataset, part of NOAA’s Global Greenhouse Gas Reference Network, is crucial for analyzing spatiotemporal patterns in greenhouse gas emissions and removals, supporting carbon management, and providing early warnings of climatic anomalies [44]. Its high temporal resolution and long duration make it essential for validating models addressing both regional phenomena and global trends, as well as for evaluating multi-temporal hybrid models.

2.2. Data Processing

This study uses a sliding window approach to create time-series datasets for multi-step forecasting. Each window consists of 100 consecutive historical data points (input data points), spanning from time t to t + 99, where t ranges from 1 to 16,488. The model predicts the next 10 time steps (output prediction points). With a total of 16,597 data points, this method generates 16,488 samples, resulting in an input matrix of size 16,488 × 100 and a target matrix of 16,488 × 10.
To capture both short-term and long-term temporal patterns, a monthly dataset is created by computing 30-day moving averages from the daily data. For each daily input sequence, a corresponding monthly sequence is generated by averaging every 30 consecutive daily values. This enables the model to effectively capture multi-scale temporal patterns.
The model incorporates an input attention mechanism that dynamically adjusts the importance of historical data within each time window. Initially, all input variables ( x t to x t + 99 , t ∈ [1, 16, 488]) are uniformly weighted. During training, these weights are iteratively optimized through backpropagation to prioritize temporally significant features. The network architecture directly maps the 100-dimensional input to a 10-dimensional output, enabling end-to-end prediction of the future 10-step sequences.
The dataset is split into 13,190 training samples (80%) and 3298 testing samples (20%) for evaluation. For each sample, the model’s 10-step output is divided into 10 separate prediction sequences (Step 1 to Step 10). Each sequence (13,190 × 1 for training and 3298 × 1 for testing) is evaluated using R2, MSE, MAE, and MAPE to assess the model’s performance across the forecast horizon. Further details on the model architecture and computational process are provided below.

2.3. Methods

The proposed model employs a hierarchical encoder–decoder architecture for multi-scale time series forecasting. Raw data are preprocessed into daily and monthly datasets (see Section 2.2). During encoding, daily data are processed using an Input Attention mechanism to capture short-term dependencies, while monthly data are handled by an Autoformer encoder to extract long-term periodic patterns. The outputs from both encoders are concatenated to form a unified latent representation. In the decoding phase, a Temporal Attention mechanism assigns dynamic weights to this representation, emphasizing key temporal features, which are then projected to the target dimension to predict the next 10 time steps. This architecture effectively combines the sensitivity of daily data with the periodicity of monthly data for multi-scale temporal modeling. Further details are provided in the following sections, with an overview in Figure 1.

2.3.1. Input Attention Mechanism

The input attention mechanism (input-attn) is particularly effective for handling fine-grained data, such as daily time series, due to its ability to adaptively select and weight input features at each time step. Given the frequent and complex variations in fine-grained data, this mechanism dynamically adjusts weights to highlight the most relevant inputs [41]. It also manages multiple driving series within the time window (set to 100 in this study) and suppresses noise, thereby enhancing model robustness [41]. By recalculating attention weights at each step, the mechanism adapts to local feature fluctuations and improves interpretability, making it well-suited for fine-grained time series forecasting.
Next, we introduce how the Input Attention Mechanism is used to process the daily input sequence X D = x D 1 , x D 2 , , x D T d . This mechanism combines an Input Attention Unit, which selects and weights relevant inputs, with LSTM units that generate the corresponding hidden state sequence.
Specifically, for each time step t, we first compute the basic hidden state h t through the LSTM unit:
h t = f 1 h t 1 , x t
where h t is the hidden state at time t, h t 1 is the previous hidden state, x t is the input sequence at t (daily data), and f 1 represents the LSTM unit.
Next, we apply the input attention mechanism to adaptively weight features of the driving sequence, enhancing focus on relevant inputs and suppressing noise. For each feature k (where k = 1, …, n and n is the number of features), we calculate an attention score as follows:
e t k = v e T tanh W e h t 1 ; s t 1 + U e x t k
where e t k is the attention score for feature k at time t, v e T is the attention weight vector, W e is the weight matrix for the hidden and cell states, h t 1 ; s t 1 concatenates the previous hidden state h t 1 and cell state s t 1 from the LSTM, U e is the weight matrix for the input feature x t k (the k-th component of x t ), and t a n h is the activation function. In this experiment, we predict only concentration values, thus n = 1, and k indexes a single feature. The attention weights are then normalized as follows:
α t k = exp e t k i = 1 n exp e t i
where α t k is the attention weight for feature k, computed via softmax over all n features.
Using these weights, we construct a new weighted driving sequence as follows:
x ~ t = α t 1 x t 1 , α t 2 x t 2 , , α t n x t n T
where x ~ t is the updated input vector, and x t k is the k-th feature of the original input x t , scaled by its corresponding weight α t k .
Using the new driving sequence, we update the hidden state h t as follows:
h t = f 1 h t 1 , x ~ t
Thus, we obtain the hidden state sequence for daily data as follows:
H D = h D 1 , h D 2 , , h D T d
where H D is the daily hidden sequence, h D i is the hidden state at daily step i, and T d is the total number of daily time window, set to 100 in this study.

2.3.2. Autoformer Encoder

The Autoformer Encoder is well-suited for coarse-grained data, such as monthly time series, due to its decomposition-based architecture and autocorrelation mechanism. By integrating time series decomposition, Autoformer progressively extracts long-term trends and seasonal components—crucial for modeling extended temporal dependencies [42]. Its autocorrelation module captures cycle-based dependencies by computing autocorrelation across subsequences, effectively identifying periodic patterns. Moreover, with a computational complexity of O(Llog L), substantially lower than the O(L2) of traditional self-attention, it efficiently handles long sequences [42]. These features make Autoformer particularly effective for modeling complex and long-term temporal patterns in coarse-grained data.
Next, we describe how the Autoformer processes the monthly input sequence X M = x M 1 , x M 2 , , x M T m . The Autoformer encoder, composed of a series decomposition block (Series Decomp), an auto-correlation mechanism (Auto-Correlation), and a feed-forward neural network (Feed Forward), extracts long-term trends and seasonal patterns from the input, producing the hidden state sequence for monthly data.
At each time step t, the input sequence is first passed through a series decomposition block. Since future sequences cannot be directly decomposed, Autoformer introduces this block as an internal operation to progressively extract long-term trends from predicted intermediate hidden states. The decomposition smooths out periodic fluctuations using moving averages, thereby emphasizing stable trends. For an input sequence X M R L × d , where L is the sequence length and d the feature dimension, the computation proceeds as follows:
X t = A v g P o o l P a d d i n g X M
X s = X M X t
where X s , X t R L × d denote the seasonal and trend components, respectively. This is denoted as X s , X t = S e r i e s D e c o m p X M , where AvgPool applies a moving average with a fixed kernel size, and Padding ensures boundary consistency during the averaging process.
Next, the auto-correlation mechanism replaces traditional self-attention by capturing cycle-based dependencies via time-delay aggregation. For a discrete-time process X t , autocorrelation at lag τ is computed as follows:
R x x τ = lim L 1 L 1 L X t X t τ
This is efficiently implemented using Fast Fourier Transform (FFT) based on the Wiener–Khinchin theorem [42], and the top k = c × l o g L lags τ 1 ,…, τ k are selected, where c is a hyperparameter.
The R o l l operation shifts the value matrix across time delays τ i , aligning repeating patterns across periods. These shifted values are weighted and aggregated using softmax-normalized autocorrelation scores:
τ 1 , , τ k = a r g T o p k R Q , K τ , τ 1 , , L
R ^ Q , K τ 1 , , R ^ Q , K τ k = s o f t m a x R Q , K τ 1 , , R Q , K τ k
A u t o C o r r e l a t i o n Q , K , V = i = 1 k R o l l V , τ i R ^ Q , K τ i
where R Q , K τ is the autocorrelation between Q and K , a r g T o p k selects the k highest τ values, R ^ are normalized weights, and R o l l V , τ i shifts V by τ i , reintroducing overflow elements at the end.
For multi-head mechanisms with hidden variable channels d m o d e l , and number of heads h , each head processes a subspace:
h e a d i = A u t o C o r r e l a t i o n Q i , K i , V i , Q i , K i , V i R L × d m o d e l h , i 1 , , h
M u l t i H e a d Q , K , V = W O C o n c a t h e a d 1 , , h e a d h
where W O   R d m o d e l × d is the output projection matrix, and C o n c a t merges head outputs.
The result, denoted h t a t t n , is transformed via a feed-forward network, mirroring Transformer’s design:
F e e d F o r w a r d x = R e L U x W 1 + b 1 W 2 + b 2
h t f f n = F e e d F o r w a r d h t a t t n
where W 1 ,   W 1   are weight matrices, b 1 ,   b 2 are biases, and R e L U is the activation function.
Residual connections and layer normalization refine the output as follows:
h t = L a y e r N o r m h t a t t n + h t f f n
Finally, the hidden sequence for monthly data are given by:
H M = h M 1 , h M 2 , , h M T m
where H M is the daily hidden sequence, h M i is the hidden state at monthly step i, and T m is the total number of monthly time steps.

2.3.3. Concating Phase

After encoding, the hidden state sequences H D , H M from the two encoders are concatenated along the feature dimension to form the final hidden sequence H = h 1 , h 2 , , h T , which integrates multi-time-resolution features. This fused sequence is then passed to the decoder for final prediction.
H = C o n c a t H D ,   H M

2.3.4. Temporal Attention Mechanism

The Temporal Attention mechanism is well-suited for decoders, as it effectively addresses long-term dependency challenges often encountered in traditional encoder–decoder architectures. Unlike RNNs, which rely heavily on the encoder’s final hidden state and struggle to retain earlier time-step information, the Temporal Attention mechanism enables the decoder to dynamically attend to hidden states across the entire input sequence [41]. This flexibility allows the decoder to focus on the most relevant time steps for each prediction, regardless of their temporal distance [41].
Next, we outline the use of the Temporal Attention mechanism for decoding the final hidden data and generating predictions. The final hidden sequence H = h 1 , h 2 , , h T passed into the Temporal Mechanism to decode the final prediction sequence. The Temporal Mechanism includes the Temporal Attention Unit and LSTM unit, where the Temporal Attention Unit is used to select the hidden states of the encoders, and the LSTM unit updates the decoder’s hidden state and outputs the final prediction ŷ T .The formulation of Temporal Attention is fundamentally consistent with the Input Attention Mechanism in Section 2.3.1. For detailed parameter meanings, please refer to that section.
Specifically, for each time step t , we first compute the time attention scores as follows:
l t i =   v d T   t a n h W d d t 1 ;   s t 1 + U d h i
where l t i is the attention score gauging the relevance of encoder hidden state h i to the current decoding step.
Then, the attention weights are normalized as follows:
β t i = exp l t i j = 1 T exp l t j
where β t i determines the contribution of h i to the context.
Next, the context vector is computed as follows:
c t = i = 1 T β t i h i  
where c t aggregates encoder information for decoding at t.
The decoder input is combined with the context vector to generate the new input as follows:
y ~ t 1 = w ~ T y t 1 ; c t 1 + b ~
The decoder’s hidden state is updated as follows:
d t =   f 1 d t 1 ,   y ~ t 1
where d t evolves the decoder’s memory, using the same LSTM unit   f 1 as in Section 2.3.1.
Finally, we output the prediction as follows:
ŷ T = v y T W y d t ; c t + b w + b v
where ŷ T is the final output at time T, derived from the combined decoder state and context. Notably, the projection vector v y T and scalar bias b v adjust the dimensions of the transformed vector W y d t ; c t + b w to produce the final prediction ŷ T .

2.4. Metric

The following are the formulas for calculating the goodness-of-fit of the model:
R 2 , Coefficient of Determination:
R 2 = 1 i = 1 n y i y ^ i 2 i = 1 n y i y ¯ 2
M S E , Mean Squared Error:
M S E = 1 n   i = 1 n y i y ^ i 2
M A E , Mean Absolute Error:
M A E = 1   n i = 1 n y i y ^ i
M A P E , Mean Absolute Percentage Error:
M A P E = 100 %   n i = 1 n y i y ^ i y i
In Equations (26)–(29), y i represents the true value, y ^ i represents the predicted value, y ¯ represents the mean of the true values, n represents the sample size.

2.5. Experiment Conditions

In this experiment, we used the Optuna framework to efficiently tune key hyperparameters such as learning rate and batch size for the MST-GHF model and 14 baseline models, ensuring fair comparison under optimal conditions. To enhance reproducibility, we fixed random seeds and GPU settings, conducting only one trial per model to eliminate variability [45]. The experiment ran for 500 iterations on an NVIDIA 3060 GPU. Optuna’s Bayesian optimization and controlled settings minimized bias and randomness, enabling consistent results that can be reliably replicated by other researchers [46].

3. Results

3.1. Overall Performances

The MST-GHF framework demonstrates exceptional predictive accuracy in GHG forecasting to improve the sustainable development, surpassing 14 baseline models across key evaluation metrics. MST-GHF achieves a Test_R2 of 0.9627, exceeding DARNN (0.9519) by 1.1%, Autoformer (0.9480) by 1.5%, and TCN (0.9408) by 2.2%, illustrating its superior ability to model greenhouse gas (GHG) concentration trends. Additionally, its Test_MSE of 0.0002878 is 22.7% lower than DARNN’s 0.0003716 and 37.1% lower than TCN’s 0.0004571, highlighting its significantly reduced forecasting error. MST-GHF also achieves the lowest Test_MAPE at 0.0147 (1.47%), outperforming Autoformer’s 0.0175 (1.75%) by 15.7% and GRU’s 0.0367 (3.67%) by nearly 60%, demonstrating its exceptional precision in minimizing relative errors. This superior accuracy results from MST-GHF’s hybrid multi-encoder architecture, which effectively integrates both daily and monthly data to capture short-term fluctuations and long-term trends. The Input Attention Encoder reduces high-frequency noise in daily data, allowing the model to prioritize the most critical short-term features, while the Autoformer Encoder leverages decomposition and autocorrelation mechanisms to extract periodic patterns and long-term trends.
The cross-scale fusion of these encoders enhances MST-GHF’s ability to outperform single-resolution models like TCN and traditional recurrent models like LSTM, which struggle to maintain stable long-term predictions. This allows MST-GHF to ensure tighter alignment with actual CO2 values, as visualized in Figure 2 across ten steps, further underscoring its robustness and precision in forecasting. All the experimental data can be found in Appendix A Table A1.
From a sustainability perspective, this enhanced predictive accuracy has significant implications. Reliable GHG forecasts provide policymakers with a robust scientific basis for evaluating climate policies, refining carbon reduction strategies, and designing proactive mitigation measures. Furthermore, minimizing forecasting errors helps prevent misallocation of resources in climate action initiatives and ensures that emission targets are based on accurate trend assessments. Improved forecasting precision also supports corporate sustainability planning, aiding businesses in optimizing energy consumption, implementing carbon offset programs, and achieving net-zero emissions more effectively.

3.2. Multi-Step Forecasting

MST-GHF exhibits remarkable stability in multi-step GHG forecasting, with minimal accuracy degradation over extended prediction horizons. Its Test_R2 decreases only 3.3% from 0.9788 at Step 1 to 0.9470 at Step 10, outperforming DARNN, which declines by 3.9% (0.9710 → 0.9334), Autoformer by 4.2% (0.9693 → 0.9284), and TCN by 4.6% (0.9651 → 0.9193). This controlled error growth underscores MST-GHF’s advantage in long-term stability, driven by its Temporal Attention mechanism, which dynamically adjusts feature weighting across different time scales to prevent the accumulation of forecasting errors.
A detailed breakdown of forecasting error metrics further highlights MST-GHF’s robustness. Its Test_MSE increases from 0.000163 at Step 1 to 0.000410 at Step 10, reflecting a 2.5 × increase, whereas Informer’s MSE rises by 2.2 × (0.000416 → 0.000926) and LSTM’s by 1.5 × (0.000678 → 0.000992), indicating that MST-GHF maintains a more controlled error propagation. Similarly, Test_MAPE increases from 1.07% at Step 1 to 1.83% at Step 10, remaining consistently lower than DARNN’s 2.03% and Autoformer’s 2.11% at the same step. Figure 3 illustrates this stability, showing that MST-GHF’s predictions remain closely aligned with actual values even in extended forecasts, unlike BiTransformer_LSTM and Informer, which exhibit greater deviation at later time steps. All the experimental data can be found in Appendix A Table A2, Table A3, Table A4 and Table A5.
This stability is crucial from a sustainability perspective. Reliable multi-step forecasting ensures that long-term climate projections remain actionable, allowing policymakers to formulate effective intervention strategies with minimal uncertainty. High accuracy over extended time horizons also enables precise planning for carbon credit markets, emissions trading schemes, and large-scale reforestation efforts. Furthermore, minimizing forecasting deviations enhances the credibility of climate impact models used to assess future environmental risks, ultimately leading to better-informed decisions on climate adaptation and mitigation policies.

3.3. Ablation Experiment

To systematically assess the contributions of different model components, we conducted ablation experiments by removing individual modules and analyzing the resulting performance changes. The tested configurations included: (1) the complete MST-GHF model, (2) MST-GHF without the Input-Attn Mechanism, (3) MST-GHF with the Autoformer Encoder replaced by a standard LSTM, and (4) MST-GHF without the Temporal-Attn Mechanism, replacing it with a simple linear layer for output mapping. To ensure comparability, all experiments maintained consistent training settings, including hyperparameters, dataset splits, and initialization seeds. The specific data from these experiments can be found in Appendix A Table A6, Table A7, Table A8, Table A9 and Table A10.
The results underscore the critical role of each module in enhancing prediction accuracy and stability. Removing the Input Attention Mechanism led to a sharp decline in performance, with Test_R2 dropping from 0.9627 to 0.9165—a 4.8% reduction—while Test_MSE surged by 123.9% (0.000288 → 0.000645). This highlights the mechanism’s effectiveness in filtering noise and selecting the most relevant short-term features from daily data.
The absence of the Autoformer Encoder resulted in a 1.5% drop in Test_R2 (0.9627 → 0.9480) and a 39.2% increase in Test_MSE, indicating its essential role in capturing long-term periodic trends. Multi-step forecasting further revealed the Autoformer’s advantages: accuracy at Step 5 declined to 0.9267 (compared to 0.9656 in the full model), and at Step 10, it fell to 0.9092 (compared to 0.9470). This sharp degradation suggests that the Autoformer’s decomposition and autocorrelation mechanisms are crucial for identifying hidden periodic patterns in CO2 variations. Without it, the model struggles to maintain stability over extended forecasting horizons, as conventional LSTMs lack the ability to extract long-range dependencies effectively.
The most severe performance deterioration occurred when removing the Temporal Attention Mechanism. Test_R2 plummeted to 0.9077 (a 5.7% decline), and Test_MSE increased by 147.6% (0.000288 → 0.000713). Multi-step evaluation showed that R2 at Step 10 dropped to 0.8883—significantly lower than the full model’s 0.9470—confirming that the Temporal Attention Mechanism is essential for maintaining prediction stability by dynamically adjusting feature importance over different time scales. Its absence leads to a loss of focus in long-term forecasting, causing greater cumulative errors.
From a sustainability perspective, these findings reinforce the necessity of advanced forecasting models in climate policy planning. The ability to accurately model both short- and long-term GHG trends provides a strong scientific foundation for emission reduction strategies, allowing policymakers to make informed decisions based on reliable, data-driven insights. By leveraging a multi-scale forecasting approach, MST-GHF enables governments and businesses to develop adaptive strategies for mitigating greenhouse gas emissions, optimizing green energy investments, and enhancing climate resilience. Additionally, its optimized computational efficiency significantly reduces the energy consumption associated with large-scale climate modeling, making it a more sustainable and resource-efficient solution for environmental decision-making and long-term climate governance.

4. Discussions

The accurate and stable prediction of greenhouse gas (GHG) concentrations is fundamental to effective climate action. Policy decisions on emissions reductions, carbon taxation, and energy transitions rely on precise forecasts to anticipate environmental trends and mitigate risks before they become unmanageable [47,48,49]. Without reliable models, short-term fluctuations in CO2 levels caused by weather variations or human activities could lead to reactive rather than proactive policymaking, while long-term uncertainties in emission trends could hinder strategic planning for sustainable development [50,51]. MST-GHF, with its advanced multi-scale forecasting capability, ensures both short-term precision and long-term trend stability, providing decision-makers with high-fidelity data for scientifically grounded climate policies.
Beyond policymaking, accurate and stable forecasts facilitate critical sustainability research. MST-GHF’s dual-encoder architecture distinguishes between high-frequency daily variations and long-term seasonal or anthropogenic trends, making it particularly valuable for scenario simulations. Researchers can use this capability to test the impact of policy interventions—such as carbon pricing or industrial decarbonization—by modifying historical data and projecting future CO2 levels under different regulatory conditions. Additionally, it supports environmental impact assessments by providing baseline CO2 concentration trajectories that integrate into broader ecological studies, such as evaluating how atmospheric carbon trends correlate with permafrost thawing or ocean acidification. By offering a scientifically robust framework for emissions forecasting, MST-GHF strengthens the analytical foundation for sustainability initiatives, aligning policy, environmental research, and climate mitigation efforts.
One of MST-GHF’s defining advantages is its computational efficiency, which significantly reduces the energy and infrastructure requirements traditionally associated with high-accuracy climate models. Unlike Transformer-based models with O(L2) complexity, MST-GHF employs an Input Attention mechanism with linear complexity and an Autoformer encoder with O(Llog L) efficiency, drastically lowering processing demands. This streamlined design allows the model to be trained within a reasonable timeframe using mid-range hardware and deployed in resource-limited environments such as remote monitoring stations or developing regions with scarce computational power. The reduction in energy consumption not only makes MST-GHF a cost-effective solution but also aligns with sustainability principles by lowering the carbon footprint of climate modeling itself. As a result, high-quality emissions forecasting becomes accessible beyond well-funded research institutions, enabling more regions to integrate data-driven decision-making into their climate strategies.
Despite its strengths, MST-GHF faces challenges, particularly in data-sparse regions where inconsistent monitoring can reduce forecast reliability. While its computational efficiency lowers deployment barriers, some level of hardware infrastructure (e.g., a GPU) is still required for training. Future research could focus on expanding its applicability by incorporating additional greenhouse gases like methane (CH4) and nitrogen oxides (NOₓ) and improving adaptability to incomplete datasets through imputation techniques or transfer learning. Further optimization for low-power edge computing could enhance its real-world usability, ensuring that even the most underserved regions benefit from advanced GHG forecasting to support sustainable climate policies.

5. Conclusions

This study introduces the MST-GHF framework, a multi-scale greenhouse gas (GHG) forecasting model that integrates daily and monthly CO2 data through a dual-encoder architecture. By leveraging an Input Attention mechanism to capture short-term variability, an Autoformer encoder to model long-term trends, and a Temporal Attention mechanism to ensure cross-scale stability, MST-GHF significantly enhances forecasting accuracy and robustness. Compared to baseline models, MST-GHF improves Test_R2 by up to 2.2% and reduces Test_MSE by 37.1%, demonstrating superior predictive performance. Its ability to deliver reliable short- and long-term GHG forecasts provides a strong data foundation for climate policy planning and environmental sustainability. Accurate and stable predictions enable policymakers to anticipate emission trends, refine carbon mitigation strategies, and proactively adjust environmental regulations. Additionally, MST-GHF facilitates scenario simulations, aiding in the evaluation of various climate interventions.
Beyond policy applications, MST-GHF contributes to sustainability science by advancing multi-scale forecasting methodologies that bridge the gap between short-term atmospheric variability and long-term climate trends. Its computational efficiency, achieved through optimized attention mechanisms, reduces resource consumption compared to traditional forecasting models, making it more accessible for deployment in resource-limited settings such as remote monitoring stations. This not only democratizes high-precision climate modeling but also minimizes the environmental footprint of data-intensive computations. Despite its reliance on high-quality continuous datasets, future work can extend the framework to include additional greenhouse gases (e.g., CH₄, NOₓ) and improve adaptability to incomplete data, further supporting global climate governance and sustainable development initiatives.

Author Contributions

H.W. contributed to conceptualization, data curation, software, methodology, visualization, writing—original draft preparation; Y.M. contributed to conceptualization, resources, writing—review and editing; J.R. contributed to conceptualization, data curation, and software; X.Z. contributed to resources and project administration; Z.Q. contributed to project administration and writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the first author.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. All test set data for 15 models.
Table A1. All test set data for 15 models.
ModelTest_MSETest_MAETest_R2Test_MAPE
MST-GHF0.00028780.01163770.96273390.0147310
DARNN0.00037160.01312110.95188570.0167605
Autoformer0.00040130.01364000.94803790.0174719
TCN0.00045710.01462280.94081470.0187982
BiTransfomer_LSTM0.00059620.01844260.92278680.0231243
Informer0.00068740.01824070.91098330.0236269
LSTM0.00086360.02439130.88816120.0295327
Bi_GRU0.00120590.02573060.84383750.0333236
Bi_LSTM0.00126790.02604540.83580320.0338971
GRU0.00127230.02924670.83521570.0367302
RNN0.00133030.03068850.82769750.0373433
CNN1D0.00182900.03512330.76314850.0422021
Bi_RNN0.00253010.04131290.67230050.0485256
CNN1D_LSTM0.00289310.04479570.62531790.0527006
ANN0.00437280.06126010.43362260.0744804
Table A2. 10 step Test_R2 data of 15 models participating in the experiment.
Table A2. 10 step Test_R2 data of 15 models participating in the experiment.
ModelSTEP1STEP2STEP3STEP4STEP5STEP6STEP7STEP8STEP9STEP10
MST-GHF0.978780.974430.970740.966840.965640.961250.958050.954580.950050.94698
DARNN0.971020.967570.960230.953830.958220.948510.948290.943550.934220.93341
Autoformer0.969300.965590.956080.948130.955780.943240.945480.939900.928430.92844
TCN0.965140.961460.948980.938000.950490.933850.939570.933610.917800.91926
BiTransfomer
_LSTM
0.942960.928760.932050.923330.913950.946720.906930.914360.911560.90723
Informer0.945960.944100.920720.894250.926640.895000.914800.909630.878460.88027
LSTM0.911900.905020.900290.883910.894880.909560.881130.857240.865880.87179
Bi_GRU0.882970.837170.907400.887880.845690.767140.913390.785060.745770.86590
Bi_LSTM0.854180.925330.880670.785460.823170.806950.769010.857710.806010.84954
GRU0.859390.847370.863610.880950.790920.827030.835800.791080.852890.80312
RNN0.857300.842500.843150.824550.823570.821050.808260.817930.820420.81823
CNN1D0.794770.799740.864950.766660.716300.786860.733720.796760.716420.65530
Bi_RNN0.721010.708510.723300.667880.618990.656010.620970.693730.661510.65108
CNN1D_LSTM0.709590.638540.656120.624040.707580.596940.673350.534870.601100.51105
ANN0.517340.447800.371910.435680.466130.430100.449040.459150.477850.28123
Table A3. 10 step Test_MSE data of 15 models participating in the experiment.
Table A3. 10 step Test_MSE data of 15 models participating in the experiment.
ModelSTEP1STEP2STEP3STEP4STEP5STEP6STEP7STEP8STEP9STEP10
MST-GHF0.0001630.0001970.0002260.0002560.0002650.0002990.0003240.0003510.0003860.000410
DARNN0.0002230.0002500.0003070.0003560.0003220.0003980.0004000.0004360.0005090.000515
Autoformer0.0002360.0002650.0003390.0004000.0003410.0004380.0004210.0004650.0005530.000554
TCN0.0002680.0002970.0003930.0004780.0003820.0005110.0004670.0005130.0006360.000625
BiTransfomer
_LSTM
0.0004390.0005490.0005240.0005910.0006640.0004110.0007190.0006620.0006840.000718
Informer0.0004160.0004310.0006110.0008160.0005660.0008110.0006580.0006990.0009400.000926
LSTM0.0006780.0007320.0007690.0008960.0008110.0006980.0009190.0011040.0010370.000992
Bi_GRU0.0009010.0012550.0007140.0008650.0011910.0017980.0006690.0016620.0019660.001037
Bi_LSTM0.0011230.0005750.0009200.0016550.0013650.0014910.0017850.0011000.0015000.001164
GRU0.0010830.0011760.0010520.0009180.0016140.0013360.0012690.0016150.0011380.001523
RNN0.0010990.0012140.0012090.0013540.0013620.0013820.0014820.0014070.0013890.001406
CNN1D0.0015810.0015430.0010410.0018000.0021900.0016460.0020580.0015710.0021930.002667
Bi_RNN0.0021490.0022460.0021330.0025620.0029410.0026570.0029290.0023680.0026180.002699
CNN1D_LSTM0.0022370.0027850.0026510.0029000.0022570.0031130.0025240.0035960.0030850.003783
ANN0.0037170.0042550.0048430.0043540.0041210.0044010.0042570.0041810.0040380.005561
Table A4. 10 step Test_MAE data of 15 models participating in the experiment.
Table A4. 10 step Test_MAE data of 15 models participating in the experiment.
ModelSTEP1STEP2STEP3STEP4STEP5STEP6STEP7STEP8STEP9STEP10
MST-GHF0.0084350.0093740.0102100.0109940.0111860.0120210.0125770.0132060.0139180.014457
DARNN0.0100610.0106300.0119070.0128690.0123020.0136800.0138290.0144690.0156570.015807
Autoformer0.0102740.0108910.0125580.0136910.0126390.0144090.0141990.0149280.0163970.016415
TCN0.0109580.0115530.0136430.0150870.0133720.0156610.0149400.0157860.0177070.017521
Informer0.0139440.0142240.0173850.0201330.0165550.0201050.0180420.0186120.0218060.021601
BiTransfomer
_LSTM
0.0160980.0183390.0178760.0186380.0201100.0138800.0208490.0194720.0194890.019675
LSTM0.0213280.0223470.0228610.0249170.0236180.0214310.0253660.0282500.0272840.026510
Bi_GRU0.0224230.0269830.0212350.0227170.0263330.0322630.0192610.0307100.0322590.023122
Bi_LSTM0.0244490.0179240.0224380.0294040.0271640.0288330.0315950.0241250.0278370.026685
GRU0.0276980.0274750.0265730.0251320.0336640.0302060.0298650.0322620.0280230.031569
RNN0.0276050.0291780.0290610.0308090.0311270.0313620.0326240.0317470.0315800.031791
CNN1D0.0329720.0325690.0263670.0352360.0385540.0333920.0374430.0331720.0387250.042803
Bi_RNN0.0375700.0382440.0375860.0414670.0448810.0425100.0449280.0401190.0425240.043299
CNN1D_LSTM0.0382340.0436310.0426430.0447880.0389070.0468890.0418320.0510890.0471420.052802
ANN0.0571030.0609650.0655370.0615440.0595350.0615730.0602710.0596270.0578730.068573
Table A5. 10 step Test_MAPE data of 15 models participating in the experiment.
Table A5. 10 step Test_MAPE data of 15 models participating in the experiment.
ModelSTEP1STEP2STEP3STEP4STEP5STEP6STEP7STEP8STEP9STEP10
MST-GHF0.0106640.0118560.0129080.0138980.0141530.0152100.0159280.0167280.0176390.018326
DARNN0.0127420.0134920.0151950.0164440.0156510.0174970.0176470.0185230.0201280.020286
Autoformer0.0130610.0138750.0160770.0175570.0161180.0184890.0181480.0191510.0211270.021117
TCN0.0139900.0147700.0175310.0194410.0171280.0201770.0191640.0202930.0228750.022614
BiTransfomer
_LSTM
0.0197780.0225800.0220510.0233310.0250920.0177700.0261300.0246490.0247670.025097
Informer0.0179830.0183450.0225110.0261430.0213990.0260940.0233210.0240850.0283290.028060
LSTM0.0255870.0268320.0275590.0300170.0285980.0260560.0309310.0343510.0331160.032278
Bi_GRU0.0289220.0348120.0272700.0288170.0339800.0420750.0248140.0400540.0423870.030106
Bi_LSTM0.0320750.0227650.0290860.0387580.0355040.0376250.0413960.0314110.0365130.033838
GRU0.0345050.0346090.0333640.0311640.0422970.0379240.0374720.0410160.0351050.039846
RNN0.0330200.0354370.0353040.0375360.0379040.0383260.0398710.0386400.0384530.038941
CNN1D0.0394740.0388100.0320600.0422280.0461830.0406660.0451550.0397770.0464880.051180
Bi_RNN0.0439910.0448040.0443060.0485380.0525530.0498510.0527120.0473340.0501560.051010
CNN1D_LSTM0.0446280.0510630.0500010.0525440.0457480.0551730.0493620.0602710.0557630.062453
ANN0.0696360.0740620.0797200.0748640.0724890.0748700.0732800.0726090.0702890.082984
Table A6. Results of Ablation Experiment.
Table A6. Results of Ablation Experiment.
ModelTest_MSETest_MAETest_R2Test_MAPE
MST-GHF0.0002880.0116380.9627340.014731
Without
Input-Attn
0.0006450.0196290.9164570.024515
Without
Autoformer Encoder
0.0004010.0136400.9480380.017472
Without
Temporal-Attn
0.0007130.0204870.9076530.025760
Table A7. 10 step Test_R2 data of 4 models participating in the Ablation Experiment.
Table A7. 10 step Test_R2 data of 4 models participating in the Ablation Experiment.
ModelSTEP1STEP2STEP3STEP4STEP5STEP6STEP7STEP8STEP9STEP10
MST-GHF0.97880.97440.97070.96680.96560.96130.95810.95460.95010.9470
Without
Input-Attn
0.93960.93510.92700.91940.92420.92170.90270.89350.90170.8997
Without
Autoformer Encoder
0.95330.95280.95920.95640.92670.93320.93830.93510.93940.9092
Without
Temporal-Attn
0.94050.92260.91680.91430.91590.90880.89030.88770.89130.8883
Table A8. 10 step Test_MSE data of 4 models participating in the Ablation Experiment.
Table A8. 10 step Test_MSE data of 4 models participating in the Ablation Experiment.
ModelSTEP1STEP2STEP3STEP4STEP5STEP6STEP7STEP8STEP9STEP10
MST-GHF0.0001630.0001970.0002260.0002560.0002650.0002990.0003240.0003510.0003860.000410
Without
Input-Attn
0.0004650.0005000.0005630.0006220.0005850.0006050.0007520.0008240.0007600.000776
Without
Autoformer Encoder
0.0003600.0003630.0003140.0003360.0005660.0005160.0004770.0005010.0004680.000702
Without
Temporal-Attn
0.0004580.0005970.0006410.0006620.0006490.0007040.0008470.0008680.0008410.000864
Table A9. 10 step Test_MAE data of 4 models participating in the Ablation Experiment.
Table A9. 10 step Test_MAE data of 4 models participating in the Ablation Experiment.
ModelSTEP1STEP2STEP3STEP4STEP5STEP6STEP7STEP8STEP9STEP10
MST-GHF0.0084350.0093740.0102100.0109940.0111860.0120210.0125770.0132060.0139180.014457
Without
Input-Attn
0.0143970.0140280.0125500.0124030.0179570.0173350.0154260.0154730.0149230.019603
Without
Autoformer Encoder
0.0169120.0170780.0183380.0196330.0184620.0189140.0211470.0227350.0216120.021463
Without
Temporal-Attn
0.0164710.0187580.0196690.0200320.0194700.0206070.0223470.0226220.0223800.022511
Table A10. 10 step Test_MAPE data of 4 models participating in the Ablation Experiment.
Table A10. 10 step Test_MAPE data of 4 models participating in the Ablation Experiment.
ModelSTEP1STEP2STEP3STEP4STEP5STEP6STEP7STEP8STEP9STEP10
MST-GHF0.0106640.0118560.0129080.0138980.0141530.0152100.0159280.0167280.0176390.018326
Without
Input-Attn
0.0207040.0212730.0228070.0243340.0230510.0236410.0266230.0285000.0271680.027045
Without
Autoformer Encoder
0.0179600.0176580.0158640.0158560.0224390.0215650.0196130.0199650.0190180.024743
Without
Temporal-Attn
0.0204290.0234670.0245780.0250370.0244040.0258190.0282820.0286970.0283600.028526

References

  1. Barnett, J. Security and Climate Change. Glob. Environ. Change 2003, 13, 7–17. [Google Scholar] [CrossRef]
  2. Adedeji, O.; Reuben, O.; Olatoye, O. Global Climate Change. J. Geosci. Environ. Prot. 2014, 2, 114–122. [Google Scholar] [CrossRef]
  3. Jeffry, L.; Ong, M.Y.; Nomanbhay, S.; Mofijur, M.; Mubashir, M.; Show, P.L. Greenhouse Gases Utilization: A Review. Fuel 2021, 301, 121017. [Google Scholar] [CrossRef]
  4. EI Kenawy, A.M.; Al-Awadhi, T.; Abdullah, M.; Jawarneh, R.; Abulibdeh, A. A Preliminary Assessment of Global CO2: Spatial Patterns, Temporal Trends, and Policy Implications. Glob. Chall. 2023, 7, 2300184. [Google Scholar] [CrossRef]
  5. Le, P.V.V.; Randerson, J.T.; Willett, R.; Wright, S.; Smyth, P.; Guilloteau, C.; Mamalakis, A.; Foufoula-Georgiou, E. Climate-Driven Changes in the Predictability of Seasonal Precipitation. Nat. Commun. 2023, 14, 3822. [Google Scholar] [CrossRef] [PubMed]
  6. Singh, S. Forest Fire Emissions: A Contribution to Global Climate Change. Front. For. Glob. Change 2022, 5, 925840. [Google Scholar] [CrossRef]
  7. Ledley, T.S.; Sundquist, E.T.; Schwartz, S.E.; Hall, D.K.; Fellows, J.D.; Killeen, T.L. Climate Change and Greenhouse Gases. Eos Trans. Am. Geophys. Union 1999, 80, 453–458. [Google Scholar] [CrossRef]
  8. Li, X.; Jiang, S.; Wang, X.; Wang, T.; Zhang, S.; Guo, J.; Jiao, D. XCO2 Super-Resolution Reconstruction Based on Spatial Extreme Random Trees. Atmosphere 2024, 15, 440. [Google Scholar] [CrossRef]
  9. Marvin, D.C.; Sleeter, B.M.; Cameron, D.R.; Nelson, E.; Plantinga, A.J. Natural Climate Solutions Provide Robust Carbon Mitigation Capacity under Future Climate Change Scenarios. Sci. Rep. 2023, 13, 19008. [Google Scholar] [CrossRef]
  10. Li, Q.; Li, Q.; Wu, J.; Li, X.; Li, H.; Cheng, Y. Wellhead Stability During Development Process of Hydrate Reservoir in the Northern South China Sea: Evolution and Mechanism. Processes 2025, 13, 40. [Google Scholar] [CrossRef]
  11. Li, Q.; Li, Q.; Cao, H.; Wu, J.; Wang, F.; Wang, Y. The Crack Propagation Behaviour of CO2 Fracturing Fluid in Unconventional Low Permeability Reservoirs: Factor Analysis and Mechanism Revelation. Processes 2025, 13, 159. [Google Scholar] [CrossRef]
  12. Selin, H.; VanDeveer, S.D. Political Science and Prediction: What’s Next for U.S. Climate Change Policy? Rev. Policy Re-Search 2007, 24, 1–27. [Google Scholar] [CrossRef]
  13. Patnaik, S. A Cross-Country Study of Collective Political Strategy: Greenhouse Gas Regulations in the European Union. J. Int. Bus. Stud. 2019, 50, 1130–1155. [Google Scholar] [CrossRef]
  14. Wang, M.; Hu, Z.; Wang, X.; Li, X.; Wang, Y.; Liu, H.; Han, C.; Cai, J.; Zhao, W. Spatio-Temporal Variation of Carbon Sources and Sinks in the Loess Plateau under Different Climatic Conditions and Land Use Types. Forests 2023, 14, 1640. [Google Scholar] [CrossRef]
  15. Wen, H.; Li, Y.; Li, Z.; Cai, X.; Wang, F. Spatial Differentiation of Carbon Budgets and Carbon Balance Zoning in China Based on the Land Use Perspective. Sustainability 2022, 14, 12962. [Google Scholar] [CrossRef]
  16. Maberly, S.C.; Stott, A.W.; Gontero, B. The Differential Ability of Two Species of Seagrass to Use Carbon Dioxide and Bicarbonate and Their Modelled Response to Rising Concentrations of Inorganic Carbon. Front. Plant Sci. 2022, 13, 936716. [Google Scholar] [CrossRef]
  17. Mauclet, E.; Villani, M.; Monhonval, A.; Hirst, C.; Schuur, E.A.G.; Opfergelt, S. Quantifying Exchangeable Base Cations in Permafrost: A Reserve of Nutrients about to Thaw. Earth Syst. Sci. Data 2023, 15, 3891–3904. [Google Scholar] [CrossRef]
  18. van den Bergh, J.; van Beers, C.; King, L.C. Prioritize Carbon Pricing over Fossil-Fuel Subsidy Reform. iScience 2024, 27, 108584. [Google Scholar] [CrossRef]
  19. Wei, X.; Xu, Y. Research on Carbon Emission Prediction and Economic Policy Based on TCN-LSTM Combined with Attention Mechanism. Front. Ecol. Evol. 2023, 11, 1270248. [Google Scholar] [CrossRef]
  20. Nagarajan, G.; Babu, L.D.D. Predictive Analytics On Big Data—An Overview. Informatica 2019, 43, 4. [Google Scholar] [CrossRef]
  21. Smith, D.M.; Cusack, S.; Colman, A.W.; Folland, C.K.; Harris, G.R.; Murphy, J.M. Improved Surface Temperature Prediction for the Coming Decade from a Global Climate Model. Science 2007, 317, 796–799. [Google Scholar] [CrossRef] [PubMed]
  22. Gualdi, S.; Navarra, A.; Guilyardi, E.; Delecluse, P. Assessment of the Tropical Indo-Pacific Climate in the SINTEX CGCM. Ann. Geophys. 2003, 46, 1–26. [Google Scholar] [CrossRef]
  23. Eiselt, K.-U.; Graversen, R.; Fredriksen, H.-B. Time dependence of climate sensitivity. In Proceedings of the EGU General Assembly 2020, EGU2020-12950. Online, 4–8 May 2020. [Google Scholar] [CrossRef]
  24. Bonsor, H.; MacDonald, A.; Calow, R. Potential Impact of Climate Change on Improved and Unimproved Water Supplies in Africa. RSC Issues Environ Sci. Technol 2010, 31, 25–50. [Google Scholar] [CrossRef]
  25. Sun, H.; Liang, L.; Wang, C.; Wu, Y.; Yang, F.; Rong, M. Prediction of the Electrical Strength and Boiling Temperature of the Substitutes for Greenhouse Gas SF6 Using Neural Network and Random Forest. IEEE Access 2020, 8, 124204–124216. [Google Scholar] [CrossRef]
  26. Cai, W.; Wei, R.; Xu, L.; Ding, X. A Method for Modelling Greenhouse Temperature Using Gradient Boost Decision Tree. Inf. Process. Agric. 2022, 9, 343–354. [Google Scholar] [CrossRef]
  27. Zhang, J.; Zhang, H.; Wang, R.; Zhang, M.; Huang, Y.; Hu, J.; Peng, J. Measuring the Critical Influence Factors for Pre-dicting Carbon Dioxide Emissions of Expanding Megacities by XGBoost. Atmosphere 2022, 13, 599. [Google Scholar] [CrossRef]
  28. Rasp, S.; Pritchard, M.S.; Gentine, P. Deep Learning to Represent Subgrid Processes in Climate Models. Proc. Natl. Acad. Sci. USA 2018, 115, 9684–9689. [Google Scholar] [CrossRef]
  29. Isphording, R.N.; Alexander, L.V.; Bador, M.; Green, D.; Evans, J.P.; Wales, S. A Standardized Benchmarking Framework to Assess Downscaled Precipitation Simulations. J. Climate 2024, 37, 1089–1110. [Google Scholar] [CrossRef]
  30. Fabbri, S.; Hauschild, M.Z.; Lenton, T.M.; Owsianiak, M. Multiple Climate Tipping Points Metrics for Improved Sustainability Assessment of Products and Services. Environ. Sci. Technol. 2021, 55, 2800–2810. [Google Scholar] [CrossRef]
  31. Ye, S. Analysis of Influencing Factors of Carbon Emissions from China’s Marine Fishery Energy Consumption under Different Development Scenarios. Front. Mar. Sci. 2024, 11, 1377215. [Google Scholar] [CrossRef]
  32. Tudor, C.; Sova, R. Benchmarking GHG Emissions Forecasting Models for Global Climate Policy. Electronics 2021, 10, 3149. [Google Scholar] [CrossRef]
  33. Sak, H.; Senior, A.; Beaufays, F. Long Short-Term Memory Recurrent Neural Network Architectures for Large Scale Acoustic Modeling. In Proceedings of the Interspeech 2014, Singapore, 14–18 September 2014; ISCA: Pune, India, 2024; pp. 338–342. [Google Scholar]
  34. Zhang, L.; Cai, Y.; Huang, H.; Li, A.; Yang, L.; Zhou, C. A CNN-LSTM Model for Soil Organic Carbon Content Prediction with Long Time Series of MODIS-Based Phenological Variables. Remote Sens. 2022, 14, 4441. [Google Scholar] [CrossRef]
  35. Panja, M.; Chakraborty, T.; Biswas, A.; Deb, S. E-STGCN: Extreme Spatiotemporal Graph Convolutional Networks for Air Quality Forecasting. arXiv 2024, arXiv:2411.12258. [Google Scholar] [CrossRef]
  36. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates, Inc.: New York, NY, USA, 2017; Volume 30. [Google Scholar]
  37. Wu, X.; Yuan, Q.; Zhou, C.; Chen, X.; Xuan, D.; Song, J. Carbon Emissions Forecasting Based on Temporal Graph Transform-er-Based Attentional Neural Network. J. Comput. Methods Sci. Eng. 2024, 24, 1405–1421. [Google Scholar] [CrossRef]
  38. Li, S.; Yuan, G.; Chen, J.; Tan, C.; Zhou, H. Self-Supervised Learning for Solar Radio Spectrum Classification. Universe 2022, 8, 656. [Google Scholar] [CrossRef]
  39. Kim, S.-J.; Chung, Y.-J. Multi-Scale Features for Transformer Model to Improve the Performance of Sound Event Detection. Appl. Sci. 2022, 12, 2626. [Google Scholar] [CrossRef]
  40. Hewage, P.; Behera, A.; Trovati, M.; Pereira, E.; Ghahremani, M.; Palmieri, F.; Liu, Y. Temporal Convolutional Neural (TCN) Network for an Effective Weather Forecasting Using Time-Series Data from the Local Weather Station. Soft Comput. 2020, 24, 16453–16482. [Google Scholar] [CrossRef]
  41. Qin, Y.; Song, D.; Chen, H.; Cheng, W.; Jiang, G.; Cottrell, G. A Dual-Stage Attention-Based Recurrent Neural Network for Time Series Prediction. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 2627–2633. [Google Scholar] [CrossRef]
  42. Wu, H.; Xu, J.; Wang, J.; Long, M. Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting. In Proceedings of the Advances in Neural Information Processing Systems, Online, 6–14 December 2021; Curran Associates, Inc.: New York, NY, USA, 2021; Volume 34, pp. 22419–22430. Available online: https://proceedings.neurips.cc/paper/2021/hash/bcc0d400288793e8bdcd7c19a8ac0c2b-Abstract.html (accessed on 14 November 2024).
  43. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting. Proc. AAAI Conf. Artif. Intell. 2021, 35, 11106–11115. [Google Scholar] [CrossRef]
  44. Thoning, K.W.; Crotwell, A.M.; Mund, J.W. Atmospheric Carbon Dioxide Dry Air Mole Fractions from Continuous Measurements at Mauna Loa, Hawaii, Barrow, Alaska, American Samoa and South Pole, 1973-Present; Version 2024-08-15, National Oceanic and Atmospheric Administration (NOAA); Global Monitoring Laboratory (GML): Boulder, CO, USA, 2024. [Google Scholar] [CrossRef]
  45. Gu, Z.; Zhu, X.; Ye, H.; Zhang, L.; Wang, J.; Zhu, Y.; Jiang, S.; Xiong, Z.; Li, Z.; Wu, W.; et al. Xiezhi: An Ever-Updating Benchmark for Holistic Domain Knowledge Evaluation. Proc. AAAI Conf. Artif. Intell. 2024, 38, 18099–18107. [Google Scholar] [CrossRef]
  46. Zhou, Y.; Dong, Z.; Bao, X. A Ship Trajectory Prediction Method Based on an Optuna–BILSTM Model. Appl. Sci. 2024, 14, 3719. [Google Scholar] [CrossRef]
  47. Farnsworth, A.; Lo, Y.T.E.; Valdes, P.J.; Buzan, J.R.; Mills, B.J.W.; Merdith, A.S.; Scotese, C.R.; Wakeford, H.R. Climate Extremes Likely to Drive Land Mammal Extinction during next Supercontinent Assembly. Nat. Geosci. 2023, 16, 901–908. [Google Scholar] [CrossRef]
  48. Rahman, M.M.; Shafiullah, M.; Alam, M.S.; Rahman, M.S.; Alsanad, M.A.; Islam, M.M.; Islam, M.K.; Rahman, S.M. Decision Tree-Based Ensemble Model for Predicting National Greenhouse Gas Emissions in Saudi Arabia. Appl. Sci. 2023, 13, 3832. [Google Scholar] [CrossRef]
  49. Kumari, S.; Singh, S.K. Machine Learning-Based Time Series Models for Effective CO2 Emission Prediction in India. Env. Sci. Pollut. Res. 2023, 30, 116601–116616. [Google Scholar] [CrossRef]
  50. McNorton, J.; Bousserez, N.; Agustí-Panareda, A.; Balsamo, G.; Cantarello, L.; Engelen, R.; Huijnen, V.; Inness, A.; Kipling, Z.; Parrington, M.; et al. Quantification of Methane Emissions from Hotspots and during COVID-19 Using a Global Atmospheric Inversion. Atmos. Chem. Phys. 2022, 22, 5961–5981. [Google Scholar] [CrossRef]
  51. Ou, Y.; Roney, C.; Alsalam, J.; Calvin, K.; Creason, J.; Edmonds, J.; Fawcett, A.A.; Kyle, P.; Narayan, K.; O’Rourke, P.; et al. Deep Mitigation of CO2 and Non-CO2 Greenhouse Gases toward 1.5 °C and 2 °C Futures. Nat. Commun. 2021, 12, 6245. [Google Scholar] [CrossRef]
Figure 1. MST-GHF encoding and decoding mechanism.
Figure 1. MST-GHF encoding and decoding mechanism.
Sustainability 17 03436 g001
Figure 2. Ten steps MST-GHF predicted values and true values on the dataset.
Figure 2. Ten steps MST-GHF predicted values and true values on the dataset.
Sustainability 17 03436 g002
Figure 3. The predicted values and true values of the nine best models.
Figure 3. The predicted values and true values of the nine best models.
Sustainability 17 03436 g003
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, H.; Mei, Y.; Ren, J.; Zhu, X.; Qian, Z. Multi-Scale Temporal Integration for Enhanced Greenhouse Gas Forecasting: Advancing Climate Sustainability. Sustainability 2025, 17, 3436. https://doi.org/10.3390/su17083436

AMA Style

Wang H, Mei Y, Ren J, Zhu X, Qian Z. Multi-Scale Temporal Integration for Enhanced Greenhouse Gas Forecasting: Advancing Climate Sustainability. Sustainability. 2025; 17(8):3436. https://doi.org/10.3390/su17083436

Chicago/Turabian Style

Wang, Haozhe, Yuqi Mei, Jingxuan Ren, Xiaoxu Zhu, and Zhong Qian. 2025. "Multi-Scale Temporal Integration for Enhanced Greenhouse Gas Forecasting: Advancing Climate Sustainability" Sustainability 17, no. 8: 3436. https://doi.org/10.3390/su17083436

APA Style

Wang, H., Mei, Y., Ren, J., Zhu, X., & Qian, Z. (2025). Multi-Scale Temporal Integration for Enhanced Greenhouse Gas Forecasting: Advancing Climate Sustainability. Sustainability, 17(8), 3436. https://doi.org/10.3390/su17083436

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop