Next Article in Journal
Robust Control of a Zeta Converter: A Feedback Linearization Approach with Digital PWM Implementation
Next Article in Special Issue
Comprehensive Benefit Evaluation Analysis of Multi-Energy Complementary Off-Grid System Operation
Previous Article in Journal
Research on the Quantitative Impact of Power Angle Oscillations on Transient Voltage Stability in AC/DC Receiving-End Power Grids
Previous Article in Special Issue
Residential Electricity Demand Modelling: Validation of a Behavioural Agent-Based Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Time Scale Optimal Scheduling Strategy for the Electro-Hydrogen Coupling System Based on the Modified TCN-PPO

1
China Energy Engineering Group Jiangsu Power Design Institute Co., Ltd., Nanjing 211100, China
2
College of Automation & College of Artificial Intelligence, Nanjing University of Posts and Telecommunications, Nanjing 210023, China
*
Author to whom correspondence should be addressed.
Energies 2025, 18(8), 1926; https://doi.org/10.3390/en18081926
Submission received: 7 March 2025 / Revised: 7 April 2025 / Accepted: 8 April 2025 / Published: 10 April 2025

Abstract

:
The regional integrated energy system, centered on electro-hydrogen technology, serves as a crucial mechanism for advancing the utilization of a high proportion of renewable energy and achieving the low-carbon transition of the energy system. In this context, a multi-time scale optimization model for distributed electro-hydrogen coupling systems is proposed, utilizing an enhanced deep reinforcement learning (DRL) method. Firstly, considering the comprehensive operation cost and real-time deviations, the optimization model of day-ahead and real-time multi-time scale electro-hydrogen coupling system is constructed. Secondly, A dynamic perception model of environmental information is established based on a time convolutional network (TCN) to achieve multi-time scale feature capture of the coupling system and to improve the ability of the agents to perceive the environment of the coupling system. Then, the proposed optimization model is transformed into the Markov decision process (MDP), and a modified Proximal Policy Optimization (PPO) algorithm is introduced to achieve optimal solutions. Finally, case studies are conducted to analyze the electro-hydrogen coupling system in a specific region. The case studies verify the effectiveness of deep reinforcement learning and the electro-hydrogen coupling system in new energy consumption.

1. Introduction

The carbon peaking and carbon neutrality goals necessitate transformative changes in China’s energy system, rendering the transition to a clean, low-carbon, and efficient energy architecture unavoidable [1]. However, the intermittent and volatile nature of both load demand and renewable energy generation presents significant challenges to the stable and secure operation of the power system [2]. In this context, hydrogen energy, recognized as a clean energy source, has garnered significant attention and emerged as a preferred solution for the integrated development of large-scale wind turbines (WT) and photovoltaic (PV) storage. The electro-hydrogen coupling system is increasingly viewed as a pivotal solution for modern power systems [3,4,5]. Investigating the optimal scheduling strategy of the electro-hydrogen coupling system holds substantial importance for enhancing new energy consumption and achieving the dual-carbon objectives within the power industry.
Currently, the primary methods to address the optimal scheduling of the electro-hydrogen coupling system are categorized into single-stage and two-stage methods. Stochastic programming (SP), a widely adopted single-stage method, involves developing an SP model and employing autoregressive moving average techniques to formulate a day-ahead scheduling model. This method optimizes the operation of the electro-hydrogen coupling system by incorporating uncertainties.
Leveraging autoregressive integrated moving average and forward selection techniques, reference [6] introduces a stochastic day-ahead scheduling model for renewable energy-based microgrids, which enables optimal energy management and dynamic energy balance in microgrids powered exclusively by renewable energy. Reference [7] examines the characteristics and requirements of electro-hydrogen shared energy storage and regional integrated energy systems, proposing an energy-sharing architecture and a hierarchical optimal scheduling method that incorporates hydrogen trading. This method utilizes particle swarm optimization (PSO) and mixed-integer linear programming algorithms, enhancing renewable energy utilization. In reference [8], a nonlinear mixed-integer dynamic optimal scheduling model is developed to minimize economic operating costs, carbon dioxide emissions, and energy consumption. The non-dominated sorting genetic algorithm is employed to derive a negative carbon optimal scheduling scheme for the integrated energy system, resulting in 7.23% reductions in operating cost and 9.94% in energy consumption compared with the conventional triple supply system.
Although SP offers a straightforward solution strategy, it is often impractical to predetermine the probability distributions of distributed energy sources, such as load and wind, in the practical implementation of the electro-hydrogen coupling system. Moreover, particle swarm optimization (PSO) and other heuristic algorithms risk trapping solutions in local optima, while the consistency and reliability of the solution strategy remain challenging to ensure. To address these challenges, reference [9] introduces an integrated energy scheduling strategy based on robust optimization (RO) combined with hydrogen energy, designed to perform effectively across all uncertainties. Reference [10] presents a novel multi-objective robust optimization model for an integrated energy system with hydrogen storage, addressing the source-load uncertainty issues in the electro-hydrogen coupling system. The model demonstrates its effectiveness in reducing cost, carbon emissions, and the abandonment of wind and solar energy. Reference [11] develops a data-driven robust optimization model by addressing the coordinated scheduling challenges in the multi-energy coupling system resulting from the expansion of wind and gas installations. The model demonstrates the effectiveness and superiority in coordinating the scheduling of the multi-energy coupling system. However, the results based on the robust optimization (RO) method tend to be conservative, and their high computational complexity renders them time-consuming and inefficient.
The two-stage optimization strategy typically accounts for prediction error characteristics across various time scales and is extensively employed in the operational methodologies of the electro-hydrogen coupling system. Reference [12] addresses the uncertainties in renewable energy generation and demand by implementing a multi-time scale rolling optimization method. This method constructs a hydrogen-based multi-energy microgrid energy management architecture, incorporating both day-ahead and real-time energy scheduling through model predictive control (MPC). This method aims to reduce microgrid operating costs and enhance system energy utilization efficiency. Authors in Reference [13] introduce a real-time energy management method and an optimal control strategy based on dynamic programming and model predictive control (DP-MPC). This strategy accounts for the characteristics and real-time operational statuses of diverse energy sources in the electro-hydrogen hybrid energy storage microgrid, facilitating effective energy management and optimal scheduling while enhancing system stability and economic performance. Zheng et al. [14] focus on the integrated energy system as their research subject, employing a multi-time scale integrated demand response method. They propose a strategic optimization operation architecture and a corresponding optimization model to achieve efficient, economical operation and equitable energy allocation within the integrated energy system. Reference [15] examines the multi-stage and multi-time scale features of the hydrogen-based integrated energy system. It develops a multi-stage and multi-time scale optimal energy management architecture to ensure efficient operation and optimal scheduling of the hydrogen-based integrated energy system.
The efficacy of model predictive control (MPC) is heavily reliant on the precision of the system model. Inaccurate models or the presence of unmodeled dynamic characteristics can result in control instability. Building on this foundation, References [16,17,18,19] investigate the application of two-stage RO in addressing system uncertainties. Reference [16], based on the regional integrated energy system, employs a two-stage RO method to propose a low-carbon economic scheduling strategy that accounts for the uncertainties of wind power and solar energy. This strategy supports the reliable operation and low-carbon development of the regional integrated energy system. Li et al. [17] examine the small-disturbance stability and dynamic response characteristics of the integrated hydrogen hybrid energy system. It introduces a two-stage RO method incorporating stability constraints and develops an associated optimization model to ensure the system’s economic stability. Leveraging the two-stage distributed RO theory within multi-region integrated energy systems, a two-stage distributed robust optimal control strategy is presented in reference [18]. This strategy optimizes the configuration of the multi-region integrated energy system by addressing the complex coupling relationships and uncertainties across regions. Reference [19] addresses energy sharing and carbon transfer by developing a two-stage distributed RO model for integrated energy system clusters. It introduces an optimal strategy rooted in RO theory to achieve economical and efficient operation of these clusters in uncertain environments.
Although the two-stage RO model offers enhanced flexibility and decision-making capabilities in managing uncertainties, such as fluctuations in wind power output, its solutions often remain overly conservative, failing to achieve an optimal balance between economic efficiency and system safety.
Deep reinforcement learning (DRL) integrates the expressive power of deep learning with the decision-making capabilities of reinforcement learning. Owing to its proficiency in handling complex and uncertain environments, it has garnered significant attention in the realm of electro-hydrogen scheduling applications. References [20,21] concentrate on the energy management of the electro-hydrogen coupling system and isolated hydrogen microgrids, utilizing DRL and predictor-driven stochastic optimal techniques to address challenges such as energy conversion, utilization, and information uncertainty. Yang et al. [22] examine the uncertainties within the electro-hydrogen integrated energy system and introduce a multi-stage stochastic scheduling method along with an associated optimal model to facilitate efficient, economical scheduling and ensure the reliable operation of the electro-hydrogen integrated energy system. In reference [23], the authors implement a day-ahead and real-time two-stage scheduling method, employing the Deep Deterministic Policy Gradient (DDPG) algorithm within DRL for the real-time stage. A data-driven two-stage scheduling strategy coupled with an associated optimal model for the multi-energy system is proposed to achieve efficient and economical scheduling of the system.
Kakodkar et al. [24] propose a DRL-based algorithm for optimal online economic dispatch in virtual power plants (VPPs), which leverages DRL to reduce computational complexity, incorporates the large and continuous state spaces arising from stochastic distributed generation characteristics, and implements an edge computing framework to address the stochasticity and high-dimensional state space challenges inherent in VPP operations.
Reference [25] addresses the challenges posed by operational uncertainties in microgrids with high renewable penetration rates to flexible operation by proposing a DRL-based collaborative energy management framework. This three-stage framework integrates internal pricing mechanism establishment, microgrid dispatch optimization, and VPP energy storage management, with comprehensive simulation experiments conducted using real-world datasets to validate the proposed methodology.
In reference [26], Huang et al. develop a DRL-based multi-scenario optimal dispatch methodology for VPPs. Their approach advances the field by designing renewable generation characteristic indices to quantify operational patterns, leveraging conditional generative adversarial networks for high-fidelity scenario data generation, and formulating a transformed VPP dispatch model that systematically integrates stochastic renewable behaviors. The methodology culminates in the application of the Soft Actor-Critic algorithm to derive robust economic dispatch strategies resilient to multi-source uncertainties. In summary, the existing research has made a great contribution to the modeling and solving of the scheduling of the electro-hydrogen coupling system. However, there are still two significant limitations in the existing literature:
(1)
Model dependency dilemma: Conventional methods frequently rely on a set of rigid assumptions and idealized scenarios, necessitating precise modeling of uncertainties and dynamic behaviors. However, in real-world applications, factors such as system parameters, fluctuations in renewable energy, variations in electro-hydrogen load demand, and other complexities exacerbate the model’s intricacy and the challenges of its resolution, thereby constraining the efficacy of conventional methods in practical settings [24].
(2)
Temporal feature neglect: DRL methods demonstrate significant potential in addressing complex decision-making challenges, relying on agents to perceive environmental states and formulate decisions. Nevertheless, current methods exhibit limited capacity to capture and analyze multi-time scale dynamic characteristics [12], making it challenging to discern precisely the temporal features of the coupling system across varying scales. Consequently, agents may overlook the equilibrium between long-term trends and short-term fluctuations during decision-making.
Unlike conventional approaches constrained by stochastic programming’s dependency on precise probability distributions, robust optimization’s tendency toward conservative strategies, and heuristic algorithms’ susceptibility to local optima, this paper proposes a novel multi-time scale optimal scheduling strategy utilizing DRL for the electro-hydrogen coupling system, which enables autonomous discovery of multi-timescale temporal dynamics through self-supervised feature extraction while achieving self-adaptive policy iteration via dual-mode advantage estimation. Taking into account the operational economy and carbon efficiency of the system, a day-ahead operational model targeting the lowest comprehensive cost, alongside a combined day-ahead and real-time optimal model aiming to minimize real-time deviations, is developed. This research incorporates the TCN model to achieve multi-scale perception and extraction of environmental information within the coupling system. Furthermore, a modified TCN-PPO algorithm is proposed for training and problem-solving, culminating in the formulation of a two-stage operational architecture for the system. This paper presents the following three key contributions:
(1)
This paper proposes a multi-time scale optimal scheduling architecture for the electro-hydrogen coupling system, integrating the modified PPO and TCN algorithms. The temporal characteristics of the coupling system are comprehensively extracted and utilized as model inputs through TCN. Additionally, the optimal paradigm of deep reinforcement learning is employed to address the real-time and stochastic challenges introduced by renewable energy sources and loads. This method not only reduces operational costs but also enhances the utilization of new energy.
(2)
A multi-time scale environmental perception model, leveraging TCN, is developed for the electro-hydrogen coupling system. By analyzing the multi-time scale temporal structure data of the coupling system, which encompasses local patterns and long-term dependencies, the TCN model facilitates a profound exploration of multi-time scale features. This enhances the model’s capacity to perceive and interpret the environmental dynamics of the coupling system.
(3)
To address the limitations of the conventional PPO algorithm, including low training efficiency, poor stability, and weak state perception, the modified PPO algorithm is introduced. By incorporating state feature enhancement, the adaptive clipping rate (ACR), and the prioritized experience replay (PER) mechanism, high-quality solutions for multi-time scale optimal challenges in the coupling system are successfully obtained.

2. Problem Formulation

Figure 1 illustrates the multi-time scale optimal scheduling architecture of the electro-hydrogen coupling system, developed based on the modified TCN-PPO architecture. The upper and lower sections of the diagram depict the system’s optimal scheduling architecture for the day-ahead and real-time stages, respectively. Additionally, the left section of the diagram is divided into a multi-time scale schematic encompassing both day-ahead and real-time stages. In the day-ahead optimization stage, the prediction results of equipment output serve as model inputs to formulate the day-ahead operation plan, whereas, in the real-time optimization stage, the model detects real-time power fluctuations to dynamically adjust the operation plan. Furthermore, the central section of the diagram displays two TCN-based environment sensing modules for the day-ahead and real-time scales, respectively, enabling the agent to acquire deep temporal characteristics of the electrical coupling system. Finally, the right side of the figure presents the policy output module incorporating the modified PPO algorithm, which optimizes model training efficiency through state feature enhancement, the adaptive clipping rate, and the PER mechanism. The optimal mapping between environmental states and scheduling policies is successfully established.

2.1. Multi-Time Scale Optimal Operation Model for Electro-Hydrogen Coupling System

2.1.1. Day-Ahead Optimization

The minimization of the integrated operating cost in the day-ahead optimization stage is set as the optimization objective F 1 . The objective includes the power purchase cost, the system operation and maintenance cost, and the start–stop cost of the electrolyzer. The optimization model objectives and constraints are as follows:
  • objective function
min F 1 = t = 1 T C t buy , long + C t op , long + C t start , long
C t buy , long = μ t TOU P t buy , long Δ t
C t op , long = μ ESS i = 1 N ESS P i , t ESS , long Δ t + μ EC i = 1 N EC P i , t EC , long Δ t + μ HFC i = 1 N HFC P i , t HFC , long Δ t
C t start , long = μ start i = 1 N EC φ i , t 1 φ i , t 1 + φ i , t 1 1 φ i , t
where C t buy , long , C t op , long , C t start , long , C t pun , long are the power purchase cost, the system operation and maintenance cost, and the start–stop cost of the electrolyzer, respectively, at time t of the previous day’s planning stage. T is the total time step. μ t TOU , μ ESS , μ EC , μ HFC , and μ start are grid industrial time-of-use tariffs, electric storage Operation and Maintenance (O&M) cost factors, electrolyzer O&M cost factors, fuel cell O&M coefficients, and the start–stop cost of the electrolyzer, respectively. P t buy , long is the power of the electricity purchased from the main grid. N ESS , N EC , and N HFC are the number of electrical storage, electrolyzer, and fuel cell, respectively. P i , t ESS , long is the charge/discharge power of the ith electric energy storage at time t, with a positive value during discharge and a negative value during charging. P i , t EC , long is the power of the ith electrolyzer at time t. φ i , t is the start/stop status of the ith electrolyzer at time t, where 1 signifies starting and 0 otherwise. P i , t HFC , long is the discharge power of the ith Hybrid Fuel Cell (HFC) at time t.
2.
Constraints
(1)
Electrical load balance constraints
P t buy , long + i = 1 N PV P i , t PV , long + i = 1 N WT P i , t WT , long + i = 1 N ESS P i , t ESS , long + i = 1 N HFC P i , t HFC , long = P t load , long + i = 1 N EC P i , t EC , long
where N PV , N WT are the number of photovoltaic and wind power respectively. P i , t PV , L , P i , t WT , L , P t Load , L are the output of the ith photovoltaic at time t, the output of the ith wind power at time t, and the size of the electrical load at time t, respectively. All electrical power variables in the equations are expressed in units of kilowatts (kW).
(2)
Hydrogen load balance constraints
i = 1 N EC η EC H i , t EC , long + i = 1 N HST η HST , dis , long H i , t HST , dis = H t load , long + i = 1 N HFC H i , t HFC , long / η HFC + i = 1 N HST H i , t HST , ch , long / η HST , ch
where H i , t EC , L is the amount of hydrogen produced by the ith electrolyzer at time t. H i , t HST , dis , L is the amount of hydrogen released from the ith hydrogen storage tank at time t. H t Load , L is the hydrogen load size at time t. H i , t HFC , L is the hydrogen consumption of the ith fuel cell at time t. H i , t HST , ch , L is the hydrogen filling of the ith hydrogen storage tank at time t. All hydrogen-related variables in the equations are expressed in units of kilograms (kg). η EC , η HST , dis , η HST , ch , η HFC ( 0 ,   1 ] represent the hydrogen production efficiency of the electrolyzer, the charging/discharging efficiency of the hydrogen storage tank, and the hydrogen utilization efficiency of the fuel cell, respectively.
(3)
Contact line power constraints
P min PN P t buy , long P max PN
where P min PN , P max PN are the lower and upper limits of the power purchased from the main grid contact line, respectively.
(4)
Electrolyzer power constraints
1 φ i , t P i , t EC , 0 + φ i , t P i , min EC P i , t EC , long 1 φ i , t P i , t EC , 0 + φ i , t P i , max EC
where P i , t EC , 0 is the standby power of the ith electrolyzer at time t. P min EC , P max EC are the minimum and maximum operating power of the electrolyzer, respectively.
(5)
Electrical storage operational constraints
P i , t ESS , long P i , max ESS
S i , min ESS S i , t ESS , long S i , max ESS
where P i , max ESS is the maximum charge/discharge power of the ith energy storage system (ESS). S i , t ESS , long is the SOC of the ith ESS at time t. S i , min ESS and S i , max ESS are the lower and upper SOC limits of the ith ESS, respectively.

2.1.2. Real-Time Optimization

To reduce the day-ahead and real-time two-stage operational deviation and improve the system’s ability to consume clean energy, the adjustment cost as well as the penalty for wind and solar abandonment are taken as the optimization objective F 2 in the real-time optimization.
  • Objective function
min F 2 = Δ C t buy , short + Δ C t op , short + C t pun , short
Δ C t buy , short = λ 1 pun Δ P t buy , short Δ t
Δ C t op , short = λ 2 pun i = 1 N ESS Δ P i , t ESS , short Δ t + i = 1 N EC Δ P i , t EC , short Δ t + i = 1 N HFC Δ P i , t HFC , short Δ t
C t pun , short = λ 3 pun i = 1 N PV P i , t PV , cut , short Δ t + i = 1 N WT P i , t WT , cut , short Δ t
where Δ C t buy , short , Δ C t op , short , C t pun , short are the intraday short-time scale power purchase adjustment cost, system operation power adjustment cost, and wind and light abandonment penalty, respectively. λ 1 pun , λ 2 pun , λ 3 pun are the penalty coefficients. Δ P t buy , short is the power purchase adjustment power at time t. Δ P i , t ESS , short is the adjustment power of the ith ESS at time t. Δ P i , t EC , short is the adjustment power of the ith EC at time t. Δ P i , t HFC , short is the adjustment power of the ith HFC at time t. P i , t PV , cut , short is the discarded power of the ith PV at time t. P i , t WT , cut , short is the discarded power of the ith wind power at time t.
2.
Constraints
Short-time scale constraints should be considered for long-time scale scheduling plan constraints. Power adjustments based on the day-ahead power plan lead to an intraday real-time power plan:
P t buy , short = P t buy , long + Δ P t buy , short
P i , t ESS , short = P i , t ESS , long + Δ P i , t ESS , short
P i , t EC , short = P i , t EC , long + Δ P i , t EC , short
P i , t HFC , short = P i , t HFC , long + Δ P i , t HFC , short
where P t buy , short is the intraday real-time power purchase. P i , t ESS , short is the charging and discharging power of the ith ESS at time t. P i , t EC , short is the power of the ith Electrolytic Cell (EC) at time t. P i , t HFC , short is the power of the ith HFC at time t.

2.2. TCN-Based Dynamic Sensing Model for Environmental Information

The operational characteristics and temporal features of various devices in the electro-hydrogen coupling system exhibit significant variability. The limited capability of conventional DRL methods to capture temporal features hinders the comprehensive identification of time-dependent and high-dimensional environmental information. Compared with the gradient vanishing issues and sequential computation constraints inherent in traditional recurrent neural networks when modeling long-term temporal dependencies, TCN addresses these limitations through causal convolutions that preserve chronological causality in the temporal dimension. By employing dilated convolutions to construct exponentially expanding receptive fields, TCN effectively captures multi-scale temporal patterns while enabling highly efficient parallel computations. This architecture demonstrates particular suitability for energy system optimization scenarios requiring simultaneous modeling of short-term operational fluctuations and long-term evolutionary trends in equipment behavior. To address this issue, a dynamic environmental sensing model based on TCN is proposed. The proposed model is designed to effectively extract multi-time scale temporal features from the environmental data of the electro-hydrogen hybrid energy storage system and precisely capture system dynamics. The model offers robust support for the development of more accurate and efficient scheduling strategies.
The TCN algorithm processes and analyzes temporal data using convolutional neural networks, effectively capturing both local patterns and long-term dependencies. The algorithm consists of the following three main modules.

2.2.1. Causal Convolution

Causal convolution constitutes a fundamental component of the TCN. This mechanism ensures that the model avoids accessing future information by employing specific convolutional operations, thereby mitigating the issue of information leakage [25]. The equation of causal convolution is as follows:
y t = τ = 0 T 1 w τ · x max ( t τ , 0 )
where y t is the output. T is the width of the convolution kernel. w τ is the weight of the convolution kernel at delay τ. x max ( t τ , 0 ) is the input at time point tτ. τ is the time delay of the convolution kernel. If t τ < 0 , then 0 is used instead.

2.2.2. Dilated Convolution

Figure 2 illustrates the schematic representation of dilated convolution. Based on the standard convolution operation, the receptive field of the convolution kernel is extended through the introduction of multiple “holes” between its neighboring elements. This method enables the TCN to substantially broaden its receptive domain and improve its ability to perceive the electro-hydrogen coupling system while maintaining the original size of the convolution kernel. The mathematical expression F ( x t ) for the dilated convolution is as follows:
F ( x t ) = i = 0 k 1 f i x t d i
where x t d i is the input at time point t d i . f is the filter, and k is the size of the filter. d is the dilation rate, which determines the number of voids in the convolution kernel.

2.2.3. Residual Connection

To mitigate the gradient vanishing issue and accelerate network convergence, the TCN incorporates skip connections via residual blocks, facilitating information flow across layers. This method allows each network layer to focus on learning residual mappings rather than reconstructing the complete transformation from scratch. It substantially enhances the training stability of deep networks and bolsters the model’s expressive capability.
y = σ ( x + G ( x ) )
where σ is the activation function. G(x) is the series residual modular transform.

2.3. System Operation Model Based on MDP

Considering the dynamics of environmental information in multi-time scales of the coupling system, we construct the multi-time scale optimal model as an MDP and use DRL to solve it. As a sequential decision-making architecture, the MDP seeks to identify an optimal policy that maximizes the cumulative rewards achieved by the agents. Within this architecture, the operational state of the coupling system is defined as the state space. The potential scheduling strategies constitute the action space, while the reward function is formulated based on the system’s operational objectives and constraints. By integrating the sequential decision-making capabilities of MDP with the high-dimensional state space processing abilities of DRL, the model effectively addresses the dynamic variations of environmental information across multi-time scales in the coupling system. The details of the modeling process are delineated below.

2.3.1. State Space

Based on the operation information of the electro-hydrogen coupling system, the long-time scale state s t long and the short-time scale state s t short are obtained as follows, respectively:
s t long = P i , t PV , long , P i , t WT , long , S i , t ESS , long , P t load , long , H t load , long
s t short = P i , t PV , short , P i , t WT , short , S i , t ESS , short , P t load , short , H t load , short , P i , t ESS , long , P i , t EC , long , P i , t HFC , long
where P i , t PV , short is the real-time output of the ith PV at time t. P i , t WT , short is the real-time output of the ith wind power at time t. S i , t ESS , short is the real-time SOC of ith ESS at time t. P t load , short is the size of electrical load at time t. H t load , short is the electrical load size at time t.

2.3.2. Action Space

Based on the system state information, the agents can select an action strategy and execute it in the action space. The long-time scale and short-time scale actions a t long and a t short can be described as follows:
a t long = P i , t ESS , long , P i , t EC , long , P i , t HFC , long
a t short = Δ P i , t ESS , short , Δ P i , t EC , short , Δ P i , t HFC , short , P i , t PV , cut , short , P i , t WT , cut , short

2.3.3. Reward Function

The reward is defined as the immediate feedback received by the agent upon executing an action policy, where a higher reward value indicates a more effective strategy. Therefore, the long-time scale and short-time scale reward functions r t long and r t short can be described as follows:
r t long = C t buy , long C t op , long C t start , long
r t short = Δ C t buy , short Δ C t op , short C t pun , short

3. Proposed Method Based on the Modified TCN-PPO

3.1. Fundamentals of the PPO Algorithm

When coordinating intermittent renewable energy integration with discrete electrolyzer operations, the PPO algorithm effectively balances the exploration–exploitation trade-off. Compared with deterministic policy methods like Deep Deterministic Policy Gradient (DDPG), PPO’s adaptive clipping mechanism prevents catastrophic policy divergence during hydrogen load transients while maintaining multi-timescale coordination capabilities.
As a DRL algorithm based on Actor–Critic architecture [26], the dominance function A ^ t ( a , s ) of the PPO algorithm can be expressed as follows:
A ^ t ( a , s ) = V ϕ s t + r t + γ r t + 1 + + γ K t + 1 r K 1 + γ K t V ϕ s K
where V ϕ s t is the state value function. γ is the discount rate. K is the horizon length.
To maximize the desired reward, the objective function L of the PPO algorithm can be expressed as follows:
L = max E t π θ k + 1 ( a | s ) π θ k ( a | s ) A ^ t ( a , s ) = max E t τ t A ^ t ( a , s )
where τ t is the ratio of old and new policy. π θ k ( · ) is the strategy function of the kth step.
The ratio between the old and new policies is typically constrained during training to maintain process stability. The trimmed objective function L clip can be expressed as follows:
L clip = max E t min τ t A ^ t , clip τ t A ^ t
clip τ t = 1 ε , τ t < 1 ε τ t , 1 ε τ t 1 + ε 1 + ε , τ t > 1 + ε
where clip( ) is the clipping function. ε is the clipping rate clip ratio.

3.2. Modified Mechanisms of the PPO Algorithm

Although the PPO algorithm is widely used for addressing continuous action decision problems, it faces significant limitations, including weak temporal feature perception, limited exploration capability, and poor stability. To address these limitations, we introduce the modified TCN-PPO algorithm, an enhanced version of TCN-PPO. The algorithm’s training performance and solution quality are enhanced through the state feature augmentation, the adaptive clipping rate, and the PER caching mechanism.

3.2.1. State Feature Enhancement

TCN-based multi-time scale environmental information sensing for an electro-hydrogen coupling system to enhance the sensing ability of the agent on the state characteristics of the environment. The state s t long of the DRL agent will be expanded to a multi-time step state s ^ t long .
s ^ t long = s t κ 1 long , s t κ 1 + 1 long , , s t long s ^ t short = s t κ 2 short , s t κ 2 + 1 short , , s t short
where s ^ t long and s ^ t short represent the long-time scale states and short-time scale states after the introduction of the TCN, respectively. κ 1 and κ 2 are the information-capturing fields of view of the long-time scale and short-time scale agents, respectively.

3.2.2. Adaptive Clipping Rate

To balance the algorithm’s convergence speed during the pre-training stage and its stability during the post-training stage, the Cosine Decay model is employed to adaptively adjust the hyper-parameter clipping rate ε . The clipping rate during training can be calculated by Equation (33).
ε = 1 2 ε 0 1 + cos n N π
where ε 0 is the initial clipping rate. n and N represent the current training round and total rounds, respectively.
As shown in Equation (33), during the initial training stage, samples offer substantial learning information and experience, facilitating significant model adjustments to enhance strategy performance. Therefore, a larger clipping rate is set to improve the algorithm update amplitude and training efficiency. During the later training stages, as the model strategy improves, a lower clipping rate is adopted to ensure convergence stability.

3.2.3. Prioritized Experience Replay

In reinforcement learning, a higher loss value indicates a greater contribution of the sample to parameter updates. This signifies a higher intrinsic value of the sample. To maximize the empirical contribution of the samples, the PER caching mechanism is employed. By sorting training samples based on their TD-error magnitudes, high-value samples are effectively utilized.
p i = 1 / ϑ i j = 1 N s 1 / ϑ j
where p i is the probability that the ith sample is sampled, i = 1 , 2 , , N s . ϑ represents the rank order number of sample among all samples. N s is the number of samples.

3.3. Training Process of the Proposed Modified TCN-PPO Algorithm

The proposed training flow based on the modified TCN-PPO method is shown in Figure 3. Firstly, initialize the parameters of the modified TCN-PPO agent network and start the environment initialization at the beginning of each round. Secondly, judge the current task time scale, and if it is the day-ahead stage, observe the environment state A based on the TCN and formulate the corresponding system day-ahead operation plan a t long based on Equation (24), which serves as the decision-making basis for the intra-day real-time scenario. After the execution of action a t long based on Equation (26), the reward r t long obtained by the agent is calculated, and the sample s ^ t long , a t long , r t long , s ^ t + 1 long is stored in the buffer Dlong for training. Then, samples are drawn every Nlong step based on the PER mechanism to update the agent network parameters. Similarly, in the case of the intraday real-time stage, the real-time scenario is adjusted based on the system’s day-ahead running schedule. Finally, at the end of each training round, the clipping rate ε is adjusted based on the adaptive clipping rate mechanism proposed in Equation (33). The above steps are repeated until the maximum training round is complete to complete the training of the proposed algorithm.

4. Case Studies

4.1. Case Study Setup

To validate the effectiveness of the proposed multi-time scale optimal scheduling strategy based on DRL for the electro-hydrogen coupling system, this paper conducts a case analysis using an industrial park’s electro-hydrogen system. Considering the operational characteristics of each component, the day-ahead and real-time scheduling intervals are set to 1 h and 15 min, respectively. The wind power unit has a rated capacity of 800 kW, while the photovoltaic system has a capacity of 600 kW. The system includes eight electrolyzers, each with a rated power of 360 kW. The energy storage system features a capacity of 800 kWh and a maximum charging/discharging power of 400 kW. Table 1 and Table 2 detail the industrial time-of-use electricity prices and the parameter settings for the case studies, respectively [27]. Figure 4 illustrates the day-ahead wind and solar power outputs alongside the load forecast. The raw dataset underwent the following preprocessing pipeline: (1) Wind/PV generation data were processed with linear interpolation for missing value imputation followed by Z-score normalization; (2) electric load profiles were transformed into time-series features using a sliding window method (24 h window with 1 h stride); (3) electrolyzer operational state variables were encoded via one-hot representation for discrete control pattern recognition.

4.2. Analysis of the Training Process

With the modified TCN-PPO algorithm, the discount rate γ is set to 0.98, the replay buffer capacity to 3000, and the mini-batch size to 120. The initial clipping rate ε 0 is 0.25, while the information captures horizons κ 1 and κ 2 for long-time scale and short-time scale agents are 24 and 16, respectively. The update step lengths ( N long and N short ) are both 16. Based on these settings, the training reward of the modified TCN-PPO algorithm is obtained, as shown in Figure 5. As shown in the figure, despite significant fluctuations in the agents’ reward during the initial stage, continuous training facilitates the attainment of stable, high rewards via an optimized system operation strategy, thereby ensuring the efficient operation of the electro-hydrogen coupling system. Specifically, the day-ahead reward exhibits relatively minor fluctuations, and the agent attains stable convergence after approximately 350 rounds, with an average reward of −1813.37. In contrast, the real-time reward demonstrates greater fluctuations and converges at a slower rate. During pre-training, the agents are required to explore the environment and refine their strategy using a higher clipping rate to adapt to real-time output and load variations. Ultimately, convergence is achieved after approximately 500 rounds, with the average reward stabilizing at −49.95, demonstrating that the agents effectively learn the optimal mapping between system states and scheduling strategies through iterative training.

4.3. Analysis of the Testing Results

Based on the proposed modified TCN-PPO model, Figure 6 and Figure 7 illustrate the results of the day-ahead system power scheduling and the electrolyzer’s day-ahead production plan, respectively. As shown in the figures, the results indicate that the proposed method successfully develops an optimal operation strategy to ensure power balance during the day-ahead stage. Specifically, the agent charges the ESS from 00:00 to 08:00, taking advantage of lower electricity costs. From 08:00 to 12:00, the agent reduces the electrolyzer load to integrate wind power and enhances ESS output to alleviate high power purchase costs. In this period, the average purchased power is 1049.13 kW, representing a 41.74% reduction compared with the 00:00–08:00 interval. As shown in Figure 7, the electrolyzer’s average load from 08:00 to 12:00 is 1340.55 kW, a 28.53% reduction compared with the earlier interval. When PV output declines sharply in the evening, the agent employs ESS discharge to partially meet the load demand.
Furthermore, Figure 8 illustrates the real-time stage system power adjustments, with hourly data averaged across four intervals to enhance readability. As illustrated, the purchased power and electrolyzer power demonstrate substantial adjustments, whereas the adjustments for the ESS and fuel cell are relatively minor. Across different time intervals, power adjustments are significantly lower from 10:00 to 17:00, averaging merely 55.89 kW per hour, representing a 50.60% reduction compared with the 00:00–10:00 interval. Ultimately, the intraday short-time scale power purchase adjustment cost amounted to 486.14 yuan, whereas the system equipment operation adjustment cost totaled 629.80 yuan. In conclusion, the proposed method successfully optimizes the operation strategy of the electro-hydrogen coupling system, thereby improving economic efficiency and reducing carbon emissions.

4.4. Comparison of Different Algorithms

To assess the efficacy of the proposed modified TCN-PPO algorithm, Figure 9 displays the cumulative reward curves of four DRL algorithms—DDPG, PPO, TCN-PPO, and the modified TCN-PPO—while Table 3 provides a summary of their simulation results. As shown, the DDPG algorithm yields the highest cumulative cost of ¥1.56 million, primarily attributed to its dependence on deterministic strategies and limited exploration capabilities, which hinder its ability to identify optimal solutions in complex environments. Consequently, the average daily power purchase cost for DDPG amounts to ¥28,792.27. In contrast, the PPO algorithm, utilizing a policy gradient method, demonstrates superior adaptability for electro-hydrogen coupling system scheduling, leading to a significant reduction in overall system cost. Furthermore, TCN-PPO enhances performance by effectively capturing temporal dynamics, achieving a cumulative cost of ¥1.37 million, which surpasses both DDPG and PPO. The proposed modified TCN-PPO attains the lowest cumulative cost of ¥1.31 million, corresponding to reductions of 16.03% and 12.67% compared with DDPG and PPO, respectively. This enhancement is attributed to improved state feature perception and adaptive reduction rate adjustments, which substantially enhance training quality and facilitate high-quality solutions in complex decision-making environments.

4.5. Analysis of Ablation Experiment

To rigorously validate the performance contributions of our algorithmic enhancements, we conduct comprehensive ablation studies comparing the following variants: PPO, TCN-PPO, TCN-PPO with adaptive clipping rate (TCN-PPO-ACR), TCN-PPO with prioritized experience replay (TCN-PPO-PER), and our proposed modified TCN-PPO. As demonstrated in Figure 10 (training reward curves) and Table 4 (quantitative metrics), PPO exhibited the fastest convergence (≈200 episodes) but yielded inferior rewards (−1983.62 and −61.23 for day-ahead and real-time phases, respectively). The integration of TCN substantially enhanced environmental state perception, increasing mean converged rewards by 6.15% in day-ahead scheduling at the cost of 27% slower convergence. The TCN-PPO-ACR variant achieved a 20.59% reduction in reward variance and a 1.71% reward improvement through dynamic clipping rate adaptation, particularly enhancing early-stage training efficiency and late-phase stability. TCN-PPO-PER demonstrated 3.08% and 8.23% reward gains in respective phases via prioritized sampling of high-learning-value experiences. Our final modified TCN-PPO synergistically integrates TCN’s multi-scale feature extraction (8.58% reward gain), ACR’s adaptive policy optimization (18.42% real-time improvement), and PER’s experience prioritization, achieving coordinated performance elevation beyond component-wise additive effects. These ablation results systematically quantify individual contribution ratios while confirming the framework’s holistic efficacy through multi-mechanism coordination.

4.6. Analysis of Model Sensitivity

To assess the influence of the initial clipping rate on model performance, Figure 11 illustrates the average rewards achieved by agents with varying initial clipping rates over the training period. All results are calculated over 10 independent runs with different random seeds.
As shown in the figure, the average reward initially increases and subsequently declines with higher initial clipping rates, reaching a peak value of −1821.39 at a clipping rate of 0.25. Beyond this threshold, the reward reduces by approximately 7.02% for each incremental rise of 0.05 in the initial clipping rate. This reduction arises due to the higher clipping rate inducing excessive oscillations during training, which impedes the convergence toward a stable and effective strategy. Meanwhile, the training duration exhibits a U-shaped pattern, reducing initially and then increasing with higher clipping rates. The shortest training duration, 90.18 min, is achieved at a clipping rate of 0.3. However, at a clipping rate of 0.25, the training duration is 95.37 min, still relatively short and acceptable.
In conclusion, an initial clipping rate of 0.25 achieves an optimal balance between model performance and training efficiency.

5. Conclusions

In this paper, a multi-time scale optimal scheduling strategy for the electro-hydrogen coupling system based on the modified TCN-PPO algorithm is proposed. The DRL agents are utilized to develop day-ahead and real-time operation strategies for the coupling system, aiming to reduce the integrated system operation cost and mitigate real-time deviations. The experimental validation based on a regional electro-hydrogen coupling system yields the following conclusions:
(1)
The proposed optimal scheduling strategy for the electro-hydrogen coupling system employs a multi-time scale architecture that integrates the feature extraction capability of TCN and the decision optimization capability of DRL. By effectively formulating the production plan and real-time adjustment scheme in advance, the system’s real-time power adjustment cost is reduced to ¥629.80, significantly enhancing both economic efficiency and low-carbon operation.
(2)
Based on the TCN to capture multi-time scale environmental time-series characteristics, the DRL agents’ environmental perception and decision-making capabilities for the coupling system are enhanced. The cumulative cost of the proposed modified TCN-PPO is only ¥1.31 million during a 30-day test cycle, which is 12.67% lower compared with the PPO algorithm.
(3)
The proposed feature enhancement, adaptive clipping rate, and PER mechanisms enable the modified TCN-PPO algorithm to address the limitations of the PPO algorithm and achieve high-quality solutions for the multi-time scale optimal problem in the coupling system. Furthermore, model sensitivity experiments demonstrate that the model achieves an optimal balance between performance and stability when the initial clipping rate falls within the interval [0.25, 0.3].
Although the current DRL-based optimal scheduling strategy for the electro-hydrogen coupling system exhibits distinct advantages, the increasing complexity of electro-hydrogen loads and the diversification of energy management system components necessitate further exploration of the characteristics of demand-response loads, supercapacitors, and other equipment to optimize the coordinated operation of such systems.
While the current DRL-based optimization strategy for electricity-hydrogen coupled systems shows operational advantages, its neglect of second-level dynamic characteristics from fast-response devices like supercapacitors may limit millisecond-level power fluctuation mitigation capabilities in high renewable penetration scenarios. Future research should integrate demand-responsive loads and supercapacitor dynamics to enhance coordinated operational performance.
For practical implementation, future research should prioritize establishing a provincial hydrogen-electricity coupling network demonstration in Jiangsu Province, particularly focusing on dynamic subsidy mechanisms for valley-period hydrogen production. Scalability validation must address model generalization across heterogeneous multi-regional systems incorporating diverse electrolysis technologies (e.g., proton exchange membrane electrolyzers) while developing self-adaptive hyperparameter tuning mechanisms. Market integration requires extending the economic objective functions (Equations (1)–(4)) through carbon trading cost internalization and electricity market bidding revenue optimization. Critical engineering challenges include mitigating cloud-edge coordination latency and addressing equipment lifespan degradation from frequent start–stop cycles, to be resolved via hardware-in-the-loop testing with accelerated aging protocols.

Author Contributions

Conceptualization, D.L. and K.Q.; Methodology, D.L. and J.Z.; Validation, Q.X.; Formal analysis, J.Z. and Q.X.; Investigation, Y.X.; Resources, K.Q.; Data curation, Z.W.; Writing—original draft, K.Q. and Y.P.; Writing—review & editing, D.L.; Visualization, Y.X. and Y.P.; Supervision, Y.X.; Project administration, D.L. and K.Q.; Funding acquisition, D.L. and K.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by Technology Project of China Energy Engineering Group Jiangsu Power Design Institute Co., Ltd. (32-JK-2024-040) and Technology Project of China Power Engineering Consulting Group Co., Ltd. (DG3-A02-2023).

Data Availability Statement

The data presented in this study are available on request from the corresponding author due to specify the privacy reasons for the restriction.

Conflicts of Interest

Author Dongsen Li, Kang Qian, Yiyue Xu, Jiangshan Zhou, Zhangfan Wang and Yufei Peng was employed by the company China Energy Engineering Group Jiangsu Power Design Institute Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The authors declare that this study received funding from China Energy Engineering Group Jiangsu Power Design Institute Co., Ltd. and China Power Engineering Consulting Group Co., Ltd. The funder was not involved in the study design, collection, analysis, interpretation of data, the writing of this article or the decision to submit it for publication.

References

  1. Yuan, K.; Zhang, T.; Xie, X.; Du, S.; Xue, X.; Abdul-Manan, A.F.; Huang, Z. Exploration of low-cost green transition opportunities for China’s power system under dual carbon goals. J. Clean. Prod. 2023, 414, 137590. [Google Scholar] [CrossRef]
  2. Dong, Y.; Shan, X.; Yan, Y.; Leng, X.; Wang, Y. Architecture, key technologies and applications of load dispatching in china power grid. J. Mod. Power Syst. Clean Energy 2022, 10, 316–327. [Google Scholar] [CrossRef]
  3. Abdelghany, M.B.; Al-Durra, A.; Zeineldin, H.H.; Gao, F. A Coordinated Multi-Time Scale Model Predictive Control for Output Power Smoothing in Hybrid Microgrid Incorporating Hydrogen Energy Storage. IEEE Trans. Ind. Inform. 2024, 20, 10987–11001. [Google Scholar] [CrossRef]
  4. Fan, G.; Liu, Z.; Liu, X.; Shi, Y.; Wu, D.; Guo, J.; Zhang, S.; Yang, X.; Zhang, Y. Two-layer collaborative optimization for a renewable energy system combining electricity storage, hydrogen storage, and heat storage. Energy 2022, 259, 125047. [Google Scholar] [CrossRef]
  5. Yue, M.; Lambert, H.; Pahon, E.; Roche, R.; Jemei, S.; Hissel, D. Hydrogen energy systems: A critical review of technologies, applications, trends and challenges. Renew. Sustain. Energy Rev. 2021, 146, 111180. [Google Scholar] [CrossRef]
  6. Daneshvar, M.; Mohammadi-Ivatloo, B.; Zare, K.; Asadi, S. Transactive energy management for optimal scheduling of interconnected microgrids with hydrogen energy storage. Int. J. Hydrog. Energy 2021, 46, 16267–16278. [Google Scholar] [CrossRef]
  7. Li, Q.; Xiao, X.; Pu, Y.; Luo, S.; Liu, H.; Chen, W. Hierarchical optimal scheduling method for regional integrated energy systems considering electricity-hydrogen shared energy. Appl. Energy 2023, 349, 121670. [Google Scholar] [CrossRef]
  8. Liu, S.; Song, L.; Wang, T.; Hao, Y.; Dai, B.; Wang, Z. Negative carbon optimal scheduling of integrated energy system using a non-dominant sorting genetic algorithm. Energy Convers. Manag. 2023, 291, 117345. [Google Scholar] [CrossRef]
  9. Lu, J.; Huang, D.; Ren, H. Data-driven source-load robust optimal scheduling of integrated energy production unit including hydrogen energy coupling. Glob. Energy Interconnect. 2023, 6, 375–388. [Google Scholar] [CrossRef]
  10. Zhao, Y.; Wei, Y.; Zhang, S.; Guo, Y.; Sun, H. Multi-Objective Robust Optimization of Integrated Energy System with Hydrogen Energy Storage. Energies 2024, 17, 1132. [Google Scholar] [CrossRef]
  11. Zhang, Y.; Liu, Y.; Shu, S.; Zheng, F.; Huang, Z. A data-driven distributionally robust optimization model for multi-energy coupled system considering the temporal-spatial correlation and distribution uncertainty of renewable energy sources. Energy 2021, 216, 119171. [Google Scholar] [CrossRef]
  12. Fang, X.; Dong, W.; Wang, Y.; Yang, Q. Multiple time-scale energy management strategy for a hydrogen-based multi-energy microgrid. Appl. Energy 2022, 328, 120195. [Google Scholar] [CrossRef]
  13. Li, Q.; Zou, X.; Pu, Y.; Chen, W. A Real-time Energy Management Method for Electric-hydrogen Hybrid Energy Storage Microgrids Based on DP-MPC. CSEE J. Power Energy Syst. 2020, 10, 324–336. [Google Scholar] [CrossRef]
  14. Zheng, B.; Hou, X.; Xu, S.; Jin, T.; Liu, W.; Li, N.; Guo, D.; Pan, C. Strategic optimization operations in the integrated energy system through multitime scale comprehensive demand response. Energy Sci. Eng. 2024, 12, 2236–2257. [Google Scholar] [CrossRef]
  15. Fang, X.; Dong, W.; Wang, Y.; Yang, Q. Multi-stage and multi-timescale optimal energy management for hydrogen-based integrated energy systems. Energy 2024, 286, 129576. [Google Scholar] [CrossRef]
  16. Zhang, M.; Wang, B.; Wei, J. The Robust Optimization of Low-Carbon Economic Dispatching for Regional Integrated Energy Systems Considering Wind and Solar Uncertainty. Electronics 2024, 13, 3480. [Google Scholar] [CrossRef]
  17. Li, Q.; Qiu, Y.; Yang, H.; Xu, Y.; Chen, W.; Wang, P. Stability-constrained two-stage robust optimization for integrated hydrogen hybrid energy system. CSEE J. Power Energy Syst. 2020, 7, 162–171. [Google Scholar] [CrossRef]
  18. Li, X.; Wu, N. A two-stage distributed robust optimal control strategy for energy collaboration in multi-regional integrated energy systems based on cooperative game. Energy 2024, 305, 132221. [Google Scholar] [CrossRef]
  19. Fan, W.; Ju, L.; Tan, Z.; Li, X.; Zhang, A.; Li, X.; Wang, Y. Two-stage distributionally robust optimization model of integrated energy system group considering energy sharing and carbon transfer. Appl. Energy 2023, 331, 120426. [Google Scholar] [CrossRef]
  20. Shi, T.; Xu, C.; Dong, W.; Zhou, H.; Bokhari, A.; Klemeš, J.J.; Han, N. Research on energy management of hydrogen electric coupling system based on deep reinforcement learning. Energy 2023, 282, 128174. [Google Scholar] [CrossRef]
  21. Dong, W.; Sun, H.; Mei, C.; Li, Z.; Zhang, J.; Yang, H. Forecast-driven stochastic optimization scheduling of an energy management system for an isolated hydrogen microgrid. Energy Convers. Manag. 2023, 277, 128174. [Google Scholar] [CrossRef]
  22. Yang, Z.; Ren, Z.; Li, H.; Sun, Z.; Feng, J.; Xia, W. A multi-stage stochastic dispatching method for electricity-hydrogen integrated energy systems driven by model and data. Appl. Energy 2024, 371, 123668. [Google Scholar] [CrossRef]
  23. Li, H.; Qin, B.; Wang, S.; Ding, T.; Wang, H. Data-driven two-stage scheduling of multi-energy systems for operational flexibility enhancement. Int. J. Electr. Power Energy Syst. 2024, 162, 110230. [Google Scholar] [CrossRef]
  24. Kakodkar, R.; He, G.; Demirhan, C.; Arbabzadeh, M.; Baratsas, S.; Avraamidou, S.; Mallapragada, D.; Miller, I.; Allen, R.; Gençer, E.; et al. A review of analytical and optimization methodologies for transitions in multi-scale energy systems. Renew. Sustain. Energy Rev. 2022, 160, 112277. [Google Scholar] [CrossRef]
  25. Li, Y.; Song, L.; Zhang, S.; Kraus, L.; Adcox, T.; Willardson, R.; Komandur, A.; Lu, N. A TCN-based hybrid forecasting architecture for hours-ahead utility-scale PV forecasting. IEEE Trans. Smart Grid 2023, 14, 4073–4085. [Google Scholar] [CrossRef]
  26. Huang, B.; Wang, J. Deep-reinforcement-learning-based capacity scheduling for PV-battery storage system. IEEE Trans. Smart Grid 2021, 12, 2272–2283. [Google Scholar] [CrossRef]
  27. Yuan, T.J.; Wan, Z.; Wang, J.J.; Zhang, D.; Jiang, D. The day-ahead output plan of hydrogen production system considering the start-stop characteristics of electrolyzers. Electr. Power 2022, 55, 101–109. [Google Scholar]
Figure 1. Multi-time scale optimal scheduling architecture of electro-hydrogen coupling system based on modified TCN-PPO.
Figure 1. Multi-time scale optimal scheduling architecture of electro-hydrogen coupling system based on modified TCN-PPO.
Energies 18 01926 g001
Figure 2. Schematic diagram of the dilated convolution.
Figure 2. Schematic diagram of the dilated convolution.
Energies 18 01926 g002
Figure 3. Proposed modified TCN-PPO algorithm training process.
Figure 3. Proposed modified TCN-PPO algorithm training process.
Energies 18 01926 g003
Figure 4. Results of wind power and load forecast for the day-ahead stage.
Figure 4. Results of wind power and load forecast for the day-ahead stage.
Energies 18 01926 g004
Figure 5. Proposed modified TCN-PPO agent reward curve.
Figure 5. Proposed modified TCN-PPO agent reward curve.
Energies 18 01926 g005
Figure 6. Day-ahead stage power scheduling results for the electro-hydrogen coupling system.
Figure 6. Day-ahead stage power scheduling results for the electro-hydrogen coupling system.
Energies 18 01926 g006
Figure 7. Electrolyzer day-ahead production program.
Figure 7. Electrolyzer day-ahead production program.
Energies 18 01926 g007
Figure 8. Power adjustment values of the system in real-time stage.
Figure 8. Power adjustment values of the system in real-time stage.
Energies 18 01926 g008
Figure 9. Cumulative reward curves from 30 days of testing of different DRL algorithms.
Figure 9. Cumulative reward curves from 30 days of testing of different DRL algorithms.
Energies 18 01926 g009
Figure 10. Training reward curves of different model variants.
Figure 10. Training reward curves of different model variants.
Energies 18 01926 g010
Figure 11. Model reward and training time for different initial clipping rates.
Figure 11. Model reward and training time for different initial clipping rates.
Energies 18 01926 g011
Table 1. Industrial time-of-use electricity prices.
Table 1. Industrial time-of-use electricity prices.
Time IntervalPrice/(Yuan/kWh)
Off-peak period0:00–8:000.31
Mid-peak period12:00–17:00, 21:00–24:000.64
Peak period8:00–12:00, 17:00–21:001.07
Table 2. Algorithm parameterization.
Table 2. Algorithm parameterization.
ParametersValue
μ ESS /(yuan/kWh)0.05
μ EC /(yuan/kWh)0.5
μ HFC /(yuan/kWh)0.04
μ start /(yuan/time)234
λ 1 pun /(yuan/kWh)0.5
λ 2 pun /(yuan/kWh)0.5
λ 3 pun /(yuan/kWh)1
Table 3. A mean value of results for different DRL methods.
Table 3. A mean value of results for different DRL methods.
DDPGPPOTCN-PPOThe Modified TCN-PPO
Power purchase cost/yuan28,792.27 27,684.03 24,143.57 22,359.29
Operation cost/yuan18,852.45 18,020.57 17,498.99 17,405.65
Start–stop cost/yuan4500.33 4209.00 4147.67 3948.33
Table 4. Comparative training metrics of model variants.
Table 4. Comparative training metrics of model variants.
Day-Ahead StageReal-Time Stage
Episodes to ConvergenceVariance of RewardMean Converged RewardEpisodes to ConvergenceVariance of RewardMean Converged Reward
PPO20024,162.28−1983.62200614.18−61.23
TCN-PPO450107,223.05−1926.86600746.24−57.09
TCN-PPO-ACR30085,146.11−1893.84350724.66−54.62
TCN-PPO-PER40083,788.75−1867.43500852.21−52.39
Modified TCN-PPO35057,540.32−1813.37500797.06−49.95
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, D.; Qian, K.; Xu, Y.; Zhou, J.; Wang, Z.; Peng, Y.; Xing, Q. A Multi-Time Scale Optimal Scheduling Strategy for the Electro-Hydrogen Coupling System Based on the Modified TCN-PPO. Energies 2025, 18, 1926. https://doi.org/10.3390/en18081926

AMA Style

Li D, Qian K, Xu Y, Zhou J, Wang Z, Peng Y, Xing Q. A Multi-Time Scale Optimal Scheduling Strategy for the Electro-Hydrogen Coupling System Based on the Modified TCN-PPO. Energies. 2025; 18(8):1926. https://doi.org/10.3390/en18081926

Chicago/Turabian Style

Li, Dongsen, Kang Qian, Yiyue Xu, Jiangshan Zhou, Zhangfan Wang, Yufei Peng, and Qiang Xing. 2025. "A Multi-Time Scale Optimal Scheduling Strategy for the Electro-Hydrogen Coupling System Based on the Modified TCN-PPO" Energies 18, no. 8: 1926. https://doi.org/10.3390/en18081926

APA Style

Li, D., Qian, K., Xu, Y., Zhou, J., Wang, Z., Peng, Y., & Xing, Q. (2025). A Multi-Time Scale Optimal Scheduling Strategy for the Electro-Hydrogen Coupling System Based on the Modified TCN-PPO. Energies, 18(8), 1926. https://doi.org/10.3390/en18081926

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop