Next Article in Journal
A Post-Disaster Fault Recovery Model for Distribution Networks Considering Road Damage and Dual Repair Teams
Previous Article in Journal
Optimizing S Chemical Looping Combustion with Cu-Fe Combined Oxygen Carriers: Performance and Mechanistic Insights
Previous Article in Special Issue
Bidding Strategy for the Alliance of Prosumer Aggregators in the Distribution Market
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on Electric Hydrogen Hybrid Storage Operation Strategy for Wind Power Fluctuation Suppression

by
Dongsen Li
1,2,
Kang Qian
1,*,
Ciwei Gao
2,
Yiyue Xu
1,
Qiang Xing
3 and
Zhangfan Wang
1
1
Department of Integrated Energy Engineering, China Energy Engineering Group Jiangsu Power Design Institute Co., Ltd., Nanjing 211100, China
2
School of Electrical Engineering, Southeast University, Nanjing 210096, China
3
School of Automation and Artificial Intelligence, Nanjing University of Posts and Telecommunications, Nanjing 210003, China
*
Author to whom correspondence should be addressed.
Energies 2024, 17(20), 5019; https://doi.org/10.3390/en17205019
Submission received: 12 August 2024 / Revised: 12 September 2024 / Accepted: 1 October 2024 / Published: 10 October 2024

Abstract

:
Due to real-time fluctuations in wind farm output, large-scale renewable energy (RE) generation poses significant challenges to power system stability. To address this issue, this paper proposes a deep reinforcement learning (DRL)-based electric hydrogen hybrid storage (EHHS) strategy to mitigate wind power fluctuations (WPFs). First, a wavelet packet power decomposition algorithm based on variable frequency entropy improvement is proposed. This algorithm characterizes the energy characteristics of the original wind power in different frequency bands. Second, to minimize WPF and the comprehensive operating cost of EHHS, an optimization model for suppressing wind power in the integrated power and hydrogen system (IPHS) is constructed. Next, considering the real-time and stochastic characteristics of wind power, the wind power smoothing model is transformed into a Markov decision process. A modified proximal policy optimization (MPPO) based on wind power deviation is proposed for training and solving. Based on the DRL agent’s real-time perception of wind power energy characteristics and the IPHS operation status, a WPF smoothing strategy is formulated. Finally, a numerical analysis based on a specific wind farm is conducted. The simulation results based on MATLAB R2021b show that the proposed strategy effectively suppresses WPF and demonstrates excellent convergence stability. The comprehensive performance of the MPPO is improved by 21.25% compared with the proximal policy optimization (PPO) and 42.52% compared with MPPO.

1. Introduction

Due to the increasing scarcity of fossil fuels and the global need for production and carbon reduction, the installed capacity of renewable energy (RE) generation is rapidly developing throughout the world. From 2013 to 2023, the combined renewable electricity and biofuels primary energy input consumption worldwide increased from 52.93 exajoules to 90.23 exajoules, with an average growth rate of 5.5% [1]. Furthermore, the average rate of increase in wind and solar installed capacity to 2035 is supposed as 450–600 GW per year—around 1.9 to 2.5 times faster than the highest rate in the past [2]. The proportion of non-fossil energy supply, dominated by RE, in China is expected to reach 60% by 2050 [3]. To accommodate a high proportion of RE, measures must be taken to address the impact of its integration [4].
Unlike traditional power sources, renewable energy (RE) sources such as wind and solar power have significant intermittency and output fluctuations, weak support from power generation equipment, and low anti-interference ability [5]. These negative characteristics hinder achieving electricity balance in the power system [6]. Additionally, the integration of RE generation has introduced a large number of power electronic devices, such as inverters and DC converters, causing issues like transient overvoltage and broadband oscillations, which seriously impact power system stability. Therefore, it is urgent to introduce flexible resources to promote the consumption of a high proportion of RE and mitigate its adverse effects [7].
Hydrogen is a clean secondary energy source with high calorific value and diverse sources [8]. Compared to electrochemical energy storage (EES), hydrogen has a longer storage and discharge cycle and a larger storage capacity. The electrolysis-hydrogen storage-gas power generation cycle allows the coupling of electricity and hydrogen to balance electricity consumption effectively [9]. It can supply power to nearby loads and provide frequency and voltage support for the power grid [10].
This paper applies an electric hydrogen hybrid storage (EHHS) system to smooth wind power fluctuations. Many studies have examined the smoothing effect of hybrid energy storage systems composed of EES and supercapacitors on new energy generation. Although the stabilizing effect of EHHS on voltage and the reducing effect on net load fluctuations have been studied [11], further research on the power coordination and allocation strategy is necessary considering the particularity of the operating points between EES and the hydrogen system [12].
This paper aims to use improved wavelet characterization theory and deep reinforcement learning (DRL) technology to suppress the fluctuation of wind power based on EHHS. Common smoothing methods include wavelet transform and model predictive control. References [13,14,15] propose decoupling wind power and hybrid energy storage (HES) power using wavelet transform or low-pass filtering. For more robust battery output processing, reference [16] used adaptive variational mode decomposition to extract frequencies from different wind scenarios, enabling effective determination of the pre-scheduled power of the HES. Additionally, the PI controller adjustment strategy proposed in reference [17] for HES can effectively control the exchange of active and reactive power between wind farms and the power grid, thereby smoothing wind power fluctuations. Moreover, scholars have used predictive control algorithms to smooth wind power [18,19,20,21,22,23,24,25]. The basic idea is to optimize the charging and discharging behavior of HES through controlling the state of charge (SOC) of the battery, thereby smoothing wind power and improving battery life. Stochastic predictive control [22,24] and intelligent predictive control algorithms [20,25] can provide more precise predictions and management of wind power uncertainty. However, research on stabilizing RE sources with HES mainly focuses on EES and supercapacitors, with relatively little complementary research on EES and hydrogen energy system. For example, the literature [26] studied the interaction between wind power and hydrogen energy systems but did not include EES, resulting in a lack of consideration for the volatility of hydrogen energy storage. Consequently, it could not achieve optimal power flow management between wind power, hydrogen energy, and EES.
However, conventional predictive control algorithms generally require prior accurate predictions, as referenced in [27,28], which are difficult to achieve in real operating scenarios of wind farms. Moreover, methods like model predictive control (MPC) and fuzzy control have complex application rules, poor scalability, and limited generalization performance. Therefore, this paper applies improved wavelet characterization theory and DRL to the allocation of EHHS power. Scholars have applied DRL to study collaborative control between wind turbines and HES [29,30,31]. These studies have shown that DRL strategies can effectively manage the nonlinearity and randomness of wind power, maximizing rewards or achieving specific goals. However, DRL has not yet been applied for control in wind EHHS systems. This paper combines the complementary characteristics of EES and hydrogen energy over different time periods to study the real-time perception and power smoothing effects of DRL in this new energy storage system.
The major contribution of this article can be concluded as:
(1)
DRL algorithm is utilized for research on smoothing on-grid WPF. Fast perception of EHHS status and formulation of charging and discharging operation strategies through DRL intelligent agents significantly suppress the on-grid WPF.
(2)
A wavelet packet power decomposition algorithm based on variable frequency entropy improvement is proposed. This algorithm addresses the drawback of the wavelet packet decomposition (WPD) algorithm that requires precise input conditions and manual setting of response time boundary points for different energy storage components.
(3)
A modified proximal policy optimization (MPPO) based on wind power deviation is proposed for training and solving the unique challenges of WPF. By dynamically adjusting the clipping rate based on real-time WPF, the training efficiency and stability of the algorithm are balanced, and the overall performance of the model is improved.
The remainder of this paper is as follows. Section 2 introduces the energy flow architecture and information exchange forms of EHHS. Section 3 establishes the power allocation strategy and smoothing effect optimization function of EHHS. Section 4 explains the optimization method of the wind power stabilization strategy using the MPPO algorithm. Section 5 illustrates the improvement effects of the proposed MPPO and the advantages of EHHS through case studies. Finally, Section 6 presents the conclusions.

2. Architecture of Electric Hydrogen Hybrid Storage System

The architecture of EHHS is shown in Figure 1. It illustrates the information exchange between wind farms and the grid, including actual on-grid power transmission, maximum wind power fluctuation, and wind power prediction. The power allocation controller manages the output strategy of EHHS. First, the controller uses an improved wavelet transform method to decompose the original wind power and obtain the direct on-grid power and fluctuation smoothing scale for EHHS. Simultaneously, the EES and HES provide real-time SOC and SOH information to the controller. After optimization using the DRL algorithm, the charging/discharging strategy is sent back to the energy storage system, which then interacts with the power grid through a power converter.

3. Wind Power Fluctuation Suppression Strategy of EHHS

The proposed EHHS wind power fluctuation suppression strategy is shown in Figure 2. First, an improved WPD algorithm is used to decompose the real-time output of wind turbines and characterize wind power in different frequency bands. Second, the state parameters (SOC, SOH, etc.) of EHHS are combined with the WPD results to form a deep reinforcement learning proximal policy optimization (PPO) agent. Using the Actor network, the agent interacts with the environment and accumulates training samples. Then, an improved PPO algorithm based on power deviation is proposed for optimization training, learning the optimal mapping relationship between environmental states and action strategies. Finally, the trained PPO agent is deployed online to smooth real-time fluctuations in wind power.

3.1. Wavelet Packet Power Decomposition Algorithm Based on Frequency Conversion Entropy Improvement

3.1.1. Traditional Wavelet Packet Decomposition Algorithm

The WPD algorithm decomposes the original wind power into low-frequency and high-frequency signals. Compared with traditional wavelet decomposition, WPD offers higher frequency decomposition rates and better time resolution. The algorithm decomposes the original signal through x layers, maps it to 2x wavelet packet spaces, and then reconstructs the 2x components of the x-th layer.
α n k , 2 x = i = 1 x u i 2 n α i k + 1 , x α n k , 2 x + 1 = i = 1 x v i 2 n α i k + 1 , x
where n denotes the number of signals obtained from wavelet decomposition, n = 1 , 2 , , 2 x ; α k , 2 x , α k , 2 x + 1 denote the low-frequency and high-frequency coefficients, u i 2 n , v i 2 n denote the decompose coefficients for low-pass and high-pass filters. The reconstruction algorithm is:
α n j + 1 = i = 1 x u n 2 i α i j , 2 x + v n 2 i α i 2 x + 2
The wavelet packet algorithm is first used to decompose and obtain the wind power output power. The xth-layer 0-node power component is taken as the grid reference power, and the others are taken as the EHHS reference power. Based on the analysis of the EHHS operating characteristics, the power component frequency division point i is determined, leading to the final EHHS power command. The power commands for EES and HES are detailed below:
P EES t = 1 i S x , i P HS t = i + 1 2 x S x , i
where P EES t , P HS t denote the power commands for EES and HES at time t, respectively; when both values are negative, it indicates that EES and the fuel cell are discharging; when both are positive, it indicates that EES is charging, and the electrolytic cell is absorbing power.

3.1.2. Variable Frequency Entropy Strategy Based on WPD

Although the WPD algorithm offers high time and frequency resolution, its application is limited because it generally requires precise input conditions and manual setting of response time boundary points for different energy storage components. To address the drawbacks of the WPD algorithm, this paper employs a frequency conversion entropy strategy based on the WPD. By combining this with the specific characteristics of EHHS, the approach optimizes the time and frequency boundary points, leading to the optimal and most economical power allocation strategy for EHHS.
Based on the WPD algorithm, the n-th node coefficient is set as B j , n = B j , n x , x = 1 , 2 , , N and the sampling interval is t. The reconstructed trajectory matrix is obtained as follows:
A j , n L M = α j , n 1 α j , n 1 + t α j , n 1 + M 1 t α j , n 2 α j , n 2 + t α j , n 2 + M 1 t α j , n L α j , n L + t α j , n L + M 1 t
where L denotes the sliding window length when performing sliding average filtering on the original wind power.
The singular values of wavelet decomposition can be obtained from Equation (4), λ j , n x 1 x m + 1 t , which can reflect the proportion of energy. Assuming that the signal energy obtained from each order of WPD is E k , the total energy is:
E k = x = 1 N λ j , n x ln λ j , n x
After normalizing Equation (5), the entropy value, P k , can be obtained as follows:
P k = E k E 0 = x = 1 N λ j , n x ln λ j , n x n = 1 2 x E n 2
The corresponding entropy spectrum can be obtained as follows:
H j , n λ j , n x = P k L M
Take the order with the fastest decrease in normalized entropy value in the entropy spectrum as the boundary point D. Based on this principle, the power command for EHHS, derived from the improved wavelet packet algorithm, is as follows:
P EES H t = 1 D S x , D P HS H t = D + 1 2 x S x , D
where x′ is the number of WPD layers determined based on entropy changes.

3.2. Modeling of Wind Power Fluctuation Suppression

3.2.1. Objective Function

In this section, the objective functions are categorized into two types of assessment: stabilization effect and operational cost.
A
The optimal effect of WPF suppression
Generally, the system’s ability to withstand fluctuations in wind power is determined via its frequency regulation capability:
P G t P G t 1 Δ P WT
where Δ P WT denotes the tolerable WPF.
This paper evaluates the degree of WPF through measuring the total amplitude that exceeds the limit. One of the objective functions is to minimize this fluctuation.
Δ Y = t = 1 T / Δ t P t P t 1 Δ P WT
where T denotes the sampling period.
B
The optimal operating cost of EHHS
The operating costs of EHHS encompass those of both HES and EES, as indicated by the following equation:
C HS + C EES = C HS , o m + C EES , o m + C HS , l o s s + C EES , l o s s
where C HS , C EES are the costs of HES and EES, respectively. The subscripts om, loss, respectively, represent operation and consumption.
The operation and maintenance costs of HES and EES are detailed below:
C HS , o m = t = 1 T i = 1 N PE σ PE , i , t C PE , i , i n v T PE , i , l + i = 1 N FC σ FC , i , i n v T FC , i , l
C EES , o m = γ EES , o m t = 1 T i = 1 N EES P EES , i , t d i s + P EES , i , t c h a
where σ PE , i , t , σ FC , i , t denote the 0–1 state variable of i-th Proton exchange membrane electrolyser (PEMEL) or Proton exchange membrane fuel cell (PEMFC), 0 indicates startup, 1 indicates shutdown. The subscript inv represents the construction cost; T PE , i , l , T FC , i , l denote the maximum operating time; γ EES , o m denotes the cost coefficient of EES charging and discharging; N denotes the number of device; P EES , i , t d i s , P EES , i , t c h a denote the discharge and charging power of the i-th battery at time t.
There is energy conversion loss in the energy storage process. The cost associated with these losses for HES and EES is considered as follows:
C HS , l o s s = γ d t = 1 T i = 1 N PE P i , t PE 1 η PE , i , t + i = 1 N FC P i , t FC 1 η FC , i
C EES , l o s s = γ d t = 1 T i = 1 N EES P EES , i , t d i s 1 η EES , i , d i s + P EES , i , t c h a 1 η EES , i , c h a
where γ d denotes the transmission and distribution tariff; η denotes the operating efficiency.
Restraints
A
Power balance
P WT t + P EES t = P load t + P PE t
where P load t denotes the fluctuating load; P PE t denotes the electrolytic cell power.
B
Unit time exchange power of EES
0 < P c h a t < P c h a max , P d i s t = 0 0 < P d i s t < P d i s max , P c h a t = 0
where P c h a max , P d i s max denote the upper limits of real-time charging and discharging power for EES.
C
Electrolytic cell operation
When the operating power of the electrolytic cell falls below a certain threshold, there is a risk that the hydrogen–oxygen mixture may exceed the explosive limit. The hydrogen production capacity of the electrolytic cell can be expressed in the following equation:
P PE min < P PE t < P PE max , P H t = η PE P PE t P PE t = 0 , P H = 0
where P PE min , P PE max denote the lower and upper ranges of the operating power of the electrolytic cell; P H t denotes the hydrogen production; η PE denotes the efficiency of hydrogen production.
D
Hydrogen storage status
SOH min SOH t SOH max
where SOH t denotes the status for hydrogen storage tanks.
E
Upper and lower limits of SOC for EES
SOC min SOC t SOC max
where SOC t denotes the status of EES batteries.

3.2.2. WPF Suppression Strategy Based on Markov Decision Process

To mitigate the impact of power generation volatility in renewable energy systems on wind power smoothing algorithms, this section converts the model into a Markov Decision Process (MDP) and trains it, using a MPPO algorithm with deep reinforcement learning.
As a deep reinforcement learning algorithm based on the Actor–Critic architecture, PPO agents observe states of the IPHS, s t , and make real-time decision actions, a t , based on the Actor network. After executing the actions a t , the agents will receive immediate rewards, r t . Repeat the above interaction to continuously accumulate samples and update the parameters of the agent network, ultimately achieving the mapping between environmental state and optimal action.
A
States
The state represents the agent’s perception of real-time environmental information to assist in decision-making. In this paper, the environmental state includes original wind power and high- and low-frequency components obtained by WPD, SOC, and SOH:
s t = P WT , t , F WT , t , S O C EES , t , S O H HS , t
where P WT , t , F WT , t denote the original wind power and high- and low-frequency components obtained from WPD, respectively.
B
Actions
Based on environmental information, the agent selects an action from the action set [32]. For EHHS, its operating space consists of the outputs of EES and HES.
a t = P EES , t , P HS , t
C
Rewards
Rewards are real-time feedback from the environment following the execution of an action, used to guide the agent toward better decision-making. In this paper, the reward comprises the cost of EHHS and a penalty for exceeding the WPD limits:
r t = r t EHHS + r t PU
r t EHHS = C EES , t C HS , t
r t PU = λ WT P WT , t P WT , t 1 , P WT , t P WT , t 1 < Δ P WT F 500 , P WT , t P WT , t 1 Δ P WT F
where r t EHHS , r t PU denote the cost of EHHS and the penalty, respectively; C EES , t , C HS , t denote the cost of EES and HES; λ WT denotes the penalty coefficient; Δ P WT F denotes the WPD limit.

4. Solution Based on the MPPO Algorithm

4.1. Basic Principles of PPO Algorithm

In the PPO algorithm, the advantage function at time t, A t a , s , can be expressed as follows:
A t a , s = V Ψ s t + r t + β r t + 1 + + β K t + 1 r K 1 + β K t V Ψ s K
where V Ψ s t denotes the state value function; β denotes the depreciation rate; K denotes the step length.
According to Equation (26), the training objective F is to obtain the agent network parameters that can maximize the expected advantage function:
F = max θ   t π θ k + 1 a s π θ k a s A t a , s = max θ   t τ t A t a , s
where π θ k denotes the function of strategy; τ t denotes the ratio of old to new strategies.
In order to ensure training stability, τ t has been limited to ensure that its value is within 1 ε , 1 + ε . The objective function can be modified to the following:
F clip = max   t min τ t A t , clip τ t A t
clip τ t = 1 ε τ t < 1 ε τ t 1 ε τ t 1 ε 1 + ε τ t > 1 + ε
where ε denotes the clipping rate.

4.2. Adaptive Clipping Rate Mechanism Based on Power Fluctuations

The proposed MPPO algorithm aims to dynamically adjust the clipping rate based on sample quality, balancing training speed, and algorithm stability for varying samples.
A heuristic search algorithm is employed for the optimization strategy. In the algorithm, the crossover factor (CR) significantly impacts performance. A small CR value can reduce population diversity and result in local optima. Conversely, a large value can decrease the search accuracy. Therefore, this paper proposes an improved differential evolution algorithm that accounts for wind power deviation.
C R = 0.5 Δ P t 2 / ξ + 0.2 , Δ P t 2 ξ r a n d ( 0.1 , 0.5 ) , Δ P t 2 > ξ
Δ P t 2 = i = 1 N P i , t P i r P i max P i min 2
where Δ P t 2 denotes the sum of squares of wind power deviation at time t; ξ denotes the threshold; rand denotes random number generation function.
Equation (30) shows that when the power deviation is small, the value of the cross factor CR is small to improve the search accuracy and convergence speed. When the power deviation is large, CR becomes a random number within the range of (0.1, 0.5) to ensure the richness of the population.
Notably, a large time difference error (TD-error) suggests that the current strategy is ineffective, necessitating a larger clipping rate, whereas a small TD-error implies that the strategy is effective, requiring a reduction in clipping rate to maintain convergence stability. The dynamic clipping rate is calculated as follows:
ε = ε max δ t δ t ¯ ε min δ t < δ t ¯
where ε max , ε min denote the maximum and minimum clipping rate; δ t denotes the time difference error; δ t ¯ denotes the mean error.

4.3. The Training Process of the Improved PPO Algorithm

Figure 3 illustrates the training process of the improved PPO algorithm. Initially, the training environment is set up, followed by the decomposition of real-time wind turbine output using the enhanced WPD, which characterizes wind power energy across different frequency bands. The real-time status is observed and the action decisions are made. Secondly, the reward, r t , is calculated and the sample, s t , a t , r t , s t + 1 , is stored in the experience pool. Next, the real-time WPF is calculated and the clipping rate, ε , is dynamically adjusted, and the parameters of the agent network are updated. These steps are repeated until the maximum number of training rounds is reached, after which the trained agent is finalized.

5. Case Study

5.1. Configuration and Parameter Setting of IPHS

This paper uses actual wind power data from a city in northwest China in 2020 and is configured with EES and HES [33]. The total installed capacity in the region is 2800 MW, comprising 1700 MW from thermal units and 1100 MW from wind units. Each wind power unit has a capacity of 5 MW. The total capacity of the HES is set at 5% to 10% of the total capacity. The charging and discharging duration for HES is assumed to be 10 h. The general cost coefficients of EES and HES are 0.005 CNY/kWh and 0.02 CNY/kWh, respectively. The WPF penalty coefficient, λ W , is set at 1 CNY/MW. The capacity configuration of EHHS is detailed in Table 1.

5.2. Analysis of Training Process

Assuming a discount rate of 0.95, an experience pool capacity of 8000, a mini-batch size of 128, and clipping rates of 0.3 and 0.1 for the maximum and minimum, respectively, the offline training reward curve of the agent is shown in Figure 4. The smoothed reward is added as the auxiliary line to better showcase the overall trend in the reward curve.
Figure 4 illustrates that the reward curve for the agent shows a significant upward trend in fluctuation. In the first stage, the agent continuously explores the training environment and obtains low and fluctuating reward values, with an average of −110.23, indicating a risk of exceeding the on-grid power limit. However, due to the PPO improvement mechanism proposed in this paper, which adapts the clipping rate based on the WPF limit, the efficiency of network parameter updates improved early in the training process, resulting in a notable increase in the reward curve between rounds 340 and 850. In the later stage of training (Stage 3), the model employed a smaller clipping rate to maintain convergence stability, resulting in the reward value converging to approximately −20.20. This demonstrates that the agent had effectively learned the optimal mapping between environmental state and action decisions.

5.3. WPF Suppression Model Application and Results Analysis

The comparison of the on-grid wind power and fluctuation before and after suppression is shown in Figure 5 and Figure 6, respectively.
The results show that the proposed modified WPD algorithm can better characterize wind power, which is helpful for power scheduling and utilization. In addition, as shown in Figure 6, there is a significant problem of the WPF limit being exceeded before suppression, especially around 3:15 am, when the power fluctuation reached 75.17 MW, exceeding the threshold by 36.67% (Figure 6 is an analysis chart for 24 h, a day and a night). The average fluctuation of wind power throughout the day before suppression reached as high as 20.11 MW, which posed a significant challenge to the stable operation of the power grid. On the contrary, the proposed algorithm effectively reduces the fluctuation amplitude to an average of 5.74 MW, representing a 71.46% reduction compared with before. Furthermore, based on the proposed method, the scenarios of short-term severe fluctuations in wind power were significantly reduced, markedly enhancing the power grid’s ability to accommodate wind energy.
By conducting online testing on the offline trained agent, the agent can formulate an EHHS operation strategy online, which is shown in Figure 7. Comparing Figure 7a,b, it can be seen that the operation strategy of EES is more flexible and can quickly adjust the charging and discharging power to mainly absorb the low-frequency components of wind power fluctuations. On the contrary, due to physical characteristics, the charging and discharging strategies of an electrolysis system with storage tend to be more conservative. The proposed method can effectively mitigate wind power grid fluctuations by coordinating the implementation and operation strategies of EHHS. In addition, the SOC curve of EES indicates that it mostly operates in a shallow charging and discharging state, which helps extend its service life. The SOH of HES generally ranged between 0.5 and 0.7, with sufficient power and capacity space to smooth out fluctuations.

5.4. Comparison of Different Algorithms

Finally, to validate the superiority of the proposed MPPO algorithm, Deep Deterministic Policy Gradient (DDPG) and PPO algorithms were used as control benchmarks for testing. The training reward curves for different DRL algorithms are shown in Figure 8. The result shows that the DDPG algorithm achieved convergence in about 600 rounds, but its stability was poor in the later stage, with an average reward of only −35.14. Although the PPO algorithm converged more slowly than the DDPG, it effectively mitigated instability during training by limiting the update step size. The reward value remained relatively stable in later stages, approximately −25.65. The proposed MPPO algorithm combined the PPO’s stability with dynamic clipping rate adjustments based on WPF to enhance early training efficiency and ensure stability in later stages. The reward value at the late stage of training of the MPPO algorithm was −20.20. The reward value increased by 21.25% from −25.65 to −20.20 compared with the conventional PPO algorithm, demonstrating the effectiveness and superiority of the proposed mechanism in WPF suppression.
The system costs and on-grid power fluctuations using different DRL algorithms are shown in Table 2.
According to Table 2, the proposed MPPO algorithm significantly outperformed in terms of cost and mean on-grid power. The proposed method utilized a WPD algorithm enhanced via frequency conversion entropy to characterize the energy features of the original wind power, optimizing the comprehensive costs of EES and HES. The MPPO algorithm was also used to achieve efficient management and scheduling of EES and HES. Compared with the PPO, the cost of EES and HES was reduced by 6.62% and 18.31%, respectively. Regarding on-grid power fluctuations, the MPPO algorithm dynamically adjusted strategy parameters based on real-time fluctuations, enabling a rapid response to WPF. The average of on-grid power fluctuations is 5.74 MW, approximately 16.81% lower than the PPO, significantly enhancing the stability and reliability of on-grid wind power. It can be seen from Table 2 that the three DRL algorithms have different effects on the suppression of WPF. The PPO represented the policy using a probability distribution and achieved policy updates through limiting the magnitude of changes between the new and old policies, thus obtaining more stable results in case experiments. The DDPG algorithm directly outputted actions instead of a probability distribution of actions. In some cases, DDPG adopted more aggressive policy updates, leading to greater fluctuations in wind power. The proposed MPPO algorithm dynamically adjusted strategy parameters based on real-time fluctuations, enabling a rapid response to WPF and achieving excellent results in various indicators.

6. Conclusions

This paper proposes an EHHS strategy based on a MPPO algorithm to mitigate real-time fluctuations of wind power. A WPD enhanced via frequency conversion entropy was employed to characterize the energy features of wind power across different frequency bands, while an MPPO algorithm based on wind power deviation was introduced to develop WPF suppression strategies.
We conducted a numerical simulation based on a wind farm and obtained the following conclusions:
(1)
This paper explores the energy flow and complementary characteristics of EHHS based on a DRL algorithm, achieving real-time perception of system status. By formulating power charging and discharging strategies for EES and HES, WPF is effectively mitigated and the overall system cost is reduced.
(2)
The proposed modified WPD algorithm can accurately characterize the wind power, thereby formulating high- and low-frequency power allocation plans. The average on-grid WPF was only 5.74 MW, a decrease of 71.46% compared with before suppression.
(3)
Compared with other DRL algorithms, the proposed MPPO algorithm can dynamically adjust the clipping rate based on WPF during the training process, effectively balancing the training efficiency and convergence stability. Compared with the conventional PPO algorithm, the MPPO algorithm increased the training reward value by 21.25% and reduced the on-grid WPF by 16.81%.
Future work will involve integrating long-term forecasting, mid-term scheduling, and short-term real-time control to enhance and optimize EHHS charging and discharging strategies, aiming for coordinated scheduling across multiple time scales. Additionally, conducting empirical research based on a Hardware-in-the-Loop platform is also a direction for our next steps.

Author Contributions

Conceptualization, D.L. and K.Q.; methodology, D.L. and C.G.; validation, D.L. and Q.X.; formal analysis, C.G. and Q.X.; investigation, Y.X.; resources, K.Q.; data curation, Z.W.; writing—original draft preparation, K.Q. and C.G.; writing—review and editing, D.L.; visualization, C.G. and Y.X.; supervision, Y.X.; project administration, D.L. and K.Q.; funding acquisition, D.L. and K.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Special Funds for Technological Innovation of Jiangsu Province (BE2022040), the China Postdoctoral Science Foundation (2023M733780), and the Technology Project of China Energy Engineering Group Jiangsu Power Design Institute Co., Ltd. (32-JK-2024-033).

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

D.L., K.Q., Y.X. and Z.W. were employed by China Energy Engineering Group Jiangsu Power Design Institute Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The authors declare that this study received funding from China Energy Engineering Group Jiangsu Power Design Institute Co., Ltd. The funder was not involved in the study design, collection, analysis, interpretation of data, the writing of this article or the decision to submit it for publication.

References

  1. Energy Institute KPMG. Statistical Review of World Energy 2024. 2024. pp. 58–65. Available online: https://www.energyinst.org/statistical-review (accessed on 30 August 2024).
  2. BP. bp Energy Outlook. 2024, 57. Available online: https://www.bp.com/content/dam/bp/business-sites/en/global/corporate/pdfs/energy-economics/energy-outlook/bp-energy-outlook-2024.pdf (accessed on 30 August 2024).
  3. Jie, D.; Xu, X.; Guo, F. The future of coal supply in China based on non-fossil energy development and carbon price strategies. Energy 2021, 220, 119644. [Google Scholar] [CrossRef]
  4. Lund, H. Renewable energy strategies for sustainable development. Energy 2007, 32, 912–919. [Google Scholar] [CrossRef]
  5. Hu, H.; Ma, C.; Wang, X.; Zhang, Z.; Aizezi, A. Anti-Interference System for Power Sensor Signals under Artificial Intelligence. In Proceedings of the 2024 Second International Conference on Data Science and Information System (ICDSIS), Hassan, India, 17–18 May 2024; pp. 1–5. [Google Scholar]
  6. Beaudin, M.; Zareipour, H.; Schellenberglabe, A.; Rosehart, W. Energy storage for mitigating the variability of renewable electricity sources: An updated review. Energy Sustain. Dev. 2010, 14, 302–314. [Google Scholar] [CrossRef]
  7. Lund, P.D.; Lindgren, J.; Mikkola, J.; Salpakari, J. Review of energy system flexibility measures to enable high levels of variable renewable electricity. Renew. Sustain. Energy Rev. 2015, 45, 785–807. [Google Scholar] [CrossRef]
  8. Hosseini, S.E.; Wahid, M.A. Hydrogen production from renewable and sustainable energy resources: Promising green energy carrier for clean development. Renew. Sustain. Energy Rev. 2016, 57, 850–866. [Google Scholar] [CrossRef]
  9. He, G.; Mallapragada, D.S.; Bose, A.; Heuberger, C.F.; Gencer, E. Hydrogen supply chain planning with flexible transmission and storage scheduling. IEEE Trans. Sustain. Energy 2021, 12, 1730–1740. [Google Scholar] [CrossRef]
  10. Serban, I.; Marinescu, C. Control strategy of three-phase battery energy storage systems for frequency support in microgrids and with uninterrupted supply of local loads. IEEE Trans. Power Electron. 2013, 29, 5010–5020. [Google Scholar] [CrossRef]
  11. Li, J.; Yang, B.; Huang, J.; Guo, Z.; Wang, J.; Zhang, R.; Hu, Y.; Shu, H.; Chen, Y.; Yan, Y. Optimal planning of Electricity-Hydrogen hybrid energy storage system considering demand response in active distribution network. Energy 2023, 273, 127142. [Google Scholar] [CrossRef]
  12. Vivas, F.J.; De las, H.A.; Segura, F.; Andújar Márquez, J.M. A review of energy management strategies for renewable hybrid energy systems with hydrogen backup. Renew. Sustain. Energy Rev. 2018, 82, 126–155. [Google Scholar] [CrossRef]
  13. Mannelli, A.; Papi, F.; Pechlivanoglou, G.; Ferrara, G.; Bianchini, A. Discrete wavelet transform for the real-time smoothing of wind turbine power using li-ion batteries. Energies 2021, 14, 2184. [Google Scholar] [CrossRef]
  14. Roy, P.; He, J.; Liao, Y. Cost minimization of battery-supercapacitor hybrid energy storage for hourly dispatching wind-solar hybrid power system. IEEE Access 2020, 8, 210099–210115. [Google Scholar] [CrossRef]
  15. Guo, T.; Liu, Y.; Zhao, J.; Zhu, Y.; Liu, J. A dynamic wavelet-based robust wind power smoothing approach using hybrid energy storage system. Int. J. Electr. Power Energy Syst. 2020, 116, 105579. [Google Scholar] [CrossRef]
  16. Wan, C.; Qian, W.; Zhao, C.; Song, Y.; Yang, G. Probabilistic forecasting based sizing and control of hybrid energy storage for wind power smoothing. IEEE Trans. Sustain. Energy 2021, 12, 1841–1852. [Google Scholar] [CrossRef]
  17. Qais, M.H.; Hasanien, H.M.; Alghuwainem, S. Output power smoothing of wind power plants using self-tuned controlled SMES units. Electr. Power Syst. Res. 2020, 178, 106056. [Google Scholar] [CrossRef]
  18. Lin, L.; Cao, Y.; Kong, X.; Lin, Y.; Jia, Y.; Zhang, Z. Hybrid energy storage system control and capacity allocation considering battery state of charge self-recovery and capacity attenuation in wind farm. J. Energy Storage 2024, 75, 109693. [Google Scholar] [CrossRef]
  19. Carvalho, W.C.; Bataglioli, R.P.; Fernandes, R.; Coury, D.V. Fuzzy-based approach for power smoothing of a full-converter wind turbine generator using a supercapacitor energy storage. Electr. Power Syst. Res. 2020, 184, 106287. [Google Scholar] [CrossRef]
  20. Syed, M.A.; Khalid, M. An intelligent model predictive control strategy for stable solar-wind renewable power dispatch coupled with hydrogen electrolyzer and battery energy storage. Int. J. Energy Res. 2023, 2023, 4531054. [Google Scholar] [CrossRef]
  21. Guo, T.; Zhu, Y.; Liu, Y.; Gu, C.; Liu, J. Two-stage optimal MPC for hybrid energy storage operation to enable smooth wind power integration. IET Renew. 2020, 14, 2477–2486. [Google Scholar] [CrossRef]
  22. Wu, C.; Gao, S.; Liu, Y.; Han, H.; Jiang, S. Wind power smoothing with energy storage system: A stochastic model predictive control approach. IEEE Access 2021, 9, 37534–37541. [Google Scholar] [CrossRef]
  23. Bao, W.; Wu, Q.; Ding, L.; Huang, S.; Teng, F.; Terzija, V. Synthetic inertial control of wind farm with BESS based on model predictive control. IET Renew. 2020, 14, 2447–2455. [Google Scholar] [CrossRef]
  24. Liu, X.; Feng, L.; Kong, X. A comparative study of robust MPC and stochastic MPC of wind power generation system. Energies 2022, 15, 4814. [Google Scholar] [CrossRef]
  25. Chen, X.; Cao, W.; Zhang, Q.; Hu, S.; Zhang, J. Artificial intelligence-aided model predictive control for a grid-tied wind-hydrogen-fuel cell system. IEEE Access 2020, 8, 92418–92430. [Google Scholar] [CrossRef]
  26. Schrotenboer, A.H.; Veenstra, A.; Broek, M.; Ursavas, E. A green hydrogen energy system: Optimal control strategies for integrated hydrogen storage and power generation with wind energy. Renew. Sustain. Energy Rev. 2022, 168, 112744. [Google Scholar] [CrossRef]
  27. Cao, D.; Hu, W.; Zhao, J.; Zhang, G.; Zhang, B.; Liu, Z.; Chen, Z.; Blaabjerg, F. Reinforcement learning and its applications in modern power and energy systems: A review. J. Mod. Power Syst. Clean Energy 2020, 8, 1029–1042. [Google Scholar] [CrossRef]
  28. Huang, S.; Li, P.; Yang, M.; Gao, Y.; Yun, J.; Zhang, C. A control strategy based on deep reinforcement learning under the combined wind-solar storage system. IEEE Trans. Ind. Appl. 2021, 57, 6547–6558. [Google Scholar] [CrossRef]
  29. Wang, X.; Zhou, J.; Qin, B.; Guo, L. Coordinated control of wind turbine and hybrid energy storage system based on multi-agent deep reinforcement learning for wind power smoothing. J. Energy Storage 2023, 57, 106297. [Google Scholar] [CrossRef]
  30. Yin, X.; Lei, M. Jointly improving energy efficiency and smoothing power oscillations of integrated offshore wind and photovoltaic power: A deep reinforcement learning approach. PCMP 2023, 8, 1–11. [Google Scholar] [CrossRef]
  31. Chen, P.; Han, D. Reward adaptive wind power tracking control based on deep deterministic policy gradient. Appl. Energy 2023, 348, 121519. [Google Scholar] [CrossRef]
  32. Liang, T.; Sun, B.; Tan, J.; Cao, X.; Sun, H. Scheduling scheme of wind-solar complementary renewable energy hydrogen production system based on deep reinforcement learning. High Volt. Eng. 2023, 49, 2264–2274. (In Chinese) [Google Scholar]
  33. Yuan, T.; Guo, J.; Yang, Z.; Feng, Y.; Wang, J. Optimal allocation of power electric-hydrogen hybrid energy storage of stabilizing wind power fluctuation. Proc. CSEE 2024, 44, 1397–1405. (In Chinese) [Google Scholar]
Figure 1. EHHS system architecture.
Figure 1. EHHS system architecture.
Energies 17 05019 g001
Figure 2. Overall architecture of wind power smoothing strategy based on EHHS.
Figure 2. Overall architecture of wind power smoothing strategy based on EHHS.
Energies 17 05019 g002
Figure 3. Training flowchart for the MPPO algorithm.
Figure 3. Training flowchart for the MPPO algorithm.
Energies 17 05019 g003
Figure 4. MPPO intelligent agent training reward curve.
Figure 4. MPPO intelligent agent training reward curve.
Energies 17 05019 g004
Figure 5. Wind power decomposition results using wavelet packet algorithm.
Figure 5. Wind power decomposition results using wavelet packet algorithm.
Energies 17 05019 g005
Figure 6. The grid-connected fluctuation limit.
Figure 6. The grid-connected fluctuation limit.
Energies 17 05019 g006
Figure 7. Operation strategy of EHHS. (a) Charging and discharging power and SOC of EES. (b) Charging and discharging power and SOH of electrolysis system with storage.
Figure 7. Operation strategy of EHHS. (a) Charging and discharging power and SOC of EES. (b) Charging and discharging power and SOH of electrolysis system with storage.
Energies 17 05019 g007
Figure 8. Comparison of reward curves after training with different DRL algorithms.
Figure 8. Comparison of reward curves after training with different DRL algorithms.
Energies 17 05019 g008
Table 1. The capacity configuration of EHHS.
Table 1. The capacity configuration of EHHS.
DeviceCapacity
EES power/MW60
EES capacity/MWh150
Electrolytic cell power/MW50
HES capacity/m32100
Fuel cell power/MW50
Table 2. The results using different DRL algorithms.
Table 2. The results using different DRL algorithms.
DeviceDDPGPPOMPPO
Cost of EES/CNY1749.311387.161295.33
Cost of HES/CNY4127.883512.282869.12
Average of on-grid power fluctuations/MW9.526.905.74
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, D.; Qian, K.; Gao, C.; Xu, Y.; Xing, Q.; Wang, Z. Research on Electric Hydrogen Hybrid Storage Operation Strategy for Wind Power Fluctuation Suppression. Energies 2024, 17, 5019. https://doi.org/10.3390/en17205019

AMA Style

Li D, Qian K, Gao C, Xu Y, Xing Q, Wang Z. Research on Electric Hydrogen Hybrid Storage Operation Strategy for Wind Power Fluctuation Suppression. Energies. 2024; 17(20):5019. https://doi.org/10.3390/en17205019

Chicago/Turabian Style

Li, Dongsen, Kang Qian, Ciwei Gao, Yiyue Xu, Qiang Xing, and Zhangfan Wang. 2024. "Research on Electric Hydrogen Hybrid Storage Operation Strategy for Wind Power Fluctuation Suppression" Energies 17, no. 20: 5019. https://doi.org/10.3390/en17205019

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop