Next Article in Journal
Life Cycle Assessment of Poplar Biomass for High Value Products and Energy
Previous Article in Journal
Eco-Sustainable Energy Production in Healthcare: Trends and Challenges in Renewable Energy Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on Q-Table Design for Maximum Power Point Tracking-Based Reinforcement Learning in PV Systems

1
NARI Technology Development Co., Ltd., Nanjing 211106, China
2
Department of Electrical and Automation Engineering, Xianlin Campus, Nanjing Normal University, Nanjing 210023, China
*
Author to whom correspondence should be addressed.
Energies 2023, 16(21), 7286; https://doi.org/10.3390/en16217286
Submission received: 6 September 2023 / Revised: 13 October 2023 / Accepted: 20 October 2023 / Published: 27 October 2023
(This article belongs to the Section A2: Solar Energy and Photovoltaic Systems)

Abstract

:
Photovoltaic (PV) power generation is considered to be a clean energy source. Solar modules suffer from nonlinear behavior, which makes the maximum power point tracking (MPPT) technique for efficient PV systems particularly important. Conventional MPPT techniques are easy to implement but require fine tuning of their fixed step size. Unlike conventional MPPT, the MPPT based on reinforcement learning (RL-MPPT) has the potential to self-learn to tune step size, which is more adaptable to changing environments. As one of the typical RL algorithms, the Q-learning algorithm can find the optimal control strategy through the learned experiences stored in a Q-table. Thus, as the cornerstone of this algorithm, the Q-table has a significant impact on control ability. In this paper, a novel Q-table of reinforcement learning is proposed to maximize tracking efficiency with improved Q-table update technology. The proposed method discards the traditional MPPT idea and makes full use of the inherent characteristics of the Q-learning algorithm such as its fast dynamic response and simple algorithm principle. By establishing six kinds of Q-tables based on the RL-MPPT method, the optimal discretized state of a photovoltaic system is found to make full use of the energy of the photovoltaic system and reduce power loss. Therefore, under the En50530 dynamic test standard, this work compares the simulation and experimental results and their tracking efficiency using six kinds of Q-table, individually.

1. Introduction

The PV power generation system is one of the most critical renewable energy systems since it has several advantages. The most significant of these are that they provide clean, accessible, and infinite resources [1]. PV arrays have different output power and voltage distributions at each irradiance/temperature level. This output power is nonlinear, and there is only one maximum power point under typical conditions. Many algorithms have been described in the literature for making PV modules at this point. Maximum power point tracking (MPPT) is the term for these algorithms [2,3]. Classic MPPT technologies are frequently used in practice due to their simplicity [4].
For example, the P&O method is simple and only involves using devices to detect photovoltaic current and voltage [5,6,7]. However, as the name suggests, this method constantly interferes with the operation of the converter by increasing/decreasing the fixed step size (FSS) to the photovoltaic source, causing oscillations in the output power. Another factor that has to be considered is that photovoltaic systems often exhibit a certain degree of uncertainty and volatility. For example, clouds and birds can cause significant changes in the photovoltaic P–V curve, at which point the FSS used leads to two main challenges: tracking speed and tracking accuracy. While a large FSS may increase the tracking speed, it may also lead to steady-state oscillations and increased power loss. In contrast, using a small FSS smoothes the oscillations but results in slow transients. These issues should be fully considered when designing a MPPT system for improved power tracking performance.
Many control strategies have been widely used in MPPT at the present stage, the most typical example being the variable step size (VSS) method [8,9,10]. In order to achieve a balance between speed and accuracy, the VSS method reduces the step size to minimize oscillations; a larger step size is adopted during transients to allow a fast speed. However, in order to adapt to changing weather conditions, the step size spin needs to be performed in real time. At the same time, the variable step size inevitably has a strong impact on the interharmonic current [11], and a smaller MPPT perturbation step size reduces the level of interharmonic emission but results in poorer tracking performance of the MPPT algorithm.
Therefore, many artificial intelligence (AI) [12]-based methods are widely applied to generate VSS optimal solutions, such as neural network-based control [13], repetitive control [14], fuzzy logic control [15], model predictive control [16], and particle swarm optimization control [17]. The above methods are suitable for theoretical modeling or cases where the subset of parameters is unknown. However, there are certain drawbacks in the application process of these control strategies. For example, neural network control requires a large amount of datasets for offline training, repetitive control cannot adapt to the rapidly changing situation of photovoltaics due to its slow dynamic response, fuzzy logic control provides high-speed tracking and zero oscillation at the cost of computational complexity, model predictive control requires high accuracy for system parameters, and particle swarm optimization control easily falls into local extreme points when the P–V curve of multiple photovoltaic arrays has multiple local extreme points, resulting in incorrect results. Due to its inherent characteristics such as fast dynamic response and simple algorithmic principle, the Q-learning algorithm has been widely studied in recent years.
As a reinforcement learning algorithm [18], the Q-learning algorithm does not require trained models or previous knowledge during implementation. The core of the Q-learning algorithm is the control strategy, because it suffices to decide the control actions by itself. The general strategy is to set up a searchable Q-table [19], which contains the state space and action space. The state space contains all the operational states of the potential system, and the action space contains all potential control actions. In recent years, many people have studied the Q-table. In [20], the clock arrival time of each register is updated using the Q-table to maximize the designed clock arrival distribution to reduce the peak current. In [21], a learning-based dual Q power management method is proposed to extend the operating frequency to improve the embedded battery life and provide a sustainable operating energy. It overturns the traditional Dynamic Voltage and Frequency Scaling (DVFS) method. In [22], a dynamic weight coefficient based on Q-learning for Dynamic Window Approach (DQDWA) is proposed. The robot state, environmental conditions, and weight coefficients are used as the Q-table for learning to better adapt to different environmental conditions.
However, little research has been performed on how to set the Q-table, which is crucial to the tracking performance in MPPT. This paper proposes a reinforcement learning Q-table design scheme for MPPT control, aiming to maximize the tracking efficiency through improved Q-table update techniques. Six kinds of Q-tables based on the RL-MPPT method are established to find the optimal discrete value state of the photovoltaic system, make full use of the energy of the photovoltaic system, and reduce the power loss. The En50530 dynamic test procedure is used to simulate the real environment and evaluate the tracking performance fairly [23], while switching between static and dynamic conditions.
The remainder of this paper is structured as follows. Section 2 describes the basic principles of the RL method. In Section 3, the tracking performance of three discrete values is evaluated through testing. Section 4 summarizes the paper.

2. Method

2.1. Fundamentals of RL

Reinforcement learning is an approach to machine learning that finds optimal behavioral strategies to maximize cumulative rewards by allowing an intelligent body to continuously try and learn in an environment. The basic concepts of reinforcement learning include intelligences, environments, states, actions, rewards, strategies, and goals.
RL is a approach of learning that maps situations to actions to maximize return. Figure 1 and Figure 2 indicate that the RL methods are typically used in frameworks that include agents and environments. Instead of being told which actions to take, the agent is a learner which uses the interaction process to find which action yields the max reward. The things with which it interacts are called the environment. A Markov Decision Process (MDP) is when an agent and an environment talk to each other. The agent sees what is happening in the environment, and then performs some action to change it. If the agent and the environment have only a few choices, then there is always a best way for the agent to act.
Many RL methods have been proposed. Q-learning, as one of the popular RL methods, is used as the value function in this paper. The Q is defined by
Q t + 1 s t , a t = η r t + 1 + γ max a A Q t + 1 s t + 1 , a t Q t s t , a t + Q t s t , a t
where η 0 , 1 and γ 0 , 1 represent learning rate and discount factor, respectively. s, a, and r are state, action, and reward, respectively. Meanwhile, there exists a Q-table that records the Q-value of each state–action space to provide a basis for formulating the optimal policy, as shown in Figure 3.
The power supply circuit and its layout are shown in Figure 4. The RL-MPPT controller, the agent, perceives the state s S by the V p v and I p v ; the operating point on the I–V curve is then determined. Further, the variance in the duty cycle Δ D is output by the agent (i.e., the controller), which is the action a A . The control signal outputs a pulse with the corresponding D through the PWM, which acts as a switch to drive the boost converter. Finally, the reward r for the previous action shall be calculated.

2.1.1. Q-Table Setup

The Q-table approach in reinforcement learning is a value-based, model-free, off-track strategy reinforcement learning algorithm. Its purpose is to guide an intelligent body to choose the optimal action in each state by learning a Q-function. The Q-function represents the expectation of the long-term cumulative reward that can be obtained after executing action a in state. The Q-function is a function of the value of the action. Q-tables can be similar to a memorized database, which can store the experience and store the reward from each step of iteration between the agent and the environment. The main dimensions in the Q-table can be created by the state and action. The original Q-value is set to zero at the beginning of the Q-table. The reward can be calculated based on the state and action in each iteration. The Q-table is then updated in each iteration to store the latest rewards. Q-learning generally consists of two parts: the process of exploration and the process of utilization. During exploration, the agent randomly chooses an action to explore the state–action space and is rewarded accordingly. The Q-table is then structured and gradually becomes complete in this process of exploration. During exploitation, the agent chooses the action with the highest Q-value to execute the optimal program. In the Q-learning algorithm, four main items should be properly defined, i.e., state space, action space, reward function, and symbols.

2.1.2. State

A Q-table in reinforcement learning is a table for storing state–action correspondences of value functions, which can be used to implement value-based reinforcement learning algorithms such as Q-learning. A state space is the set of all possible states, e.g., in a maze game, each grid is a state. In reinforcement learning, the size and complexity of the state space affects the feasibility and efficiency of Q-tables. If the state space is discrete and finite, the Q-table can be represented by a two-dimensional array, where each row corresponds to a state, each column corresponds to an action, and each element stores the value of that state–action pair. In this case, the Q-table can be continuously updated to approximate the optimal value function. Only enough descriptive information is available to define the state space, rather than utilize the agent to make control decisions. However, the large amount of information may result in a complex state space, which inevitably increases the difficulty. In contrast, too little information may result in a weak ability to discriminate between types of states, which will result in a less effective decision-making ability. Throughout this paper, the location of the run point can be described by the measurable V p v and I p v . The state space is represented as
S = V p v , I p v
The voltage variable V p v is discrete from 0 to 25 V and the current variable I p v is discrete from 0 to 4 A. As shown in Figure 5, three discrete values are set in this paper; in this way the tracking performance of different state spaces is compared.

2.1.3. Action

The action space is the set of all possible actions that can be taken by an intelligent body in reinforcement learning. The size and complexity of the action space affects the design and effectiveness of the reinforcement learning algorithm. Action spaces can be categorized into two types: discrete and continuous. The advantage of a discrete action space is that it is easy to represent and compute. However, the disadvantage is that it may not be able to cover all the action choices or lead to an action space that is too large and difficult to explore. The advantage of continuous action spaces is that they allow for more precise control of an intelligent agent’s behavior. However, the disadvantage is that it is difficult to represent and optimize with tables or discrete functions, and it needs to be solved using methods such as function approximation or policy gradient. In this paper, for the control of MPPT the discrete action space should be realized using the control duty cycle.
This study specifies a discrete, finite action space for applying Q-learning to MPPT [24]. The action space needs to follow rules as (a) both positive and negative changes must be included in the action space; (b) the action needs to be given enough small resolution to attain optimum power; (c) in order to eliminate oscillations between states, a zero-charge action must be provided. For MPPT, the state is determined after measuring V p v and I p v . The movement of the running point can then be determined based on the actions chosen by the exploration strategy and the optimal strategy. Here, the action space is described in terms of five duty cycle Δ D steps (i.e., five actions) and is defined as
A = a | a 1 , a 2 , a 3 , a 4 , a 5
where the values of a 1 and a 5 indicate large positive and negative changes. Further, smaller positive and negative changes are indicated by a 2 and a 4 . Finally, no change in duty cycle is denoted by a 3 . To examine the tracking performance of several Q-tables, this study built up two separate action spaces, as shown in Table 1.

2.1.4. Reward

The effect of the interaction is measured by the reward function returning a scalar value [25]. The reward is a response from the “environment” as a result of action a for a state transition from s to s .
Usually, the previous action is evaluated by it. And the agent is taught how to choose the action. The reward function is expressed by
r t + 1 = w p Δ P t + 1 Δ t , i f Δ P t + 1 Δ t < δ w n Δ P t + 1 Δ t , i f Δ P t + 1 Δ t > δ 0 , i f Δ P t + 1 Δ t < δ
Δ P t + 1 = P t + 1 P t
where the current power is P t + 1 and the previous power is P t . Setting δ achieves the elimination of incorrect agent actions affected by measurement noise. The chosen action will result in a power rise indicated by a positive reward and vice versa. The reward weights can be set to w p = 1.5 , w n = 1 , thus reducing the convergence time.
The RL method flowchart is shown in Figure 6. Firstly, the discount factor γ , the maximum number of exploring E m a x , and the Q-table Q ( s , a ) are initialized. The voltage and current of PV system are measured to perceive the corresponding state in one iteration. Whether the current iteration is an exploration process or an exploitation process can be determined. During exploration, the next action will be randomly selected from space A. The action with the maximum Q-value will be selected in the exploitation process in the Q-table. According to (4), the reward r t + 1 of the previous action will be calculated and then stored in the Q-table.

2.2. The RL Algorithm

In this paper, Q-learning is applied to MPPT control. Through the interaction with the environment, the intelligent body can learn the optimal policy by feeding back the rewards, and then improve the performance of maximum power point tracking. In general, Q-learning is divided into two parts: the exploration process and exploitation process. In the exploration process, the intelligent body randomly selects an action to explore the environment and obtains the corresponding reward. The agent should explore the environment as much as possible and accumulate enough state–action space exploration experiences to accumulate experience. However, the algorithm cannot keep exploring; the algorithm needs to enter the exploitation phase to apply the previously accumulated experience and execute the optimal action through the previously acquired experience.
In an RL environment, the agent uses a trial-and-error process to learn the optimal policy as it interacts with the environment, rather than utilizing the model’s a priori knowledge [26]. The main meaning of the term “learning” is that the agent translates its experience that it gains from the interaction into knowledge. Depending on the objective optimal policy or optimal value function, RL algorithms can be categorized to two groups: the first are value-based approaches and the second are policy-based approaches. Q-learning is an approach that is value-based and model-free. For an MPPT algorithm, the state information that is measurable is accepted as an input signal, and the change in duty cycle is considered as an output signal. When initialization is complete, the current state s t is firstly observed by the algorithm. With the exploration count E n u m , the algorithm can determine whether the current iteration is an exploitation process or an exploration process. During exploration, this algorithm will choose the next action a randomly. However, convergence to the optimal Q-function is still ensured under the assumption that the state–action set is finite, discrete, and has an infinite number of visits. The algorithm will switch to the exploitation process when the exploration process is over (i.e., the E n u m is equal to the threshold E max ). In this process, the action with the largest Q-value in the Q-table will be chosen by the agent (i.e., it extracts the optimal policy). For both an exploration process or an exploitation process, the award that corresponds to the previous operation will be counted and then updated in the Q-table.
In order to achieve the optimal action selection strategy, traditional RL methods require a large number of exploratory iterations. However, it is time-consuming and infeasible to explore a huge state–action space. In addition, if the exploration process ends, traditional RL methods cannot gain experience beyond optimal decision-making through interaction with the environment. Once the solar irradiation changes rapidly, traditional RL methods must return to the exploration process and re-accumulate the exploration experience to update the optimal action strategy, which may lead to tracking failure.

2.3. Comparison Results with Other Methods

For a clear comparison with other methods. We have added results comparing the P&O method with a fixed step size, the P&O method with a variable step size, and the proposed Q-table method under the En50530 dynamic test program.
Although the Q-table RL method is compared with the fixed-step as well as variable-step MPPT methods. However, it can be seen from Figure 7, Figure 8 and Figure 9 that the RL method becomes more and more accurate after continuous exploration and iteration. In addition, when the PV module has problems such as corrosion and aging, the MPPT points will be shifted, and the learning iteration RL method can avoid this problem.

3. Results

The simulation model was implemented using Matlab/Simulink software. Table 2 presents the crucial data of the solar module. It should be noted that the is set to 0.1 s.
In order to validate the effectiveness of the proposed Q-table method, the experimental setup is physically shown in Figure 10 using hardware-in-the-loop (HIL). The DSP used is the TMS320F28335 model manufactured by Texas Instruments. The DSP can sample the voltage and current of the MT6016 through an analog-to-digital converter (ADC), and then generate PWM control signals through the controller area network (CAN). The hardware configuration is simulated by a transient simulation software developed by StarSim, Inc. (Shanghai, China). The HIL system provides PV voltage and current measurements and obtains PWM control signals through an interface.
The parameters of the experimental tests are the same as those of the simulations. The RL-MPPT algorithm is applied and implemented by Simulink Matlab 2018b functions through the Embedded Encoder Support Package for the Texas Instruments C2000 processor, without the need for any additional libraries. The parameters of the experimental test circuit are the same as in the simulation. The sampling time (frequency) for the MPPT controller is taken as 0.1 s.

3.1. En50530 Dynamic Test Condition

The test conditions closely simulate the operational environment of real photovoltaic systems by considering both and. Moreover, the test is particularly handy for indoor use. This paper focuses on the most difficult aspect of the test procedure [2]. As depicted in Figure 11, this specific part exhibits a gradient variation in irradiance ranging from 300 W/m 2 to 1000 W/m 2 at a rate of 100 W/m. The dynamic efficiency can be mathematically expressed as follows,
η d = i P p v × Δ T i j P max × Δ T j
where P max and Δ T j represent theoretical power and corresponding time, while P p v and Δ T i denote real power and corresponding time.

3.2. Simulation Result

The contrast of power, voltage, and duty cycle is illustrated among Figure 12 and Figure 13, with the red dashed lines representing theoretical values. Among the three types of state space, the second type exhibits the highest at 95.55%, surpassing that of the first type (92.32%) and the third type (93.44%).
It is worth noting that a smaller discretization value results in more convergence time, leading to power loss. Furthermore, once the Q-table has converged, employing a reasonable discretization value contributes to more precise tracing accuracy compared to using a larger one. Therefore, it is recommended to utilize a moderate discretization value for the long-term operation of the RL-MPPT method. Additionally, regardless of the chosen discretization value, the RL-MPPT method will generate a bigger voltage deviation during irradiance decrease; however, this issue can be mitigated by adopting a smaller discretization value. Conversely, utilizing a huge discretization value may bring about significant vibration in the stable state as it fails to precisely locate and maintain operation at the MPP.
The simulation results for type 1 of the action space with different types of state spaces are described in Figure 13. It is seen that the first and third state spaces can lead to significant trace failure for the first pattern. Notably, the second state space mitigates voltage drop during irradiance descent, resulting in an increase in harvested energy. In comparison to the first and second state spaces, it is found that the overall efficiency of the third state space is lower, achieving successful tracking of power maximum point at high irradiance steady-state only for the fourth pattern. Generally speaking, type 2 exhibits slower tracking compared to type 1, particularly during the rise of irradiance. Efficiency variations under different Q-tables are illustrated in Table 3.

3.3. Experimental Evaluation

Figure 10 shows the hardware in the loop (HIL) used in this experimental setup, and the TMS320F28335 model microcontroller produced by Texas Instruments is used.
The voltage and current from the MT6061 are converted by an analog-to-digital converter (ADC) and sampled by the microcontroller, and then a PWM control signal is generated by the controller area network (CAN). The simulation process is carried out through the Simulink Matlab 2018b.
This experiment was conducted on the MT6016 machine and obtained the same experimental test parameters as the simulation results. Figure 14 and Figure 15 show the power and voltage waveforms measured according to the En50530 test procedure. Where the green dashed box indicates the initial learning iteration phase of the Q-table. It also reflects the learning speed for different discretizations.The simulation results are roughly the same as the experimental results in the type 1 action space. The second type’s efficiency is 93.78%, which is higher than 90.77% of the first type and 90.78% of the third type.
It is noteworthy that the smaller the discrete value, the closer the working point is to the MPP, thus reducing power loss. However, when the discretization value is smaller rather than larger, it requires a longer convergence time. Therefore, achieving fast convergence and high tracking accuracy can be achieved by selecting an appropriate discretization value.
Figure 14 presents the experimental results for the type 2 action space with different kinds of state space. The efficiency of the second kind, at 88.55%, is the highest among the three kinds of state space, which is similar to the type 1 action space. Regardless of how the state space is sampled, type 2 leads to a lower efficiency compared to type 1. Thus, type 1 is better than type 2. Furthermore, due to sampling smaller step sizes (a1 and a5), the tracking speed is slow during the uphill part of the second pattern. But as the experience accumulates, the tracking speed improves greatly during the uphill part of the irradiance increasing process, and its efficiency increases. In addition, unlike in type 1, the voltage dip during the downhill slope is more severe in type 2. Table 4 shows the efficiency under different Q-tables.

4. Conclusions

In this paper, six kinds of Q-tables of the RL-MPPT method under the En50530 test procedure are discussed. By comparing the simulation and experimental results with the other two kinds of discretization value, it is shown that the highest efficiency can be achieved when the discretization value states are 1 V and 0.2 A. Additionally, the type 1 action space can lead to higher efficiency than the type 2 action space. Thus, this work recommends that the Q-table be set to a moderate resolution of state space and type 1 action space, which can lead to less power loss.

Author Contributions

Methodology, Y.C., D.L. and X.L.; Software, W.W.; Formal analysis, F.X.; Investigation, S.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by 2022 Jiangsu Carbon Peak and Neutrality Technology Innovation Special Fund (Industrial Foresight and Key Core Technology Research) “Research and Development of Key Technologies for Grid Integration Operation and Control of Renewable Energy Sources” (Project Number: BE2022003).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fernandes, D.A.; de Almeida, R.G.; Guedes, T.; Filho, A.J.S.; Costa, F.F. State feedback control for DC-photovoltaic systems. Electr. Power Syst. Res. 2017, 143, 794–801. [Google Scholar] [CrossRef]
  2. Li, X.; Wen, H.; Hu, Y.; Du, Y.; Yang, Y. A Comparative Study on Photovoltaic MPPT Algorithms Under EN50530 Dynamic Test Procedure. IEEE Trans. Power Electron. 2021, 36, 4153–4168. [Google Scholar] [CrossRef]
  3. Li, X.; Wang, Q.; Wen, H.; Xiao, W. Comprehensive Studies on Operational Principles for Maximum Power Point Tracking in Photovoltaic Systems. IEEE Access 2019, 7, 121407–121420. [Google Scholar]
  4. Kjaer, S.B. Evaluation of the “Hill Climbing” and the “Incremental Conductance” Maximum Power Point Trackers for Photovoltaic Power Systems. IEEE Trans. Energy Convers. 2012, 27, 922–929. [Google Scholar] [CrossRef]
  5. Femia, N.; Petrone, G.; Spagnuolo, G.; Vitelli, M. Optimization of perturb and observe maximum power point tracking method. IEEE Trans. Power Electron. 2005, 20, 963–973. [Google Scholar] [CrossRef]
  6. Bianconi, E.; Calvente, J.; Giral, R.; Mamarelis, E.; Petrone, G.; Ramos-Paja, C.A.; Spagnuolo, G.; Vitelli, M. A Fast Current-Based MPPT Technique Employing Sliding Mode Control. IEEE Trans. Ind. Electron. 2013, 60, 1168–1178. [Google Scholar] [CrossRef]
  7. Kumar, N.; Hussain, I.; Singh, B.; Panigrahi, B.K. Framework of Maximum Power Extraction From Solar PV Panel Using Self Predictive Perturb and Observe Algorithm. IEEE Trans. Sustain. Energy 2018, 9, 895–903. [Google Scholar] [CrossRef]
  8. Abdelsalam, A.; Massoud, A.; Ahmed, S.; Enjeti, P. High-Performance Adaptive Perturb and Observe MPPT Technique for Photovoltaic-Based Microgrids. IEEE Trans. Power Electron. 2011, 26, 1010–1021. [Google Scholar] [CrossRef]
  9. Xiao, W.; Dunford, W.G. A modified adaptive hill climbing MPPT method for photovoltaic power systems. In Proceedings of the 2004 IEEE 35th Annual Power Electronics Specialists Conference (IEEE Cat. No.04CH37551), Aachen, Germany, 20–25 June 2004; Volume 3, pp. 1957–1963. [Google Scholar]
  10. Liu, F.; Duan, S.; Liu, F.; Liu, B.; Kang, Y. A Variable Step Size INC MPPT Method for PV Systems. IEEE Trans. Ind. Electron. 2008, 55, 2622–2628. [Google Scholar]
  11. Hou, J.; Chen, H.; Ding, G.; Zhang, J.; Wu, C. Interharmonics suppression scheme in pv system with variable step size mppt algorithm. In Proceedings of the 2020 IEEE/IAS Industrial and Commercial Power System Asia (I&CPS Asia), Weihai, China, 13–16 July 2020; pp. 1369–1374. [Google Scholar]
  12. Zhao, S.; Blaabjerg, F.; Wang, H. An Overview of Artificial Intelligence Applications for Power Electronics. IEEE Trans. Power Electron. 2021, 36, 4633–4658. [Google Scholar] [CrossRef]
  13. Fu, X.; Li, S. Control of single-phase grid-connected converters with LCL filters using recurrent neural network and conventional control methods. IEEE Trans. Power Electron. 2015, 31, 5354–5364. [Google Scholar] [CrossRef]
  14. Chen, D.; Zhang, J.; Qian, Z. An improved repetitive control scheme for grid-connected inverter with frequency-adaptive capability. IEEE Trans. Ind. Electron. 2012, 60, 814–823. [Google Scholar] [CrossRef]
  15. Li, X.; Wen, H.; Hu, Y.; Jiang, L. A novel beta parameter based fuzzy-logic controller for photovoltaic MPPT application. Renew. Energ. 2019, 130, 416–427. [Google Scholar] [CrossRef]
  16. Lashab, A.; Sera, D.; Guerrero, J.M. A Dual-Discrete Model Predictive Control-Based MPPT for PV Systems. IEEE Trans. Power Electron. 2019, 34, 9686–9697. [Google Scholar] [CrossRef]
  17. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, WA, USA, 27 November–1 December1995; Volume 4, pp. 1942–1948. [Google Scholar]
  18. Lin, D.; Li, X.; Ding, S.; Wen, H.; Du, Y.; Xiao, W. Self-Tuning MPPT Scheme Based on Reinforcement Learning and Beta Parameter in Photovoltaic Power Systems. IEEE Trans. Power Electron. 2021, 36, 13826–13838. [Google Scholar] [CrossRef]
  19. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  20. Beheshti-Shirazi, S.A.; Nazari, N.; Gubbi, K.I.; Latibari, B.S.; Rafatirad, S.; Homayoun, H.; Sasan, A.; Manoj, S. Advanced Reinforcement Learning Solution for Clock Skew Engineering: Modified Q-Table Update Technique for Peak Current and IR Drop Minimization. IEEE Access 2023, 11, 87869–87886. [Google Scholar] [CrossRef]
  21. Huang, H.; Lin, M.; Yang, L.T.; Zhang, Q. Autonomous power management with double-Q reinforcement learning method. IEEE Trans. Ind. Inform. 2019, 16, 1938–1946. [Google Scholar] [CrossRef]
  22. Kobayashi, M.; Zushii, H.; Nakamura, T.; Motoi, N. Local Path Planning: Dynamic Window Approach with Q-learning Considering Congestion Environments for Mobile Robot. IEEE Access 2023, 11, 96733–96742. [Google Scholar] [CrossRef]
  23. Kantasewi, N.; Marukatat, S.; Thainimit, S.; Manabu, O. Multi Q-table Q-learning. In Proceedings of the 2019 10th International Conference of Information and Communication Technology for Embedded Systems (IC-ICTES), Bangkok, Thailand, 25–27 March2019; pp. 1–7. [Google Scholar]
  24. Bellman, R. A Markovian decision process. J. Math. Mech. 1957, 6, 679–684. [Google Scholar] [CrossRef]
  25. Brys, T.; Harutyunyan, A.; Vrancx, P.; Taylor, M.E.; Kudenko, D.; Nowé, A. Multi-objectivization of reinforcement learning problems by reward shaping. In Proceedings of the 2014 international joint conference on neural networks (IJCNN), Beijing, China, 6–11 July 2014; pp. 2315–2322. [Google Scholar]
  26. Watanabe, T.; Sawa, T. Instruction for reinforcement learning agent based on sub-rewards and forgetting. In Proceedings of the International Conference on Fuzzy Systems, Barcelona, Spain, 18–23 July 2010; pp. 1–7. [Google Scholar]
Figure 1. The agent–environment interface in an MDP.
Figure 1. The agent–environment interface in an MDP.
Energies 16 07286 g001
Figure 2. Schematic diagram of the interaction process between agent and environment.
Figure 2. Schematic diagram of the interaction process between agent and environment.
Energies 16 07286 g002
Figure 3. Schematic diagram of the Q-table.
Figure 3. Schematic diagram of the Q-table.
Energies 16 07286 g003
Figure 4. RL-MPPT method simulation.
Figure 4. RL-MPPT method simulation.
Energies 16 07286 g004
Figure 5. PV source I–V curves in three resolution state spaces: (a) 1 V and 0.1 A for discrete value states; (b) 1 V and 0.2 A for discrete value states; (c) 1.5 V and 0.1 A for discrete value states.
Figure 5. PV source I–V curves in three resolution state spaces: (a) 1 V and 0.1 A for discrete value states; (b) 1 V and 0.2 A for discrete value states; (c) 1.5 V and 0.1 A for discrete value states.
Energies 16 07286 g005
Figure 6. Q-learning flowchart for the conventional RL method.
Figure 6. Q-learning flowchart for the conventional RL method.
Energies 16 07286 g006
Figure 7. Simulation results of MPPT algorithm based on fixed step size.
Figure 7. Simulation results of MPPT algorithm based on fixed step size.
Energies 16 07286 g007
Figure 8. Simulation results of MPPT algorithm based on variable step size.
Figure 8. Simulation results of MPPT algorithm based on variable step size.
Energies 16 07286 g008
Figure 9. Simulation results of MPPT algorithm based on proposed Q-table method.
Figure 9. Simulation results of MPPT algorithm based on proposed Q-table method.
Energies 16 07286 g009
Figure 10. Setup of HIL system.
Figure 10. Setup of HIL system.
Energies 16 07286 g010
Figure 11. En50530 dynamic test condition diagram.
Figure 11. En50530 dynamic test condition diagram.
Energies 16 07286 g011
Figure 12. Simulation results for different discrete values under action space type 1: (a) 1 V and 0.1 A discrete state; (b) 1 V and 0.2 A discrete state; and (c) 1.5 V and 0.1 A discrete state.
Figure 12. Simulation results for different discrete values under action space type 1: (a) 1 V and 0.1 A discrete state; (b) 1 V and 0.2 A discrete state; and (c) 1.5 V and 0.1 A discrete state.
Energies 16 07286 g012
Figure 13. Simulation results for different discrete-valued states under action space type 2: (a) 1 V and 0.1 A state; (b) 1 V and 0.2 A state; (c) 1.5 V and 0.1 A state.
Figure 13. Simulation results for different discrete-valued states under action space type 2: (a) 1 V and 0.1 A state; (b) 1 V and 0.2 A state; (c) 1.5 V and 0.1 A state.
Energies 16 07286 g013
Figure 14. Experimental waveform results under the type 2 action space: (a) the discretization values are 1 V and 0.1 A; (b) the discretization values are 1 V and 0.2 A; (c) the discretization values are 1.5 V and 0.1 A.
Figure 14. Experimental waveform results under the type 2 action space: (a) the discretization values are 1 V and 0.1 A; (b) the discretization values are 1 V and 0.2 A; (c) the discretization values are 1.5 V and 0.1 A.
Energies 16 07286 g014
Figure 15. Experimental waveform results for different discrete value states under action space type 1: (a) 1 V and 0.1 A discrete state; (b) 1 V and 0.2 A discrete state; (c) 1.5 V and 0.1 A discrete state.
Figure 15. Experimental waveform results for different discrete value states under action space type 1: (a) 1 V and 0.1 A discrete state; (b) 1 V and 0.2 A discrete state; (c) 1.5 V and 0.1 A discrete state.
Energies 16 07286 g015
Table 1. Two different type action spaces.
Table 1. Two different type action spaces.
a 1 a 2 a 3 a 4 a 5
Type 1−0.02−0.00500.0050.02
Type 2−0.01−0.00500.0050.01
Table 2. Main product parameters of the PV module MSX-60W.
Table 2. Main product parameters of the PV module MSX-60W.
ParameterSymbolValue
Maximum power P m p p 60 W
Current at MPP I m p p 3.5 A
Voltage at MPP V m p p 17.1 V
Short-circuit current I s c 3.8 A
Open-circuit voltage V o c 21.1 V
Temperature coefficient of I s c K i 0.065%/ ° C
Temperature coefficient of V o c K v −80 mV/ ° C
Table 3. Comparisonof simulation efficiency results of different Q-tables.
Table 3. Comparisonof simulation efficiency results of different Q-tables.
1 V and 0.1 A1 V and 0.2 A1.5 V and 0.1 A
Action Type 192.32%95.55%93.44%
Action Type 290.21%91.90%90.09%
Table 4. Comparisonof experiment efficiency results of different Q-tables.
Table 4. Comparisonof experiment efficiency results of different Q-tables.
1 V and 0.1 A1 V and 0.2 A1.5 V and 0.1 A
Action Type 190.77%93.78%90.78%
Action Type 287.35%88.55%85.98%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Y.; Lin, D.; Xu, F.; Li, X.; Wang, W.; Ding, S. Research on Q-Table Design for Maximum Power Point Tracking-Based Reinforcement Learning in PV Systems. Energies 2023, 16, 7286. https://doi.org/10.3390/en16217286

AMA Style

Chen Y, Lin D, Xu F, Li X, Wang W, Ding S. Research on Q-Table Design for Maximum Power Point Tracking-Based Reinforcement Learning in PV Systems. Energies. 2023; 16(21):7286. https://doi.org/10.3390/en16217286

Chicago/Turabian Style

Chen, Yizhi, Dingyi Lin, Fei Xu, Xingshuo Li, Wei Wang, and Shuye Ding. 2023. "Research on Q-Table Design for Maximum Power Point Tracking-Based Reinforcement Learning in PV Systems" Energies 16, no. 21: 7286. https://doi.org/10.3390/en16217286

APA Style

Chen, Y., Lin, D., Xu, F., Li, X., Wang, W., & Ding, S. (2023). Research on Q-Table Design for Maximum Power Point Tracking-Based Reinforcement Learning in PV Systems. Energies, 16(21), 7286. https://doi.org/10.3390/en16217286

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop