Next Article in Journal
Coupling Control Strategy and Experiments for Motion Mode Switching of a Novel Electric Chassis
Next Article in Special Issue
Emulation Strategies and Economic Dispatch for Inverter-Based Renewable Generation under VSG Control Participating in Multiple Temporal Frequency Control
Previous Article in Journal
Efficient Stereo Visual Simultaneous Localization and Mapping for an Autonomous Unmanned Forklift in an Unstructured Warehouse
Previous Article in Special Issue
Phase Balancing and Reactive Power Support Services for Microgrids
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Global MPPT Based on Machine-Learning for PV Arrays Operating under Partial Shading Conditions

by
Christos Kalogerakis
,
Eftichis Koutroulis
* and
Michail G. Lagoudakis
School of Electrical & Computer Engineering, Technical University of Crete, GR 73100 Chania, Greece
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(2), 700; https://doi.org/10.3390/app10020700
Submission received: 24 December 2019 / Revised: 12 January 2020 / Accepted: 14 January 2020 / Published: 19 January 2020
(This article belongs to the Special Issue Advancing Grid-Connected Renewable Generation Systems 2019)

Abstract

:
A global maximum power point tracking (GMPPT) process must be applied for detecting the position of the GMPP operating point in the minimum possible search time in order to maximize the energy production of a photovoltaic (PV) system when its PV array operates under partial shading conditions. This paper presents a novel GMPPT method which is based on the application of a machine-learning algorithm. Compared to the existing GMPPT techniques, the proposed method has the advantage that it does not require knowledge of the operational characteristics of the PV modules comprising the PV system, or the PV array structure. Additionally, due to its inherent learning capability, it is capable of detecting the GMPP in significantly fewer search steps and, therefore, it is suitable for employment in PV applications, where the shading pattern may change quickly (e.g., wearable PV systems, building-integrated PV systems etc.). The numerical results presented in the paper demonstrate that the time required for detecting the global MPP, when unknown partial shading patterns are applied, is reduced by 80.5%–98.3% by executing the proposed Q-learning-based GMPPT algorithm, compared to the convergence time required by a GMPPT process based on the particle swarm optimization (PSO) algorithm.

1. Introduction

The economic competitiveness of photovoltaic (PV) systems, compared to conventional power production technologies, has continuously improved during the last years. Therefore, they exhibit significant potential for deployment in both stand-alone and grid-connected applications, also in combination with electrical energy storage units [1].
The PV modules comprising the PV array of a PV energy production system may operate under non-uniform incident solar irradiation conditions due to partial shading caused by dust, nearby buildings, trees, etc. As shown in Figure 1, under these conditions the power–voltage curve of the PV array exhibits multiple local maximum power points (MPPs). However, only one of these MPPs corresponds to the global MPP (GMPP), where the PV array produces the maximum total power [2]. Therefore, the controller of the power converter that is connected at the output of the PV array must execute an effective global MPP tracking (GMPPT) process in order to continuously operate the PV array at the GMPP during the continuously changing incident solar irradiation conditions. That process results in the continuous maximization of the energy production of the installed PV system and enables the reduction of the corresponding levelized cost of electricity (LCOE) [1].
The simplest algorithm for detecting the position of GMPP is to scan the entire power-voltage curve of the PV source by iteratively modifying the duty cycle of the PV power converter [3], which results in the continuous modification of the PV source output voltage. Alternatively, the current–voltage characteristic of the PV array is scanned by iteratively increasing the value of the reference current in a direct current/direct current (DC/DC) PV power converter with average current-mode control [4]. Compared to the power-voltage scan process, this method has the advantage of faster detection of the GMPP, since it avoids searching the parts of the current-voltage curve located on the left-hand side of the local MPPs. In both of these methods, the accuracy and speed of detecting the GMPP under complex shading patterns depend on the magnitude of the perturbation search-step of the duty cycle or reference current, respectively. In [5], the current–voltage curve of the PV source is scanned by setting in the ON and OFF states the power switch of a boost-type DC/DC converter, thus causing the PV source current to sweep from zero to the short-circuit value. The PV source current and voltage are sampled at high speed during this process in order to calculate the PV array output voltage where the PV power production is maximized (i.e., GMPP). The effective implementation of this technique requires the use of high-bandwidth current and voltage sensors, a high-speed analog-to-digital converter and a fast digital processing unit [e.g., a field-programmable gate array (FPGA)], thus resulting in high manufacturing cost of the GMPPT controller. Also, the application of this method requires the use of an inductor in series with the PV source, thus it cannot be applied in non-boost type power converter topologies.
The GMPPT algorithms proposed in [6,7] are based on searching for the global MPP close to integer multiples of 0.8 × Voc, where Voc is the open-circuit voltage of the PV modules. Thus, this technique requires knowledge of both the value of Voc and the number of series-connected PV modules (or sub-modules with bypass diodes) within a PV string. A GMPPT controller based on machine learning has been proposed in [8], using Bayesian fusion to avoid convergence to local MPPs under partial shading conditions. For that purpose, a conditional probability table was trained using values of the PV array output voltage at integer multiples of 80% of the PV modules open-circuit voltage under various shading patterns. In [9], the GMPP search process was based on the value of an intermediate parameter (termed as “beta”), which is calculated according to the PV modules open-circuit voltage and the number of PV modules connected in series in the PV string. All of the techniques described above cannot be applied to PV arrays with unknown operational characteristics. Therefore, they cannot be incorporated in commercial PV power converter products used in PV applications, where the operational characteristics of the PV array are determined by the end-user without having been known during the power converter manufacturing stage.
In [10] a distributed maximum power point tracking (DMPPT) method is proposed, where the power converters of each PV module communicate with each other in order to detect deviations among the power levels generated by the individual PV modules of the PV array. That process enables the identification of the occurrence of partial shading conditions within the submodules that they comprise. The global MPP is then detected by searching with a conventional maximum power point tracking (MPPT) method (e.g., incremental-conductance) at particular voltage ranges derived similarly to the 0.8 × Voc GMPPT technique. Therefore, the implementation of this technique requires knowledge of the PV source arrangement/configuration and its operational characteristics.
Evolutionary algorithms have also been applied, considering the GMPPT process as an optimization problem where the global optimum (i.e., GMPP) must be discovered so that the objective function corresponding to the output power of the PV array is maximized. The advantage of these types of algorithm is that they are capable of searching for the position of the GMPP on the power-voltage curve of the PV source with a lower number of search steps compared to the exhaustive scanning/sweeping process. These algorithms have been inspired by natural and biological processes and they differ on the operating principle used to produce the alternative operating points of the PV array, which are investigated during searching for the GMPP position. The particle swarm optimization (PSO) [11,12,13], grey wolf optimization (GWO) [14,15,16], flower pollination algorithm (FPA) [17,18], Jaya algorithm [19] and differential evolution (DE) algorithm [20] are the most frequently employed evolutionary algorithms. Genetic algorithms (GAs) inspired by the process of natural evolution are applied in [21]. However, this type of optimization algorithm exhibits high computational complexity compared to other evolutionary algorithms, due to the complex operations (i.e., selection, mutation and crossover) which must be performed during the chromosomes search process.
During operation, the evolutionary algorithms avoid convergence to local MPPs and converge to an operating point that is located close to the GMPP. Thus, they should be combined with a traditional MPPT method [e.g., perturbation and observation (P&O), incremental-conductance (INC) etc.], which is executed after the evolution algorithm execution has been finished, in order to: (i) fine-tune the PV source operating point to the GMPP and (ii) maintain operation at the GMPP during short-term changes of solar irradiation or shading-pattern, which do not alter significantly the shape (e.g., relative position of the local and global MPPs) of the power–voltage curve (e.g., [15]). This enables to avoid frequent re-executions of the evolutionary algorithm, which would result in power loss due to operation away from the GMPP during the search process. The execution of the evolutionary algorithm is either re-initiated periodically (e.g., every few minutes), or after the detection of significant changes of the PV power production, in order to track any new position of the GMPP. The performance of many metaheuristic optimization techniques implementing the GMPPT process in partially-shaded PV systems has been compared in [22].
This paper presents a new GMPPT method, which is based on a machine-learning algorithm. Compared to the past-proposed GMPPT techniques, the Q-learning-based method proposed in this paper has the following advantages: (a) it does not require knowledge of the operational characteristics of the PV modules and the PV array comprised in the PV system; and (b) due to its inherent learning capability, it is capable of detecting the GMPP in significantly less search steps. Thus, the proposed GMPPT method is suitable for employment in PV applications, where the shading pattern may change quickly (e.g., wearable PV systems, building-integrated PV systems, where shading is caused by people moving in front of the PV array, etc.). The numerical results presented in this paper, validate the capabilities and advantages of the proposed GMPPT technique.
This paper is organized as follows: the operating principles and structure of the proposed Q-learning-based GMPPT method are analyzed in Section 2; the numerical results by applying the proposed method, as well as the PSO evolutionary GMPPT algorithm, for various shading patterns of a PV array are presented in Section 3; and, finally, the conclusions are discussed in Section 4.

2. The Proposed Q-Learning-Based Method for Photovoltaic (PV) Global Maximum Power Point Tracking (GMPPT)

A block diagram of the PV system under consideration is illustrated in Figure 2. The PV array comprises multiple PV modules connected in series and parallel. The GMPPT process is executed by the GMPPT controller, which produces the duty cycle, D, of the pulse width modulation (PWM) control signal of the DC/DC converter using measurements of the PV array output voltage and current. Each value of D determines the PV array output voltage according to (1):
V P V = ( 1 D ) V o
where V o (V) is the DC/DC converter output voltage and V P V (V) is the PV array output voltage.
The proposed technique considers the PV GMPPT operation as a Markov decision process (MDP) [23], which constitutes a discrete-time stochastic control process and models the interaction between an agent implemented in the GMPPT control unit and the PV system (i.e., PV array and DC/DC power converter). The MDP consists of: (a) the state-space S, (b) the set of all possible actions A, and (c) the reinforcement reward function, which represents the reward, when applying action a in state S , which leads to a transition to state S [23]. The Markovian property dictates that each state transition is independent of the history of previous states and actions and depends only on the current state and action; likewise, each reward is independent of the past states, actions, and rewards and depends only on the most recent transition. A temporal-difference Q-learning algorithm is applied for solving the PV GMPPT optimization problem. The Q-learning algorithm’s goal is to derive an action selection policy, which will maximize the total expected discounted rewards that the system will receive in the future. A simplified representation of process implemented by the Q-learning algorithm in order to control a PV system for the implementation of the GMPPT process is presented in Figure 3.
In Q-learning, an agent interacts with the unknown environment (i.e., the PV system) and gains experience through a specific set of states, actions and rewards encountered during this interaction [24,25,26,27]. Q-learning strives to learn the Q-values of state-actions pairs, which represent the expected total discounted reward in the long term. Typically, experience for learning is recorded in terms of samples (St, at, Rt, St+1), meaning that at some time step t, action at was executed in state St and a transition to the next state St+1 was observed, while reward Rt was received. The Q-learning update rule, given a sample (St, at, Rt, St+1) at time step t , is defined as follows:
Q ( S t , a t ) = Q ( S t a t ) + a l p h a [ R t + γ max a t + 1 { Q ( S t + 1 , a t + 1 ) } Q ( S t , a t ) ]
where Q ( S t , a t ) is the Q-value of the current state-action pair that will be updated, Q ( S t + 1 , a t + 1 ) is the Q-value of the next state-action pair, a l p h a is the learning rate, R t is the immediate reward and γ is the discount factor.
In each time-step, the agent observes its current state and chooses its action according to its action-selection policy π . After the action selection, it observes its future state S t + 1 . Then, it receives an immediate reward R t and selects the future Q-value that offers the maximum value over all possible actions, i.e., max a t + 1 Q ( S t + 1 , a t + 1 ) . The learning rate determines how much the new knowledge acquired by the agent will affect the existing estimate in the update of Q ( S t , a t ) .

2.1. Action Selection Policy

Each possible pair of state-action must be evaluated for each state that the agent has visited in order to ensure that the estimated/learned Q-value function is the optimal one (i.e., corresponds to the most suitable action policy). In this work, the Boltzmann exploration policy has been used [28]. Each possible action has a probability of selection by the agent, which is calculated as follows:
p ( S t , a i ) = e Q ( S t , a i ) τ 1 | A | e Q ( S t , a i ) τ
where τ is the Boltzmann exploration “temperature”, α i is the i-th possible action, | A | (size of the action-space) is the total number of alternative actions in each state and Q ( S t , a i ) is the Q-value for the i-th action in state S t .
The temperature function is given by:
τ = { τ min + ( 1 N N max ) ( τ max τ min ) if   N N max τ min if   N   >   N max
where τ min , τ max are the minimum and maximum, respectively, “temperature” values, N is the current number of visits of the agent in the specific state and N max is the maximum number of visits of the agent.
The value of τ is positive and controls the randomness of the action selection. If the value of τ is high, then all values of the probabilities of each action, which are calculated using (3), are similar. This results in a random action selection depending on the parameter N r , which is a random number in the range (0, 1). Section 2.6 presents the way that N r affects the action-selection policy. As the number of visits, N , to a specific state increases, the value of τ reduces. After a certain number of visits, N max , the value of τ becomes equal to its minimum value. This means that the exploration process has been accomplished in that state and the agent chooses the action with the highest Q-value in that state.

2.2. State-Space

During the execution of the PV GMPPT process, each state depends on the current value of the duty cycle of the DC/DC power converter (Figure 2), the power generated by the PV array and the duty cycle during the previous time-step [28]:
S t = { S | S i , j , k = ( D i , P P V   j , D O k ) ,   i [ 1 , 2 , 3 , . . . . , n ] ,   j [ 1 , 2 , 3 , . . . . , m ] ,   k [ 1 , 2 , 3 , . . . . , p ] }
where n is the number of equally quantized steps of the duty cycle ( D ) range, m is the number of equally quantized steps of the PV array output power ( P P V ) range and k is the number of equally quantized steps of the duty cycle range for the previous time-step ( D O ) . Selecting high values for n , m and k will result in a higher accuracy of detecting the MPP, but the learning time and storage requirements are increased. For each value of D , the state is determined jointly by the current and the previous quantization steps that the duty cycle value belongs to, as well as by the quantization step that the PV array output power value belongs to. For example, assume that n = 10 , m = 10 , k = 5 , p max = 100 W , the duty cycle value at the current time-step is equal to 0.64, the PV array output power is equal to 53 W and the duty cycle value at the previous time-step equals 0.5. In this case, the agent’s state is equal to (7, 6, 3), because 53 W belongs to the 6th quantization step of the PV power range, 0.64 belongs to the 7th quantization step of the current duty cycle range and 0.5 belongs to the 3rd quantization step of the previous-duty-cycle range.

2.3. Action-Space

The action-space of the PV system GMPPT controller is expressed by the following equation:
D t + 1 = D t + Δ D
where D t + 1 is the duty cycle for the next time-step, D t is the current duty cycle value and Δ D indicates the “increment”, “decrement” or “no-change” of the current duty cycle value.
If the agent selects action “no change”, then it will remain in the same state during the next time-step. The actions “increment” and “decrement” of the current duty cycle value are further classified as “low”, “medium” and “high”, as analyzed in Section 3, in order to ensure that most of the duty cycle values are selected. Thus, there is a total of seven actions in the action-space with the “no change” action being the last one. This is necessary in order to minimize the probability that the agent will converge to a local MPP. Each action is selected according to the Boltzmann exploration policy, which was analyzed in Section 2.1. When the exploration process has been finished, the action with the highest Q-value is selected for that particular state. It is considered that the GMPP of the PV array is detected when the action “no change” has the highest Q-value. In this case, the Q-values of all other actions are lower due to the reduction of the PV array output power by the selection of the corresponding actions compared to the Q-value for action “no change”.

2.4. Reward

The criterion for the selection of the most suitable action is the immediate reward, R t , which evaluates the action selection according to:
R t = { Positive   Reward   if   P P V ( t + 1 ) P P V ( t ) > + d 0 if   d P P V ( t + 1 ) P P V ( t ) + d Negative   Reward if   P P V ( t + 1 ) P P V ( t ) < d
where P P V ( t ) ( W ) is the PV array output power at time step t and d is the desired power difference considered necessary in order for the agent to realize that a significant power change has occurred. In case the difference P P V ( t + 1 ) P P V ( t ) is higher than + d , then a small positive reward will be given for that specific pair of state-action, in order to “encourage” the agent to continue this map-selection strategy during the next time-step that it will visit the current state. If P P V ( t + 1 ) P P V ( t ) is within [ d , + d ] it is considered that there is no change in the generated power, and the reward will be equal to zero. Lastly, if P P V ( t + 1 ) P P V ( t ) is lower than d , then a small negative reward will be given to the agent such that in the next time-step that the agent will visit the current state again, the action that caused the power reduction will have a small probability of re-selection according to (3). This will result in the selection of actions which promote the increment of the PV array power production during the GMPP exploration process.

2.5. Discount Factor and Learning Rate

The discount factor is a value which is used to balance immediate and future rewards. The agent is trying to maximize the total future rewards, instead of the immediate reward. It is used to adjust the weight of the reward of the current optimal value and indicate the significance of future rewards [24]. The discount factor is necessary in order to help the agent find optimal sequences of future actions that lead to high rewards soon and not only the current optimal action.
A visit frequency criterion is applied for ensuring that the proposed Q-learning-based GMPPT process will converge to the GMPP. The agent is encouraged to consider that knowledge acquired from states with a small number of visits is more important than knowledge acquired from states with a higher number of visits. Thus, states that have been visited fewer times will have a higher learning rate than states that have been explored more times. This is accomplished by calculating the value of the learning rate a l p h a in (2) as follows:
a l p h a = k 1 k 2 + k 3 N
where k 1 , k 2 and k 3 are factors which determine the initial learning rate value for a state with no visits and N is the number of times that the agent visited the specific state. The rationale behind the variable value of alpha is that its value decreases as the number of visits increases to enable convergence in the limit of many visits.

2.6. The Overall Q-Learning-Based GMPPT Algorithm

A flowchart of the overall Q-learning-based PV GMPPT algorithm, which is proposed in this paper, is presented in Figure 4. The PV GMPPT process is performed as follows:
  • Step 1: Q-table initialization. As analyzed in Section 2.2, a Q-table is created using four dimensions according to (5); three dimensions for the state and one for the action. This table is initially filled with zeros, because there is no prior knowledge. Also, a table that contains the number of visits of each state is defined. The size of this table is equal to the number of states. Lastly, an initial value is given to the duty cycle.
  • Step 2: After the PV array output voltage and current signals reach the steady-state, the PV array-generated power is calculated. According to the value of the duty cycle in the current time-step, its value during the previous time-step and the PV array-generated power, the state is determined using (5).
  • Step 3: The learning rate is calculated by (8).
  • Step 4: The temperature τ is calculated by (4) and the probability of every possible action for the current state is also calculated by (3).
  • Step 5: If the number of visits to the current state is higher than or equal to the predefined value of N max and the Q-value for the action corresponding to “no-change” of the duty cycle value is the maximum compared to the other actions, then it is concluded that convergence to an operating point close to the GMPP has been achieved and operation at the local MPPs has been avoided. Then, the proposed GMPPT process executes the P&O MPPT algorithm in order to help the agent fine-tune its operation at the GMPP.
  • Step 6: If the number of visits of the current state is lower than N max or if the Q-value for the action corresponding to “no-change” of the duty cycle value is not the maximum, a number in the range (0, 1) is randomly selected and compared to the probabilities of every possible action for the specific state (step 4). Each probability encodes a specific duty cycle change, as follows:
    If   0 < N r p 1 :   Action   1 Else   if   p 1 < N r p 1 + p 2 :   Action   2 Else   if   p 1 + p 2 < N r k = 1 3 p k :   Action   3 Else   if   k = 1 3 p k < N r k = 1 4 p k :   Action   4 Else   if   k = 1 4 p k < N r k = 1 5 p k :   Action   5 Else   if   k = 1 5 p k < N r k = 1 6 p k :   Action   6 Else :   Action   7  
    where p k (k = 1, 2, …6) is the probability of each possible action for the current state that is calculated by using (3), while Ν r is a random number in the range ( 0 ,   1 ) .
  • Step 7: After the change of the action, a new duty cycle value is applied to the control signal of the DC/DC power converter and the algorithm waits until the PV array output voltage and the current signals reach the steady state. Then, the proposed GMPPT algorithm calculates the difference between the power produced by the PV array at the previous and the current time-steps, respectively, and assigns the reward according to (7).
  • Step 8: The future state is determined and the Q-function is updated according to (2).
  • Step 9: Return to step 2.

3. Numerical Results

In order to evaluate the performance of the proposed Q-learning-based PV GMPPT method, a model of the PV system has been developed in the MATLABTM/Simulink software platform (Figure 5). The PV system under study consists of the following components: (a) a PV array connected according to the series-parallel topology, (b) a DC/DC boost-type power converter [29], (c) a GMPPT control unit which produces the duty cycle of the power converter PWM control signal, and (d) a battery that is connected to the power converter output terminals.
Table 1 presents the operational characteristics οf the PV modules comprising the PV array. Table 2 displays the values of the operational parameters of the proposed Q-learning-based GMPPT method, while Table 3 presents the action-space that was used for the Q-learning-based method. For each state, the action of ΔD “increment” or “decrement” with the highest Q-value is performed, in order to ensure the detection of the GMPP with the minimum number of time-steps. There are three categories of duty cycle change (i.e., “low”, “medium” and “high”), which may be selected depending on how close the agent is to the GMPP. The action of “no change” of ΔD is used such that the agent is able to realize that the position of GMPP has now been detected. The operation of the PV system with the proposed Q-learning-based GMPPT method was simulated for 9 different shading patterns of the PV array. Regarding the state-space, the duty cycle was quantized in 18 equal steps, the PV array output power was quantized in 30 equal steps and the duty cycle value of the previous time-step was quantized in 9 equal steps. For comparison with the proposed method, the PSO-based GMPPT method [30] was also implemented and simulated as analyzed next. Table 4 displays the values of the operational parameters of the PSO-based GMPPT method. The values of the operational parameters of the proposed and PSO-based GMPPT algorithms (i.e., Table 2, Table 3 and Table 4) were selected, such that they converge as close as possible to the global MPP and with the minimum number of search steps for the partial shading patterns under consideration, which are presented in the following section.

3.1. Shading Patterns Analysis

In order to evaluate the performance of the proposed Q-learning-based GMPPT method, the shading patterns 1–3 were used during the learning process and then the acquired knowledge was exploited during the execution of the GMPPT process for the rest of the shading patterns. The distribution of incident solar irradiation (in W / m 2 ) over each PV module of the PV array for each shading pattern is presented in Figure 6, while the resulting output power-voltage curves of the PV array are presented in Figure 7. Each number (i.e., SP1, SP2 etc.) in Figure 7 indicates the order of application of each shading pattern during the test process. These shading patterns, as well as their sequence of application, were formed such that the power-voltage characteristic of the PV array exhibits multiple local MPPs at varying power levels (depending on the shading pattern). This approach enabled the learning capability of the agent employed in the proposed Q-learning-based GMPPT technique and its effectiveness in reducing the time required for convergence to the GMPP to be investigated.

3.2. Tracking Performance of the Proposed Q-Learning-Based GMPPT Method

Figure 8 presents the variations of the duty cycle and PV array output power during the execution of the proposed Q-learning-based GMPPT process for shading pattern 1. It is observed that, in this case, the Q-learning-based GMPPT algorithm needed a relatively long period of time in order to converge close to the GMPP. This is due to the fact that initially the agent has no prior knowledge of where the GMPP may be located at. Thus, it has to visit many states, until it converges to the duty cycle value, which corresponds to an operating point close to the GMPP. When the oscillations start (i.e., time = 29 s), the exploration by the agent of the Q-learning-based GMPPT algorithm stops and the P&O MPPT algorithm starts its execution.
Figure 9 and Figure 10, respectively, illustrate the duty cycle and PV array output power versus time during the execution of the proposed Q-learning-based GMPPT process for shading patterns 2 and 3. The knowledge gained by the agent during the execution of the GMPPT process for shading pattern 1 did not affect the speed of the GMPP location-detection process for shading pattern 2, since the agent needs to be trained in more shading patterns. Similarly, the knowledge gained during the GMPPT process for shading patterns 1 and 2 did not affect the GMPPT process for locating the GMPP for shading pattern 3. The variations of duty cycle and PV array output power versus time for shading pattern 4, which corresponds to an intermediate power-voltage curve with respect to those of shading patterns 1–3 (Figure 7), are displayed in Figure 11. Since the GMPPT process for shading patterns 1–3 had been executed previously, the agent was now able to detect the GMPP in a shorter period of time by exploiting the knowledge acquired before. As shown in Figure 12, when shading pattern 5 was applied, which was unknown till that time, the time required for detecting the position of the GMPP was further reduced.
Figure 13 and Figure 14 present the results of the simulations for shading patterns 6 and 7. After this point, the agent of the Q-learning-based GMPPT process is able to detect the GMPP in unknown intermediate power–voltage curve conditions with respect to those of shading patterns 1 and 2 (Figure 8 and Figure 9) in much less time compared to the previous shading patterns. This happens because the agent had previously acquired enough knowledge from the learning process performed during the GMPPT execution for shading patterns 1–5.
Finally, the simulation results for the unknown shading patterns 8 and 9 in Figure 15 and Figure 16, respectively, demonstrate that the agent is capable to detect the GMPP in significantly less time. It is therefore concluded that the inherent learning process integrated in the proposed Q-learning algorithm enhances significantly the convergence speed of the proposed GMPPT method.

3.3. Tracking Performance of the Particle Swarm Optimization (PSO)-Based GMPPT Method

The PSO GMPPT process was also applied for shading patterns 1–9. As an indicative example, Figure 17 and Figure 18 illustrate the duty cycle and PV array output power variations during the execution of the PSO-based GMPPT algorithm for shading patterns 1 and 9, respectively. It can be observed that the variability of the duty cycle and the PV array output power are not affected significantly by the shape of the shading pattern applied. In contrast, as illustrated in Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13, Figure 14, Figure 15 and Figure 16, the learning capability of the agent incorporated in the proposed Q-learning-based GMPPT approach enabled the progressive reduction of the duty cycle and the PV array output power variability, when shading patterns 1–9 were applied. The PSO-based GMPPT method required a constant period of time (approximately 11.5 s) in order to detect the location of the GMPP. A similar behavior of the PSO-based algorithm was also observed for shading patterns 2–8.

3.4. Comparison of the Q-Learning-Based and PSO-Based GMPPT Methods

The time required for convergence, the number of search steps performed during the GMPPT process until the P&O process starts its execution and the MPPT efficiency were calculated for each shading pattern for both the Q-learning-based and PSO-based GMPPT algorithms. The execution of the P&O MPPT process was not included in this analysis. The corresponding results are presented in Table 5 and Table 6, respectively. The MPPT efficiency is defined as follows:
n = P P V P G M P P
where P P V ( W ) is the PV array output power after convergence of the Q-learning-based, or the PSO-based GMPPT method, respectively and P G M P P ( W ) is the GMPP power of the PV array for the solar irradiance conditions (i.e., shading pattern) under study.
The results presented in Table 5 indicate that the knowledge acquired initially by the Q-learning agent during the learning process performed when executing the proposed GMPPT algorithm for shading pattern 1 does not affect the number of search steps required when applying shading patterns 2 and 3, respectively. After the learning process of shading patterns 1–3 has been accomplished, the agent needs less time in order to detect the GMPP location for shading pattern 4, which was unknown till that time, with an MPPT efficiency of 99.7%. At that stage, the agent knows how to react when the unknown shading patterns 5–7 are applied, which further reduces significantly the number of search steps required to converge close to the GMPP. Finally, after the execution of the proposed GMPPT algorithm for shading patterns 1–7, the agent needed only 12 and 4 search steps, respectively, when shading patterns 8 and 9 were applied, which were also unknown till that time, and simultaneously achieved an MPPT efficiency of 99.3%–99.6%. Therefore, it is clear that the training of the agent was successful and the subsequent application of the P&O MPPT process is necessary only for compensation of the short-term changes of the GMPP position without re-execution of the entire Q-learning-based GMPPT process.
As demonstrated in Table 6, the PSO-based GMPPT algorithm needs 11.3–11.6 sec in order to detect the position of the GMPP for each shading pattern, with an MPPT efficiency of 99.6%–99.9%. This is due to the fact that each time that the shading pattern changes, the positions of the particles comprising the swarm are re-initialized and any prior knowledge that was acquired about the location of the GMPP is lost. In contrast, due to its learning capability, the proposed Q-learning-based GMPPT algorithm reduced the convergence time when shading patterns 5–9 were applied by 80.5%–98.3%, compared to the convergence time required by the PSO-based GMPPT process to detect the GMPP with a similar MPPT efficiency.
The MPPT efficiency achieved by the proposed Q-learning-based GMPPT algorithm was lower than that of the PSO-based algorithm by 1.3% when shading pattern 1 was applied, since the learning process was still in evolution. However, the application of a few additional shading patterns (i.e., shading patterns 2–7) offered additional knowledge to the Q-learning agent. Therefore, the resulting MPPT efficiency had already been improved during application of shading patterns 8 and 9 and reached a value similar to that obtained by the PSO-based algorithm, but with significantly less search steps, as analyzed above.
Figure 19a presents an example of the trajectory followed during the PV array output power maximization process, when the proposed Q-learning-based GMPPT method is applied for shading pattern 9. Figure 19b illustrates the initial positions of the particles comprising the PSO swarm (red dots), the positions of the particles during all intermediate swarm generations (black dots), as well as their final positions (blue dots) for the same shading pattern. It is observed that the PSO algorithm must visit a large number of alternative operating points of the PV array in order to be able to detect the position of the GMPP with an MPPT efficiency similar to that of the proposed Q-learning-based GMPPT algorithm (Table 5 and Table 6). In contrast, the proposed Q-learning-based GMPPT technique significantly reduces the number of search steps (i.e., only four steps are required), even if an unknown partial shading pattern is applied, since prior knowledge acquired during its previous executions is retained and exploited in the future executions of the proposed GMPPT algorithm.

4. Conclusions

The PV modules synthesizing the PV array of a PV energy production system may operate under fast-changing partial shading conditions (e.g., in wearable PV systems, building-integrated PV systems, etc.). In order to maximize the energy production of the PV system under such operating conditions, a global maximum power point tracking (GMPPT) process must be applied to detect the position of the global maximum power point (MPP) in the minimum possible search time.
This paper presented a novel GMPPT method which is based on the application of a machine-learning algorithm. Compared to the existing GMPPT techniques, the proposed method has the advantage that it does not require knowledge of either the operational characteristics of the PV modules comprising the PV system or the PV array structure. Also, due to its learning capability, it is capable of detecting the GMPP in significantly fewer search steps, even when unknown partial shading patterns are applied to the PV array. This feature is due to the inherent capability of the proposed GMPPT algorithm to retain the prior knowledge acquired about the behavior and attributes of its environment (i.e., the partially-shaded PV array of the PV system) during its previous executions and exploit this knowledge during its future executions in order to detect the position of the GMPP in less time.
The numerical results presented in the paper demonstrated that by applying the proposed Q-learning-based GMPPT algorithm, the time required for detecting the global MPP when unknown partial shading patterns are applied is reduced by 80.5%–98.3% compared to the convergence time required by a GMPPT process based on the PSO algorithm, while, simultaneously achieving a similar GMPP detection accuracy (i.e., MPPT efficiency).
Future work includes the experimental evaluation of the proposed GMPPT method in order to assess its effect on the long-term energy production performance of PV systems, subject to rapidly changing incident solar irradiation conditions.

Author Contributions

Conceptualization, E.K. and M.G.L.; Methodology, C.K., E.K. and M.G.L.; Software, C.K.; Validation, C.K. and E.K.; Writing—Original Draft Preparation, C.K. and E.K.; Writing—Review & Editing, C.K., E.K. and M.G.L.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lai, C.S.; McCulloch, M.D. Levelized cost of electricity for solar photovoltaic and electrical energy storage. Appl. Energy 2017, 190, 191–203. [Google Scholar] [CrossRef]
  2. Koutroulis, E.; Blaabjerg, F. Overview of maximum power point tracking techniques for photovoltaic energy production systems. Electr. Power Compon. Syst. 2015, 43, 1329–1351. [Google Scholar] [CrossRef]
  3. Radjai, T.; Gaubert, J.-P.; Rahmani, L. A new maximum power point tracking algorithm for partial shaded photovoltaic systems. In Proceedings of the 7th International Conference on Systems and Control (ICSC), Valencia, Spain, 24–26 October 2018; pp. 169–173. [Google Scholar]
  4. Hosseini, S.; Taheri, S.; Farzaneh, M.; Taheri, H. A high-performance shade-tolerant MPPT based on current-mode control. IEEE Trans. Power Electron. 2019, 34, 10327–10340. [Google Scholar] [CrossRef]
  5. Selvakumar, S.; Madhusmita, M.; Koodalsamy, C.; Simon, S.P.; Sood, Y.R. High-speed maximum power point tracking module for PV systems. IEEE Trans. Ind. Electron. 2019, 66, 1119–1129. [Google Scholar] [CrossRef]
  6. Aquib, M.; Jain, S.; Agarwal, V. A time-based global maximum power point tracking technique for PV system. IEEE Trans. Power Electron. 2020, 35, 393–402. [Google Scholar] [CrossRef]
  7. Başoğlu, M.E. An improved 0.8VOC model based GMPPT technique for module level photovoltaic power optimizers. IEEE Trans. Ind. Appl. 2019, 55, 1913–1921. [Google Scholar] [CrossRef]
  8. Keyrouz, F. Enhanced Bayesian based MPPT controller for PV systems. IEEE Power Energy Technol. Syst. J. 2018, 5, 11–17. [Google Scholar] [CrossRef]
  9. Li, X.; Wen, H.; Hu, Y.; Jiang, L.; Xiao, W. Modified beta algorithm for GMPPT and partial shading detection in photovoltaic systems. IEEE Trans. Power Electron. 2018, 33, 2172–2186. [Google Scholar] [CrossRef]
  10. Ghassami, A.A.; Sadeghzadeh, S.M. A communication-based method for PSC detection and GMP tracking under PSC. IEEE Trans. Ind. Inform. 2019, 1. [Google Scholar] [CrossRef]
  11. Li, H.; Yang, D.; Su, W.; Lü, J.; Yu, X. An overall distribution particle swarm optimization MPPT algorithm for photovoltaic system under partial shading. IEEE Trans. Ind. Electron. 2019, 66, 265–275. [Google Scholar] [CrossRef]
  12. Kermadi, M.; Salam, Z.; Ahmed, J.; Berkouk, E.M. An effective hybrid maximum power point tracker of photovoltaic arrays for complex partial shading conditions. IEEE Trans. Ind. Electron. 2019, 66, 6990–7000. [Google Scholar] [CrossRef]
  13. Ram, J.P.; Pillai, D.S.; Rajasekar, N.; Strachan, S.M. Detection and identification of global maximum power point operation in solar PV applications using a hybrid ELPSO-P&O tracking technique. IEEE J. Emerg. Sel. Top. Power Electron. 2019, 1. [Google Scholar] [CrossRef] [Green Version]
  14. Ma, X.; Jiandong, D.; Xiao, W.; Tuo, S.; Yanhang, W.; Ting, S. Research of photovoltaic systems MPPT based on improved grey wolf algorithm under partial shading conditions. In Proceedings of the 2nd IEEE Conference on Energy Internet and Energy System Integration (EI2), Beijing, China, 20–22 October 2018; pp. 1–6. [Google Scholar]
  15. Sampaio, L.P.; da Rocha, M.V.; da Silva, S.A.O.; de Freitas, M.H.T. Comparative analysis of MPPT algorithms bio-inspired by grey wolves employing a feed-forward control loop in a three-phase grid-connected photovoltaic system. IET Renew. Power Gener. 2019, 13, 1–12. [Google Scholar]
  16. Mohanty, S.; Subudhi, B.; Ray, P.K. A new MPPT design using grey wolf optimization technique for photovoltaic system under partial shading conditions. IEEE Trans. Sustain. Energy 2016, 7, 181–188. [Google Scholar] [CrossRef]
  17. Nansur, A.R.; Murdianto, F.D.; Hermawan, A.S.L. Improving the performance of MPPT coupled inductor SEPIC converter using flower pollination algorithm (FPA) under partial shading condition. In Proceedings of the 2018 International Electronics Symposium on Engineering Technology and Applications (IES-ETA), Bali, Indonesia, 29–30 October 2018; pp. 1–7. [Google Scholar]
  18. Ram, J.P.; Rajasekar, N. A novel flower pollination based global maximum power point method for solar maximum power point tracking. IEEE Trans. Power Electron. 2017, 32, 8486–8499. [Google Scholar]
  19. Huang, C.; Wang, L.; Yeung, R.S.-C.; Zhang, Z.; Chung, H.S.-H.; Bensoussan, A. A prediction Model-Guided Jaya Algorithm for the PV System Maximum Power Point Tracking. IEEE Trans. Sustain. Energy 2018, 9, 45–55. [Google Scholar] [CrossRef]
  20. Tey, K.S.; Mekhilef, S.; Seyedmahmoudian, M.; Horan, B.; Oo, A.T.; Stojcevski, A. Improved differential evolution-based MPPT algorithm using SEPIC for PV systems under partial Shading conditions and load variation. IEEE Trans. Ind. Inform. 2018, 14, 4322–4333. [Google Scholar] [CrossRef]
  21. Megantoro, P.; Nugroho, Y.D.; Anggara, F.; Pakha, A.; Pramudita, B.A. The implementation of genetic algorithm to MPPT technique in a DC/DC buck converter under partial shading Condition. In Proceedings of the 2018 3rd International Conference on Information Technology, Information System and Electrical Engineering (ICITISEE), Yogyakarta, Indonesia, 13–14 November 2018; pp. 308–312. [Google Scholar]
  22. Eltamaly, A.M.; Farh, H.M.H.; Al-Saud, M.S. Grade point average assessment for metaheuristic GMPP techniques of partial shaded PV systems. IET Renew. Power Gener. 2019, 13, 1–16. [Google Scholar] [CrossRef]
  23. Nasr, A.N.A.; Saied, H.; Mostafa, M.Z.; Abdel-Moneim, T.M. A survey of MPPT techniques of PV systems. IEEE Energytech 2012, 1, 1–17. [Google Scholar] [CrossRef]
  24. Sutton, R.; Barto, A. Reinforcement Learning; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  25. Watkins, C.J.; Dayan, P. Q-learning. Mach. Learn. 1992, 8, 279–292. [Google Scholar] [CrossRef]
  26. Liu, T.; Zou, Y.; Liu, D.; Sun, F. Reinforcement learning–based energy management strategy for a hybrid electric tracked vehicle. Energies 2015, 8, 7243–7260. [Google Scholar] [CrossRef] [Green Version]
  27. Fachantidis, A.; Taylor, M.E.; Vlahavas, I. Learning to teach reinforcement learning agents. Mach. Learn. Knowl. Extr. 2019, 1, 21–42. [Google Scholar] [CrossRef] [Green Version]
  28. Wei, C.; Zhang, Z.; Qiao, W.; Qu, L. Reinforcement-learning-based intelligent maximum power point tracking control for wind energy conversion systems. IEEE Trans. Ind. Electron. 2015, 62, 6360–6370. [Google Scholar] [CrossRef]
  29. Li, H.; Liu, X.; Lu, J. Research on linear active disturbance rejection control in DC/DC Boost converter. Electronics 2019, 8, 1249. [Google Scholar] [CrossRef] [Green Version]
  30. Liu, Y.H.; Huang, S.C.; Huang, J.W.; Liang, W.C. A particle swarm optimization-based maximum power point tracking algorithm for PV systems operating under partially shaded conditions. IEEE Trans. Energy Convers. 2012, 27, 1027–1035. [Google Scholar] [CrossRef]
Figure 1. An example of the power-voltage characteristic of a photovoltaic (PV) array under partial shading conditions.
Figure 1. An example of the power-voltage characteristic of a photovoltaic (PV) array under partial shading conditions.
Applsci 10 00700 g001
Figure 2. A block diagram of the PV system under consideration.
Figure 2. A block diagram of the PV system under consideration.
Applsci 10 00700 g002
Figure 3. A simplified representation of the process implemented by the Q-learning algorithm in order to control a PV system for the implementation of the global maximum power point tracking (GMPPT) process.
Figure 3. A simplified representation of the process implemented by the Q-learning algorithm in order to control a PV system for the implementation of the global maximum power point tracking (GMPPT) process.
Applsci 10 00700 g003
Figure 4. A flowchart of the proposed Q-learning-based PV GMPPT algorithm.
Figure 4. A flowchart of the proposed Q-learning-based PV GMPPT algorithm.
Applsci 10 00700 g004
Figure 5. The model of the PV system under study in MATLABTM/Simulink.
Figure 5. The model of the PV system under study in MATLABTM/Simulink.
Applsci 10 00700 g005
Figure 6. The distribution of incident solar irradiation in W/m2 over each PV module of the PV array for: (a) shading pattern 1, (b) shading pattern 2, (c) shading pattern 3, (d) shading pattern 4, (e) shading pattern 5, (f) shading pattern 6, (g) shading pattern 7, (h) shading pattern 8 and (i) shading pattern 9.
Figure 6. The distribution of incident solar irradiation in W/m2 over each PV module of the PV array for: (a) shading pattern 1, (b) shading pattern 2, (c) shading pattern 3, (d) shading pattern 4, (e) shading pattern 5, (f) shading pattern 6, (g) shading pattern 7, (h) shading pattern 8 and (i) shading pattern 9.
Applsci 10 00700 g006aApplsci 10 00700 g006b
Figure 7. Τhe PV array output power–voltage curves for shading patterns 1–9.
Figure 7. Τhe PV array output power–voltage curves for shading patterns 1–9.
Applsci 10 00700 g007
Figure 8. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 1: (a) Duty cycle versus time and (b) PV array output power versus time.
Figure 8. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 1: (a) Duty cycle versus time and (b) PV array output power versus time.
Applsci 10 00700 g008
Figure 9. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 2: (a) duty cycle versus time and (b) PV array output power versus time.
Figure 9. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 2: (a) duty cycle versus time and (b) PV array output power versus time.
Applsci 10 00700 g009
Figure 10. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 3: (a) duty cycle versus time and (b) PV array output power versus time.
Figure 10. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 3: (a) duty cycle versus time and (b) PV array output power versus time.
Applsci 10 00700 g010
Figure 11. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 4: (a) duty cycle versus time and (b) PV array output power versus time.
Figure 11. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 4: (a) duty cycle versus time and (b) PV array output power versus time.
Applsci 10 00700 g011
Figure 12. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 5: (a) duty cycle versus time and (b) PV array output power versus time.
Figure 12. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 5: (a) duty cycle versus time and (b) PV array output power versus time.
Applsci 10 00700 g012
Figure 13. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 6: (a) duty cycle versus time and (b) PV array output power versus time.
Figure 13. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 6: (a) duty cycle versus time and (b) PV array output power versus time.
Applsci 10 00700 g013
Figure 14. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 7: (a) duty cycle versus time and (b) PV array output power versus time.
Figure 14. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 7: (a) duty cycle versus time and (b) PV array output power versus time.
Applsci 10 00700 g014
Figure 15. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 8: (a) duty cycle versus time and (b) PV array output power versus time.
Figure 15. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 8: (a) duty cycle versus time and (b) PV array output power versus time.
Applsci 10 00700 g015
Figure 16. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 9: (a) duty cycle versus time and (b) PV array output power versus time.
Figure 16. Results during execution of the proposed Q-learning-based GMPPT process for shading pattern 9: (a) duty cycle versus time and (b) PV array output power versus time.
Applsci 10 00700 g016
Figure 17. Results during execution of the PSO-based GMPPT process for shading pattern 1: (a) duty cycle versus time and (b) PV array output power versus time.
Figure 17. Results during execution of the PSO-based GMPPT process for shading pattern 1: (a) duty cycle versus time and (b) PV array output power versus time.
Applsci 10 00700 g017
Figure 18. Results during execution of the PSO-based GMPPT process for shading pattern 9: (a) duty cycle versus time and (b) PV array output power versus time.
Figure 18. Results during execution of the PSO-based GMPPT process for shading pattern 9: (a) duty cycle versus time and (b) PV array output power versus time.
Applsci 10 00700 g018
Figure 19. The GMPPT process trajectory for shading pattern 9 during the execution of: (a) the proposed Q-learning-based method and (b) the PSO-based method.
Figure 19. The GMPPT process trajectory for shading pattern 9 during the execution of: (a) the proposed Q-learning-based method and (b) the PSO-based method.
Applsci 10 00700 g019
Table 1. Operational characteristics of each PV module.
Table 1. Operational characteristics of each PV module.
Maximum Power ( W ) 11.6
Open-circuit voltage, Voc ( V ) 7.25
Voltage at maximum power point VMP ( V ) 5.75
Temperature coefficient of Voc ( % / deg ° C ) −0.322
Cells per module (Ncell)24
Short-circuit current Isc ( A ) 2.204
Current at maximum power point Imp ( A ) 2.016
Temperature coefficient of Isc ( % / deg ° C ) 0.071996
Table 2. Operational parameters of the proposed Q-learning-based GMPPT method.
Table 2. Operational parameters of the proposed Q-learning-based GMPPT method.
Initial duty cycle0.71Minimum duty cycle0.19
τmin, τmax0.08, 0.8Maximum duty cycle0.83
d1k1, k2, k310, 25, 0.6
Discount factor0.9Nmax20
Table 3. Action-space of the proposed Q-learning-based GMPPT method.
Table 3. Action-space of the proposed Q-learning-based GMPPT method.
Low Change of Duty CycleMedium Change of Duty CycleHigh Change of Duty Cycle
Increment of ΔD +0.04+0.12+0.28
Decrement of ΔD -0.04-0.12-0.28
No change of ΔD000
Table 4. Operational parameters of the particle swarm optimization (PSO)-based GMPPT method.
Table 4. Operational parameters of the particle swarm optimization (PSO)-based GMPPT method.
Number of particles8Inertia weight w0.6
Maximum duty cycle0.83Cognitive coefficient c11.6
Minimum duty cycle0.19Social coefficient c21.5
Table 5. The simulation results of the proposed Q-learning-based GMPPT method for various shading patterns.
Table 5. The simulation results of the proposed Q-learning-based GMPPT method for various shading patterns.
Convergence Time (s)Number of Search Steps P G M P P   ( W ) P P V   ( W ) MPPT Efficiency (%)
Shading pattern 12946483.282.098.6
Shading pattern 23251267.066.298.8
Shading pattern 32844875.273.097.1
Shading pattern 41727271.170.999.7
Shading pattern 523279.178.799.5
Shading pattern 61.52480.780.399.5
Shading pattern 72.23569.568.598.6
Shading pattern 80.731275.975.499.3
Shading pattern 90.25476.976.699.6
Table 6. The simulation results of the PSO-based GMPPT method for various shading patterns.
Table 6. The simulation results of the PSO-based GMPPT method for various shading patterns.
Convergence Time (s)Number of Search Steps P G M P P   ( W ) P P V   ( W ) MPPT Efficiency (%)
Shading pattern 111.618583.283.199.9
Shading pattern 211.518467.0 66.9 99.9
Shading pattern 311.618575.2 75.1 99.9
Shading pattern 411.518571.1 71.0 99.9
Shading pattern 511.518679.1 79.0 99.9
Shading pattern 611.518580.7 80.4 99.6
Shading pattern 711.318469.5 69.2 99.6
Shading pattern 811.418575.975.899.9
Shading pattern 911.618776.976.899.9

Share and Cite

MDPI and ACS Style

Kalogerakis, C.; Koutroulis, E.; Lagoudakis, M.G. Global MPPT Based on Machine-Learning for PV Arrays Operating under Partial Shading Conditions. Appl. Sci. 2020, 10, 700. https://doi.org/10.3390/app10020700

AMA Style

Kalogerakis C, Koutroulis E, Lagoudakis MG. Global MPPT Based on Machine-Learning for PV Arrays Operating under Partial Shading Conditions. Applied Sciences. 2020; 10(2):700. https://doi.org/10.3390/app10020700

Chicago/Turabian Style

Kalogerakis, Christos, Eftichis Koutroulis, and Michail G. Lagoudakis. 2020. "Global MPPT Based on Machine-Learning for PV Arrays Operating under Partial Shading Conditions" Applied Sciences 10, no. 2: 700. https://doi.org/10.3390/app10020700

APA Style

Kalogerakis, C., Koutroulis, E., & Lagoudakis, M. G. (2020). Global MPPT Based on Machine-Learning for PV Arrays Operating under Partial Shading Conditions. Applied Sciences, 10(2), 700. https://doi.org/10.3390/app10020700

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop