Next Article in Journal
Optimization Environment Definition for Beam Steering Reflectarray Antenna Design
Previous Article in Journal
Stability for Representations of Hecke Algebras of Type A
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A VMD–CISSA–LSSVM Based Electricity Load Forecasting Model

1
School of Information and Electrical Engineering, Shandong Jianzhu University, Jinan 250101, China
2
School of Electrical Engineering and Automation, Qilu University of Technology (Shandong Academy of Sciences), Jinan 250353, China
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(1), 28; https://doi.org/10.3390/math10010028
Submission received: 1 December 2021 / Revised: 9 December 2021 / Accepted: 20 December 2021 / Published: 22 December 2021

Abstract

:
Accurate power load forecasting has an important impact on power systems. In order to improve the load forecasting accuracy, a new load forecasting model, VMD–CISSA–LSSVM, is proposed. The model combines the variational modal decomposition (VMD) data preprocessing method, the sparrow search algorithm (SSA) and the least squares support vector machine (LSSVM) model. A multi-strategy improved chaotic sparrow search algorithm (CISSA) is proposed to address the shortcomings of the SSA algorithm, which is prone to local optima and a slow convergence. The initial population is generated using an improved tent chaotic mapping to enhance the quality of the initial individuals and population diversity. Second, a random following strategy is used to optimize the position update process of the followers in the sparrow search algorithm, balancing the local exploitation performance and global search capability of the algorithm. Finally, the Levy flight strategy is used to expand the search range and local search capability. The results of the benchmark test function show that the CISSA algorithm has a better search accuracy and convergence performance. The volatility of the original load sequence is reduced by using VMD. The optimal parameters of the LSSVM are optimized by the CISSA. The simulation test results demonstrate that the VMD–CISSA–LSSVM model has the highest prediction accuracy and stabler prediction results.

1. Introduction

Electricity load forecasting is of great importance to the development of modern power systems. Stable and efficient management and scheduling strategies for power systems rely heavily on accurate forecasts of future loads at different times [1]. Accurate short-term load forecasting can help national grids and energy suppliers cope with the increasing complexity of pricing strategies in future smart grids, further increase the utilization of renewable energy and meet the challenges posed by the development of electricity [2].
In recent years, the research on electricity load forecasting can be divided into traditional forecasting methods based on mathematical statistics and forecasting methods based on artificial intelligence (AI). Traditional forecasting models can be classified into exponential smoothing [3], Kalman filtering [4] and multiple linear regression models [5]. Traditional forecasting methods rely on statistical models to analyze the regularity of electrical loads during stochastic variations and cannot effectively solve complex problems of non-linearity. In order to better solve the problem of complex nonlinear time series, AI-based forecasting methods have been widely discussed and applied. AI-based prediction methods include artificial neural networks (ANNs) [6,7], support vector machines (SVM) [8,9] and fuzzy prediction methods [10]. For example, [11] combined a real number coded genetic algorithm (GA) with a BP neural network (BPNN) for short-term gas load forecasting. In addition, the feasibility of combining the GA algorithm with the BPNN and the superiority of the prediction model performance were verified by simulation experiments. Another study [12] proposed a hybrid annual electricity load forecasting model combining a fruit fly optimization algorithm (FOA) and a generalized regression neural network (GRNN). The FOA algorithm was used to determine the parameter values in the GRNN model to improve the forecasting accuracy of the model. A third study [13] proposed a novel evolutionary algorithm based on the “leader-following” behavior of sheep. This algorithm was combined with an ANN for electricity load forecasting. Simulation experiments demonstrated that this approach can achieve the desired results satisfactorily. A fourth study [14] combined particle a swarm algorithm (PSO) with an ANN. The PSO algorithm was used to find the optimal network weights to improve the prediction performance of the prediction model.
In addition, the least square support vector machine (LSSVM) was well suited to handle complex nonlinear power load sequences because of its superior self-learning and self-adaptive capabilities. However, the performance of the LSSVM was highly susceptible to the penalty factor gam and the RBF kernel parameter sig. Therefore, researchers used artificial intelligence algorithms to find the optimal parameters. For example, [15] proposed an electricity load forecasting model that combines the K-average algorithm, the reinforcement learning improved Q-PSO algorithm and the LSSVM algorithm. Finding parameters for LSSVM using a PSO algorithm was improved by the reinforcement learning strategy. The production patterns of industrial customers were identified using the K-average algorithm. Improving The data-specific prediction accuracy was improved by sacrificing the data generality. Another study [16] proposed an electricity load forecasting model combining variational modal decomposition (VMD), maximum relevance minimum redundancy (MRMR), a BPNN neural network and the LSSVM model. VMD was used to decompose the original data series into several high-, medium- and low-frequency components. High-frequency variables were predicted using the BPNN model. Medium- and low-frequency variables were predicted using the LSSVM model. Another study [17] proposed a combined load prediction model based on a singular spectrum analysis data preprocessing approach, a neural network model and the LSSVM model. The artificial intelligence algorithm was used to find the optimal parameters of the model. The prediction results were obtained using a weighted combination of computational methods.
At the same time, researchers have found that standard intelligent optimization algorithms have difficulties with achieving their theoretical optimality due to their own structural deficiencies. Therefore, a multi-strategy optimization approach was used to improve and optimize the intelligent optimization algorithm to address the shortcomings of the algorithm. For example, [18] proposed a new electricity load forecasting model combining the grey wolf algorithm (GWO) and Elman neural network. Chaotic sequences and chaotic cosine inertia weighting strategies were introduced on the basis of the GWO algorithm to improve the shortcomings of the algorithm’s insufficient global merit-seeking ability. Another study [19] proposed a chicken flock optimizer based on a nonlinear dynamic convergence factor algorithm. In addition, the nonlinear dynamic inertia weights and Levy variation strategy were introduced into the optimizer to improve the convergence speed of the algorithm. The improved algorithm was used to find the optimal initial weights and thresholds of the ELM neural network in order to build a new electricity load forecasting model. Another study [20] proposed a sparrow search algorithm (SSA) and ELM neural network based electricity load forecasting model. The tent chaos strategy and firefly perturbation strategy were used to improve a problem with the SSA algorithm, namely that it had insufficient global search capability and easily fell into a local optimum. Comparative experiments with several competing models also confirmed that the method was effective in improving the prediction accuracy of the model. Another study [21] proposed a storm flood loss estimation model based on the SSA algorithm, mean impact value (MIV) and LSSVM model for economic losses caused by flooding in metro stations. The paper demonstrated not only the good prediction accuracy of the model but also the reliability of the LSSVM model when dealing with large data time series.
The sparrow search algorithm is a new type of swarm intelligence optimization algorithm proposed in 2020 and is widely used in various fields [22]. The SSA algorithm has a strong global optimization capability and stability, but it still suffers from an insufficient optimization capability and slow convergence speed, and it easily falls into local optimality when encountering complex problems. Researchers have proposed a number of solutions to address the shortcomings of the SSA algorithm. For example, [23] proposed a fused cross-variant sparrow search algorithm. The algorithm used tent chaotic mapping to initialize the population to increase the population diversity. The crossover and variation ideas of the genetic algorithm were used to improve the position update equation of the SSA algorithm and help the algorithm to jump out of the local optimum. The chaotic flying sparrow search algorithm was proposed in [24]. The improved algorithm was optimized mainly in the position update phase of the sparrow. In the search discovery phase, dynamic adaptive weights and levy flight mechanisms were combined to improve the search range and flexibility of the algorithm. The backward learning strategy based on lens imaging was introduced into the follower’s position update process to help the algorithm balance local and global search. Another study [25] presented an improved sparrow search algorithm applied to the field of photovoltaic microgrids. The improved algorithm used a gravity inverse learning mechanism to initialize the population. Learning coefficients are introduced into the sparrow finder position update process to improve the global optimization capability. The variation operator is introduced into the joiner position update process to help the algorithm jump out of the local optimum.
In addition, it was found that data pre-processing techniques can effectively reduce the effect of noise in the raw data on the prediction results. For example, [26] proposed a combined forecasting model based on improved empirical modal decomposition (IEMD), autoregressive integrated moving average (ARIMA) and wavelet neural network (WNN) optimized based on the FOA algorithm. IEMD was used to reduce the noise of the original data. Simulation experiments not only verified the excellent prediction performance of the model, but also confirmed that data pre-processing has a positive impact on the prediction results. Another study [27] proposed a novel electricity load forecasting model based on data preprocessing and a multi-objective cuckoo search algorithm based on non-dominated ranking to optimize the GRNN. Fast empirical modal decomposition by integration (FEEMD) was used to reduce the interference of raw data. Another study [28] used the ensemble empirical mode decomposition (EEMD) to decompose the raw load data and then used the Elman neural network to make predictions. Although empirical mode decomposition (EEMD) and empirical mode decomposition (EEMD) could automatically decompose the modal components based on the data, the addition of white noise to the EEMD and EEMD during the decomposition process could create endpoint effects and cause distortion. VMD enables the effective separation of the intrinsic modal components and the division of the frequency domain of the signal to avoid distortions caused by the endpoint effect.
Table 1 shows a further summary and analysis of the above literature. The following conclusions are obtained from Table 1: the use of the idea of combined models to construct prediction models, a reasonable signal noise reduction approach and a multi-strategy optimization approach can effectively improve the prediction accuracy of power load prediction models. Studies [11,12,13,14,21] used standard intelligent optimization algorithms to optimize the network weights of neural networks in order to construct prediction models. Although such prediction models also had high prediction accuracy, the authors did not take into account the impact of the standard intelligent optimization algorithm’s own shortcomings on the optimization process and the impact of non-linear fluctuations in the original data on the prediction results. While [15,18,19,20] considered the impact of multi-strategy optimization approaches on intelligent optimization algorithms, the authors ignored the improvement of the prediction accuracy by data pre-processing methods. In addition, although [26,27,28] integrated the idea of combinatorial modeling, multi-strategy optimization and data pre-processing, the authors did not consider the endpoint effects inherent in EMD denoising that can also affect the final prediction results.
In summary, a new combined power load forecasting model based on variational modal decomposition (VMD) and an improved chaotic sparrow search algorithm (CISSA) to optimize the LSSVM model is proposed. First, we address the problem that the standard sparrow search algorithm is prone to fall into local extremes as the population diversity decreases in the late iterations. In this paper, an improved chaotic sparrow optimization algorithm (CISSA) is proposed based on the analysis of the SSA algorithm. The improved tent mapping strategy, the random following strategy in the chicken flock optimization algorithm idea and the Levy flight strategy in the cuckoo algorithm idea are improved for the population initialization phase, algorithm iteration phase and global optimization search phase of the algorithm, respectively. Second, the original load sequence is decomposed into several modal components of different frequencies by VMD. The CISSA algorithm is used to calculate the two parameters of the LSSVM model, the penalty factor gam and the RBF kernel parameter sig. The CISSA-LSSVM prediction model is then used to train and predict the components at different frequencies separately. Finally, the predicted values of each component are integrated to produce the final prediction results.
In order to verify the performance of the CISSA algorithm proposed in this paper, 8 test benchmark functions are used to evaluate the optimization capability of the CISSA algorithm. The comparison with two improved SSA algorithms and three basic algorithms verifies that the CISSA algorithm has better search accuracy, convergence performance and stability. Finally, simulation experiments using real historical load data are conducted to verify the prediction accuracy and stability of the model. The simulation results compared with several competing models also demonstrate the excellent prediction accuracy and performance of the VMD–CISSA–LSSVM prediction model.

2. Theory and Methods

This section presents the mathematical theory and models of the variational modal decomposition, the LSSVM model, the sparrow search algorithm, the improved chaotic sparrow search algorithm and the VMD–CISSA–LSSVM model.

2.1. Variational Modal Decomposition

VMD is an adaptive decomposition method for non-smooth signals, which can determine the number of modal decompositions according to the actual situation of the sequence. The optimal solution is obtained by adaptively matching the frequency bandwidth of each mode to the optimal frequency bandwidth of each class of modes during the solution process. The specific mathematical model of VMD is shown in [29]. The specific process of VMD decomposition is shown as follows:
  • The Hilbert transform is applied to each sub-mode and the one-sided spectrum ( δ ( t ) + j π t ) u n ( t ) of the resolved signal is obtained;
  • There is a transformation of the spectrum to a baseband where the spectrum is multiplied by the central frequency of an exponential signal estimate: ( ( δ ( t ) + j π t ) u n ( t ) ) e j ω n t ;
  • The bandwidth is estimated by demodulating the signal and its constrained variational problem can be expressed as Equation (1):
    { min | u n | , | w n | = { n = 1 N ( ( δ ( t ) + j π t ) u n ( t ) ) e j ω n t 2 2 } n = 1 N u n ( t ) = f ( t ) ;
  • The quadratic penalty factor α and Lagrange multiplier λ ( t ) are introduced to turn it into an unconstrained variational problem to be solved:
    L ( | u n | , | w n | , λ ) = α n = 1 N t [ ( δ ( t ) + j π t ) u n ( t ) ] e j ω n t 2 2 + f ( t ) n = 1 N u n ( t ) + λ ( t ) , f ( t ) n = 1 N u n ( t ) ;
  • The alternating direction multiplier method is used to update the values of A and B, as shown in Equation (3):
    { u n k + 1 ( ω ) = f ( w ) n = 1 N u n ( ω ) + λ 2 1 + 2 α ( ω ω n ) 2 ω n k + 1 = 0 ω | u n ( ω ) | 2 d ω 0 | u n ( ω ) | 2 d ω ,
    where δ ( t ) is the unit pulse signal; n is the n-th modal component obtained after the signal decomposition; N is the total number of modal decompositions; k is the number of iterations; ω n is the central frequency of the modal; t is the sign of the partial derivative operation; α is the penalty factor; j is the unity of the imaginary number; is the convolution operator; λ is the Lagrange multiplier; f ( w ) , u n ( ω ) and λ are the Fourier transforms of f ( t ) , u n ( t ) and ω n ( t ) , respectively; u n ( t ) is the finite bandwidth of the component; and w n is the central frequency of the component.

2.2. Least Square Support Vector Machines

The inequality constraint in the SVM algorithm is replaced by an equation constraint and the sum of squared errors is used as the empirical loss. In addition, the selection of penalty factor and kernel function parameters in LSSVMs directly affects LSSVMs’ anti-interference ability and generalization ability. The specific mathematical model of LSSVMs is shown in [30].
For a given training set T = x 1 , y 1 , , x n , y n , its regression function can be defined as Equation (4):
y = f ( t ) = ω T x + b ,
where x is the sample input, y is the sample output and ω and b are the normal vector and intercept of the hyperplane in the higher dimensional space, respectively. According to the risk minimization principle, the regression problem can be transformed into a constraint problem:
{   min ω , b , e J ( ω , e ) = 1 2 ω T ω + 1 2 γ i = 1 N e i 2 s . t . y i = ω T φ ( x i ) + b + e i , i = 1 , , N ,
where e i is the relaxation variable and γ is the regularization factor. By introducing the Lagrange multiplier α , the above problem is transformed into Equation (6):
L ( ω , b , e , α ) = J ( ω , e ) i = 1 N α i ω T φ ( x i ) + b + e i y i ,
The optimal values are obtained by the partial differentiation of ω , b, e and α , respectively, and the regression function is then established:
y ( x ) = i = 1 N α i K ( x , x i ) + b ,
where K ( x , x i ) is the kernel function and the RBF kernel function is used in this paper. The expression is as shown in Equation (8):
K ( x , x i ) = exp ( x i , x j 2 2 σ 2 ) ,
where σ is the RBF kernel parameter.

2.3. Sparrow Search Algorithm

The sparrow search algorithm [31] is a new swarm intelligence optimization algorithm proposed by Xue in 2020. In this paper, a rational analysis is carried out based on the SSA algorithm so as to develop a reasonable optimization scheme.
The initial sparrow individuals in the sparrow search algorithm are randomly generated in the search space and gradually aggregated during the iterative process, making it difficult to obtain a good population diversity and maintain it at a certain level. This leads to a poor convergence performance and an inconsistency between the global search capability and local exploitation performance of the algorithm.
Sparrow populations are divided into searchers, followers and vigilantes, depending on their individual capabilities. The searcher’s position is updated by the following Equation (9):
x i d t + 1 = { x i d t exp ( i T ) , R 2 < S T x i d t + Q L , R 2 S T ,
where t is the current number of iterations; is the random number between [ 0 , 1 ] ; R 2 [ 0 , 1 ] and S T [ 1 2 , 1 ] represent the warning value and the safety threshold, respectively; L is a matrix of 1 × d whose elements are all 1; and Q is a random number subject to a normal distribution.
The equation for updating the position of a follower is as follows:
x i d t + 1 = { x i d t exp ( x w d t x i d t i 2 ) , i > n 2 x b d t + 1 + ( r a n d { 1 , 1 } | x i d t x b d t + 1 | ) × L , i n 2 ,
where x w d t denotes the worst position of the sparrow in the d-th dimension in the t-th iteration of the population, x b d t + 1 denotes the optimal position of the sparrow in the d-th dimension in the t+1-th iteration of the population and L is the unit matrix of 1 × d .
The equation for updating the location of the vigilantes is as follows:
x i d t + 1 = { x b t + β ( x i d t x b d t ) , f i f g x i d t + K ( x i d t x w d t | f i f w | + ) , f i = f g ,
where is the minimum constant; K is a random number within [ 1 , 1 ] ; f i , f g and f w are the current adaptation fitness, the best adaptation fitness and the worst adaptation fitness, respectively; β is the number of iteration steps. When f i f g , the sparrow is at the edge of the population and is vulnerable to predators; when f i = f g , the sparrow is in the middle of the population, is aware of the threat of predators and adjusts its search strategy by moving closer to other sparrows in time to avoid being attacked by predators.
From Equation (10), it can be seen that the follower position update process is mainly guided by x w d t and x b d t + 1 . This also shows that the SSA algorithm does not take full advantage of the information carried by most common individuals in the population. As a result, the effective exploration area for sparrows is small and the algorithm is weak at the global scale.
The operation flow of the standard SSA algorithm is shown in Figure 1. The iterative search process for individual sparrows shows that the strength of the sparrow search algorithm is influenced by the quality of the individuals in the population and the location update parameters. The individual sparrow position updating relies on the inter-individual following and interactions. Due to the lack of variation in the iterative update process of individuals, once the local optimum stagnation is reached it is difficult for sparrows to jump out of the current local space.

2.4. Improved Chaotic Sparrow Search Algorithm

In this paper, the CISSA algorithm is proposed based on the analysis of the SSA algorithm. First, an improved tent chaotic mapping is used to generate the initial population to improve the quality of the initial solution and lay the foundation for global optimization. Second, in the iteration of the algorithm the random following strategy of the chicken flock algorithm is used to optimize the position update process of the followers in the SSA algorithm, thus balancing the local exploitation performance and global search capability of the algorithm. Finally, the Levy flight strategy of the cuckoo algorithm is introduced to improve the global search capability of the algorithm and help it to jump out of local constraints. The multi-strategy fusion approach helps the algorithm to balance local exploitation and global search capabilities, while improving the algorithm’s local extreme value escape capability.

2.4.1. Improved tent Mapping Strategy

Chaos is a nonlinear system between deterministic and stochastic systems [32,33]. Chaotic mappings are capable of traversing all states without repetition within a certain range.
Figure 2 shows the bifurcation diagrams of four common chaotic mappings. From Figure 2, it is clear that the tent chaotic map covers a larger area and is more uniformly distributed. Therefore, the tent chaos mapping is chosen to initialize the sparrow population distribution and help the sparrow population to be uniformly distributed in the mapping space. In addition, random variables are introduced into the tent chaos mapping to improve the diversity and randomness of the population.
The tent chaotic mapping can be expressed by Equation (12):
y i + 1 = { 2 y i , 0 y < 0.5 2 ( 1 y i ) , 0.5 y 1 .
Adding random variables r a n d ( 0 , 1 ) × 1 N to Equation (12), Equation (13) is obtained:
y i + 1 = { 2 y i × r a n d ( 0 , 1 ) × 1 N , 0 y < 0.5 2 ( 1 y i ) + r a n d ( 0 , 1 ) × 1 N , 0.5 y 1 .
Finally, the improved tent chaotic mapping is obtained after the Bernoulli shift transformation of y i + 1 = ( 2 y i ) mod 1 + r a n d ( 0 , 1 ) × 1 N . The initial position of the sparrow population in the feasible domain is obtained by Equation (14):
y i d = l b i d + q ( l b i d u b i d ) × x i d ,
where q is a random number within [ 0 , 1 ] ; l b i d and u b i d represent the upper and lower bounds of the feasible solution interval, respectively; and y i d is the individual after mapping. The process can be expressed as follows: a d-dimensional vector is randomly generated in [ 0 , 1 ] as the initial individual. Then N-1 new individuals are generated by iterating over each dimension of the vector in Equation (13). Finally, Equation (14) is used to map the values of the variables generated by the modified tent chaos mapping onto the sparrow individuals.

2.4.2. Random Following Strategy

The followers in the SSA algorithm are prone to rapid population clustering in a short period of time as they move towards the optimal position. Although a fast convergence can be achieved, the probability of the algorithm falling into a local optimum is greatly increased by the sudden drop in population diversity. Therefore, the random following strategy of the chicken flock optimization algorithm is used to improve the position update of the followers in the SSA algorithm. The mathematical model of the chicken swarm optimization is shown in [34]. The random-following strategy of the chicken swarm optimization algorithm is to move the hens closer to the roosters with a certain probability. This ensures a convergence without reducing the diversity of the population and provides a good balance between local exploitation and global search. The Equation for updating the position of the hen is as follows:
X i , j t + 1 = X i , j t + S 1 r a n d ( 0 , 1 ) ( X r , j t X i , j t ) + S 2 r a n d ( 0 , 1 ) ( X r , j t X i , j t ) ,
S 1 = exp ( ( f i f r ) / ( | ( f i ) | + ε ) ) ,
S 2 = exp ( ( f s f i ) ) ,
where r denotes any r-th rooster as the hen’s mate and s denotes any s-th rooster or hen in the flock, r s ; f is the fitness of a randomly selected rooster s; f i is the fitness value of the i-th sparrow.
The improved follower position update formula can be expressed as:
x i d t + 1 = { x i d t exp ( x w d t x i d t i 2 ) , i > n 2 x b d t + 1 + | x i d t x b d t + 1 | × S 3 r a n d ( 0 , 1 ) , i n 2 ,
where S 3 = exp ( ( f s f i ) ) .

2.4.3. Levy Flight Strategy

In the late iterations of the SSA algorithm, individual sparrows have already completed their position updates and are prone to local optimum stagnation at this point. In order to solve this problem, the Levy flight strategy in the cuckoo algorithm is used to update and mutate the population after the SSA algorithm position update [35].
The Levy flight strategy is based on a combination of long-term small-step searches and short-term large-step jumps. The short distance search ensures that a small area around the individual is carefully searched during foraging. Longer walks ensure that the individual is able to move into another area and search over a wider area. Currently, the Mantegna method is commonly used to generate random step sizes that obey the Levy distribution. The formula proposed by Mantegna for simulating the Levy flight path can be expressed as:
s = ρ | v | 1 β ,
{ ρ N ( 0 , σ ρ 2 ) v N ( 0 , σ v 2 ) ,
{ σ ρ = { Γ ( 1 + β ) sin ( π β 2 ) Γ [ ( 1 + β 2 ) ] 2 ( β 1 ) 2 β } σ v = 1 ,
where s is the flight path of Levy; β is a constant, usually taken as 1.5; and ρ and v are normally distributed random numbers, which obey the normal distribution of Equation (20). The standard deviations σ ρ and σ v of the corresponding normal distribution in Equation (20) take values that satisfy Equation (21). The position update formula for Levy’s flight can be expressed as:
x i ( t ) = x i ( t ) + l s ,
l = 0.01 ( x i ( t ) x b e s t ) ,
where x i ( t ) denotes the i-th solution at generation t, x b e s t denotes the optimal solution at this point, l denotes the weight of the control step and denotes the point multiplication.
Figure 3 illustrates the two-dimensional plane-based Levy flight path generated using the Mantegna method. It is clear from Figure 3 that the Mantegna method can be effectively implemented in the search for the optimal solution based on a long-term small-step search and a short-term large-step jump change interphase. By expanding the search space in the short term with large steps, the individual is able to escape from the local stagnation at this point. In addition, the long-term small-step search method is used to enhance the local search capability, effectively solving the problem of individuals falling into local optima. In the standard SSA algorithm, once the sparrow’s position is updated it enters the next phase of the cycle or ends, at which point it tends to fall into a local optimum. By introducing Levy flight variation into the sparrow population in the global search phase, the variation is updated again, helping the population to update its position again and move away from the local optimum at this point.
This paper accomplishes a selective variation update of sparrows after a position update by comparing the size of rand with the inertia weight factor f :
f = 1 i t e r / M a x i t e r ,
where i t e r is the current iteration number. M a x i t e r is the maximum iteration number and rand is a random number within ( 0 , 1 ) . If the selected random number rand is greater than f , the selected sparrow is subjected to Levy flight variation according to Equation (22). If the selected random number rand is less than f , the variation is skipped and the next step is carried out.

2.4.4. The CISSA Algorithm

As shown in Figure 4, the operational flow of the CISSA algorithm can be summarized as follows:
  • Initialize the relevant parameters of the SSA algorithm;
  • Initialize the sparrow population using a tent chaotic mapping with increased random variables. The improved tent chaotic mapping is used to improve the diversity of the sparrow population by using the ergodicity and randomness of the mapping, thus providing a basis for the global optimization of the algorithm. It generates a d-dimensional vector in the initial space as the initial individual. Then, N-1 new individuals are generated by iterating over each of its dimensions by the equation y i + 1 = { 2 y i × r a n d ( 0 , 1 ) × 1 N , 0 y < 0.5 2 ( 1 y i ) + r a n d ( 0 , 1 ) × 1 N , 0.5 y 1 . Finally, the values of the variables generated by the chaotic mapping are mapped onto individual sparrows by the equation x i d = l b i d + q ( l b i d u b i d ) × y i d ;
  • Calculate and rank the fitness values of the sparrows at this time and record the best and worst positions of the sparrows at this time;
  • Update the position of the spotter sparrow at this point according to the equation x i d t + 1 = { x i d t exp ( i T ) , R 2 < S T x i d t + Q L , R 2 S T ;
  • The position of the follower sparrow at this point is updated according to the random following strategy employed in the equation x i d t + 1 = { x i d t exp ( x w d t x i d t i 2 ) , i > n 2 x b d t + 1 + | x i d t x b d t + 1 | × S 3 r a n d ( 0 , 1 ) , i n 2 . The local exploitation performance and global search capability of the algorithm is balanced using the random following strategy;
  • Update the position of the spotter alert at this point according to the equation x i d t + 1 = { x b t + β ( x i d t x b d t ) , f i f g x i d t + K ( x i d t x w d t | f i f w | + ) , f i = f g ;
  • Recalculate and rank the fitness values of the sparrows, recording the best and worst positions of the sparrows at this time;
  • Calculate the inertia weighting factor f = 1 i t e r / M a x i t e r . Whether sparrow populations undergo Levy variation is determined by comparing the magnitude of rand to f . If the selected random number is greater than f , then the selected individual sparrow is subjected to Levy flight variation according to the equation x i ( t ) = x i ( t ) + l s . The Levy flight strategy in the cuckoo algorithm idea is used to improve the global search ability of the algorithm and help the algorithm to jump out of local restrictions;
  • Recalculate the fitness and record of the optimal and worst positions of the sparrow at this time;
  • Determine whether if the stop condition is met. If the stop condition is met, output the result; otherwise, repeat steps 2–9.

2.5. VMD–CISSA–LSSVM Electricity Load Forecasting Model

In summary, a new combined power load forecasting model based on VMD, the CISSA algorithm and the LSSVM model is proposed.
The VMD algorithm is used to decompose the original data to obtain multiple IMF components and Res residual components. The effect of denoising the raw load data is then achieved by means of modal reconstruction. The accuracy of the prediction model is reduced if the sub-series data is fed directly into the LSSVM model for load power prediction. The reason for this is that the penalty factor gam and the RBF kernel parameter sig of the LSSVM have a significant impact on the prediction results. To improve the prediction accuracy, the CISSA algorithm proposed in this paper is used to find the optimal kernel width and penalty factor for these two important parameters and input them into the LSSVM model for load prediction.
The flow of the VMD–CISSA–LSSVM power load forecasting model is shown in Figure 5. The specific operational flow can be expressed as follows:
  • Modal decomposition of load data using the VMD algorithm;
  • The input sub-series data has a large variance in peak values, which can have a significant impact on the prediction results if entered directly without processing. Therefore, the data needs to be normalized before the individual subsequences are fed into the LSSVM. The normalization formula can be expressed as x = x x min x max x min , where x represents the original data and x min and x max represent the minimum and maximum values in the original data;
  • The kernel function width and penalty factor of the LSSVM are optimized using the CISSA algorithm proposed above;
  • The decomposed sub-series data of the original load prediction are fed into the LSSVM prediction model optimized by the CISSA algorithm;
  • The prediction results of each sub-series are summed to obtain the final prediction result.
In addition, the proposed combined prediction model can be applied in a real power transmission environment, as shown in Figure 6. The VMD–CISSA–LSSVM model is applied in the first conversion phase. The model’s forecasting performance is continuously improved by continuous learning from historical electricity load data from previous years. Highly accurate forecasting results are used to give effective feedback to the power sector, helping decision makers to develop reasonable power supply and production plans and reduce unnecessary losses and waste in the supply–consumption process.

3. Results

3.1. CISSA Algorithm Performance Testing and Analysis

3.1.1. Selection of the Test Function

In order to verify the effectiveness and stability of the CISSA, eight benchmark test functions were used to perform a comparative test of function optimization. The test functions and their specific information are shown in Table 2 below. F1–F5 denote unimodal testbench functions and F6–F8 denote multimodal testbench functions.

3.1.2. Experimental Comparison of the CISSA Algorithm with Other Intelligent Algorithms

The sparrow search algorithm, chaotic sparrow search algorithm, particle swarm optimization algorithm, grey wolf optimization algorithm, FA–CSSA algorithm proposed in [20] and CISSA algorithm proposed in this paper were selected for the test function finding comparison experiments. The common parameters of all algorithms were kept the same, the population size was set to 50, and the maximum number of iterations was set to 300. The relevant parameters of each algorithm are shown in Table 3 below.
In order to remove the error caused by chance, 30 independent trials were conducted for each of the 8 test functions. Table 4 shows the experimental results of the PSO, GWO, SSA, CSSA, FA–CSSA and the proposed CISSA algorithms after 30 independent runs on several standard test functions. The best values are marked in bold. In addition, the iterative convergence curves of the benchmark test functions are plotted to further visualize the convergence of each algorithm and the optimization results of the algorithms. The iterative convergence curves are shown in Figure 7. The different colored curves represent the convergence of the different algorithms. The horizontal axis represents the number of iterations and the vertical axis represents the fitness value.
The analysis in Table 4 shows that the CISSA algorithm outperformed the other three standard comparison algorithms and the two modified sparrow algorithms for the same test constraints for eight sets of test benchmark functions.
For the single-peak test functions F1–F5, the CISSA algorithm performed better than the PSO and GWO standard algorithms in the calculation of the mean and standard deviation. Even the FA–CSSA and CSSA algorithms achieved very good results for some of the test functions. In particular, the SSA, CSSA and FA–CSSA algorithms also achieved theoretical optima, but still did not outperform the CISSA algorithm in the calculation of the mean and standard deviation. This also demonstrates that for the single-peaked test functions F1–F5, CISSA not only achieved optimal results, but also showed a better convergence accuracy and stability.
For the multi-peaked functions F6–F8, the CISSA algorithm performed better than the PSO standard algorithm and the GWO standard algorithm in the calculation of the mean and standard deviation. In addition, the SSA, CSSA, FA–CSSA and CISSA algorithms all found their theoretical optimal values and performed very well in the calculation of the mean and standard deviation. The fact that the optimal solution was closer to the theoretical value proves that the CISSA algorithm is efficient in exploring the search space and guarantees a strong global search and local exploration capability.
In addition, compared to the SSA algorithm, the CISSA algorithm with the introduction of three improvement strategies had improved the optimal value, mean and standard deviation of the search results by several orders of magnitude to several tens of magnitude. The improvement of the SSA by a single strategy was limited, and the optimization results could not be maintained at a high level for different functions. The CISSA algorithm combining the three improved strategies showed a better overall solution performance, with most of the optimization accuracy and stability in 30 dimensions significantly better than the other four compared algorithms, which demonstrates the all-round improvement of the algorithm by combining multiple strategies.
From the iteration curves of the eight tested functions shown in Figure 7, it can be found that: for the F1 and F2 functions, the CISSA algorithm converged quickly and obtained the best fitness value, although the number of iterations to reach the optimal fitness was slightly greater than those of the other algorithms; for the F3, F4 and F5 functions, the CISSA algorithm was superior not only in terms of its convergence speed but also in terms of the number of iterations it required to achieve the optimal result; for the F6, F7 and F8 functions, the CISSA algorithm outperformed the other algorithms in terms of the number of iterations it required to reach the optimal value and the speed of convergence.
To further evaluate the performance of the CISSA algorithm, a Wilcoxon signed-rank test was performed at the α = 5 % significance level on the best results of the CISSA algorithm and the other five algorithms at 30 independent operations. The symbols “+”, “−” and “=” indicate that CISSA outperformed, underperformed and was equivalent to the comparison algorithm, respectively, and N/A indicates that the algorithms were close to each other and no significance could be determined. The results are shown in Table 5, where CISSA outperformed SSA in five of the eight benchmark functions, PSO in eight functions, GWO in eight functions, CSSA in five functions and FACSSA in five functions. Moreover, the p-values of CISSA were basically less than 0.05, indicating that the superiority of CISSA was statistically significant.
The CISSA algorithm performed the best, which proves the superiority of the algorithm itself. The mean and standard deviation of CISSA were smaller than other algorithms in the process of multiple search for both single-peaked and multi-peaked functions, which shows that the stability and robustness of CISSA is significantly better than other algorithms. The improved sparrow search algorithm was also able to explore the search space sufficiently and efficiently, and ensure a strong global search capability and local exploration capability. For different types of test functions, CISSA required the lowest number of iterations and the fastest iteration speed when several algorithms converged to the optimum, demonstrating the superiority of the convergence performance of the CISSA algorithm.

3.2. Electricity Load Forecasting Simulation Experiment

3.2.1. Simulation Experimental Data

For this paper, the real historical load data from 30 April 2007 to 12 September 2007 in a region of Shandong were selected as the simulation data. The data were collected every 0.5 h for a total of 20 weeks of historical data. As can be seen in Figure 8, the raw data series fluctuated considerably and was generally consistent with the “peak and trough” characteristics of the electricity load. We divided the dataset into seven datasets, Monday, Tuesday, Wednesday, Thursday, Friday, Saturday and Sunday. The 20 weeks of data are divided into seven datasets according to Monday to Sunday. The data from the first 19 weeks were used for training and the data from the last week were predicted. This division allowed the prediction of loads from a particular month to a precise day.

3.2.2. Evaluation Functions

For the performance evaluation of the prediction model, four common performance evaluation metrics were used: mean squared error (MSE), mean absolute error (MAE) and mean absolute percentage error (MAPE). Table 6 shows the mathematical models of the three evaluation functions.

3.2.3. Experimental Analysis of VMD Noise Reduction

From Figure 8, it can be seen that the historical load data had a strong volatility and non-linear characteristics. Therefore, it was necessary to use VMD for noise reduction of the load data. The parameters of the VMD were set as follows: a penalty factor of 2000, an initial center frequency of 0 and a convergence factor of 1 × 10−7. Figure 9 shows the decomposition of the load data to obtain a number of IMF and Res components. From Figure 9, it can be observed that the decomposed sequence was regular and had a certain periodicity. The frequencies were relatively stable and there was no obvious spectrum aliasing.
The training set of each component sequence after VMD decomposition was normalized and input into the CISSA–LSSVM power load forecasting model, and the final forecasting results were obtained by summing. The smaller the value of the MAPE as a function of the fitness was, the higher the accuracy of the prediction model was. Table 7 shows the MAPE values of the VMD–CISSA–LSSVM model compared with the CISSA–LSSVM model without VMD noise reduction.
The overall prediction performance of VMD–CISSA–LSSVMD was superior compared to the CISSA–LSSVM model without the VMD noise reduction process. The MAPE values for the seven subsets of data from Monday to Sunday were reduced by 0.1284%, 0.1479%, 0.1275%, 0.1885%, 0.0989%, 0.081% and 0.1088%, respectively. Simulation experiments verify that VMD could effectively improve the prediction accuracy of the model.

3.2.4. Analysis of Electrical Load Simulation Results

To further analyze the forecasting performance of the proposed VMD–CISSA–LSSVM power load forecasting model, three common independent forecasting models, three combined models based on the sparrow standard algorithm and three efficient combined models based on the improved sparrow standard algorithm were used as competitive models to participate in the forecasting performance comparison simulation experiments.
The three common prediction models were the LSSVM model, the ELM model and the Elman model. The three combined models based on the underlying optimization algorithm were the SSA–LSSVM model, GWO–ELMAN model and PSO–ELM model. The three efficient combinatorial models based on the improved sparrow standard algorithm were the FA–CSSA–ELM model proposed in [20], the CISSA1–LSSVM model and the CISSA2–LSSVM model. CISSA1–LSSVM represents the SSA-algorithm-optimized LSSVM model based on the Levy flight strategy and tent chaos mapping improvement. CISSA2–LSSVM represents the SSA algorithm optimized LSSVM based on tent chaos mapping and random following strategy improvement.
Figure 10 clearly shows the prediction results and trends of VMD–CISSA–LSSVM with three independent prediction models and six combined prediction models. The different colored curves represent the prediction results and trends of the different prediction models. For more details, please refer to Appendix A.
From the analysis of the prediction performance and trends of the different competing models shown in Figure 10, the prediction results of VMD–CISSA–LSSVM were more accurate and the prediction results were very stable and very close to the trend of the real load data.
Second, the FA–CSSA–ELM model, CISSA1–LSSVM model and CISSA2–LSSVM model performed relatively well. However, the above models also showed large fluctuations, resulting in unstable prediction results. The three independent forecasting models had the worst performance. In particular, the Elman model showed the most dramatic fluctuations and was the furthest from the real load data.
In addition, the predictive performance of the different competing models was analyzed more closely according to the four evaluation functions described above. Table 8 shows the performance of the evaluation functions of the different competing models for the seven data subsets. In particular, the smaller the evaluation value was, the better the prediction performance the model was. Based on the values of the three evaluation functions presented in Table 8, the corresponding bar graphs are plotted for a more visual analysis, as shown in Figure 11.
From the analysis of the data presented in Table 8 and Figure 10 and Figure 11, we can draw the following conclusions: compared to the LSSVM, ELM and ELMAN independent forecasting models, the average MSE values of the VMD–CISSA–LSSVM model were reduced by 69.8%, 87.8% and 86.7%, respectively; the average MAPE values were reduced by 66.1%, 89.0% and 87.6%, respectively; and the average MAE values were reduced by 69.2%, 87.6% and 86.5%, respectively. This also demonstrates the inability of independent forecasting models to achieve a forecasting accuracy that meets modern requirements.
Compared to independent forecasting models, combined forecasting models provided more accurate forecasts and trends that were closer to the true historical load. Compared with the SSA–LSSVM, PSO–Elman and GWO–ELM models, the average MSE values of the VMD–CISSA–LSSVM model were reduced by 62.3%, 61.7% and 60.9%, respectively; the average MAPE values were reduced by 59.7%, 58.2% and 60.1%, respectively; and the average MAE values were decreased by 59.1%, 58.0% and 60.2%, respectively.
The combined model based on the underlying optimization algorithm also achieved good prediction results. However, it still had the disadvantage of a low prediction accuracy. Compared with the CISSA1–LSSVM, FA–CSSA–ELM and CISSA2–LSSVM models, the average MSE values of the VMD–CISSA–LSSVM model decreased by 25.3%, 36.3% and 54.8%, respectively; the average MAPE values decreased by 13.7%, 34.4% and 54.4%, respectively; and the average MAE values decreased by 12.0%, 32.8% and 53.7%, respectively.
In summary, the combined VMD–CISSA–LSSVM prediction model proposed in this paper had the most outstanding prediction performance and was the most stable model; it could follow the trend of historical load data very well. In addition, the combined VMD–CISSA–LSSVM model had the lowest MSE evaluation value, indicating that the predicted and true values of the model had the lowest deviation. Having had the smallest MAE evaluation value indicates that the model had the smallest error in the prediction value. With the smallest MAPE evaluation values, the combined VMD–CISSA–LSSVM model was the most outstanding.

4. Conclusions

The innovations of this paper can be summarized as follows:
  • A detailed analysis of the iterative optimization search process of the SSA algorithm is presented. The CISSA algorithm is proposed to address the drawbacks of an uneven initial population distribution, the ease of falling into a local optimum and the slow convergence of the sparrow search algorithm. The CISSA algorithm uses the improved tent mapping strategy in the initialization phase of the population. The random following strategy taken from the chicken flock optimization algorithm idea is used in the iteration stage of the algorithm. The Levy flight strategy from the cuckoo algorithm idea is used in the global optimization phase for improvement;
  • The experimental results of the eight benchmark functions prove that the improved strategies are collaborative and complementary. The CISSA algorithm had a better convergence performance and overall synergy of the search and stability of the solution, and the overall performance was significantly improved compared with the SSA algorithm;
  • A new VMD–CISSA–LSSVM model for power load forecasting is proposed. The load forecasting simulation results demonstrate that the VMD–CISSA–LSSVM forecasting model had the highest forecasting accuracy and more stable forecasting results, and could follow the trend of historical load data very well. The numerical comparison with the four evaluation functions of competing models shows the superiority of the VMD–CISSA–LSSVM prediction model. Therefore, the VMD–CISSA–LSSVM can provide reasonable decision making and production guidance to the authorities.
The VMD–CISSA–LSSVM model had satisfactory prediction results. However, this paper does not take into account the effects of other factors such as temperature and holidays. In future studies, the effect of temperature, holidays and other factors on load forecasting will be taken into account.

Author Contributions

Conceptualization, X.W., G.W. and Z.W.; methodology, X.W., G.W. and Z.W.; software, X.W., G.W. and Z.W.; validation, X.W., C.M. and Z.S.; formal analysis, G.W., C.M. and Z.S.; investigation, X.W. and G.W.; resources, Z.W., G.W. and Z.S.; writing—original draft preparation, X.W. and Z.S; writing—review and editing, X.W., G.W. and Z.W.; visualization, X.W., C.M. and Z.S.; supervision, X.W. and Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science, Education, and Industry Integration Innovation Pilot Project of Qilu University of Technology (Shandong Academy of Sciences), grant number 2020KJC-ZD04; the Postgraduate Tutors’ Guidance Ability Improvement Project of Shandong Province, grant number SDYY17076; and the Empirical Research on Innovation of Cultivation Model of Control Graduate Students Based on System Synergy Theory, grant number SDYY18151.

Institutional Review Board Statement

The study did not involve humans or animals.

Informed Consent Statement

The study did not involve humans.

Data Availability Statement

As required by relevant policies, we are unable to provide more detailed raw data.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Figure A1. Final daily power load forecast values of different forecast models. (a) Prediction results of different prediction models in the Monday subset; (b) prediction results of different prediction models in the Tuesday subset; (c) prediction results of different prediction models in the Wednesday subset; (d) prediction results of different prediction models in the Thursday subset; (e) prediction results of different prediction models in the Friday subset; (f) prediction results of different prediction models in the Saturday subset; (g) prediction results of different prediction models in the Sunday subset.
Figure A1. Final daily power load forecast values of different forecast models. (a) Prediction results of different prediction models in the Monday subset; (b) prediction results of different prediction models in the Tuesday subset; (c) prediction results of different prediction models in the Wednesday subset; (d) prediction results of different prediction models in the Thursday subset; (e) prediction results of different prediction models in the Friday subset; (f) prediction results of different prediction models in the Saturday subset; (g) prediction results of different prediction models in the Sunday subset.
Mathematics 10 00028 g0a1aMathematics 10 00028 g0a1bMathematics 10 00028 g0a1cMathematics 10 00028 g0a1d

References

  1. Tan, M.; Yuan, S.; Li, S.; Su, Y.; Li, H.; He, F. Ultra-short-term industrial power demand forecasting using LSTM based hybrid ensemble learning. IEEE Trans. Power Syst. 2019, 35, 2937–2948. [Google Scholar] [CrossRef]
  2. Feng, C.; Sun, M.; Zhang, J. Reinforced deterministic and probabilistic load forecasting via Q-learning dynamic model selection. IEEE Trans. Smart Grid 2019, 11, 1377–1386. [Google Scholar] [CrossRef]
  3. Dudek, G.; Pełka, P.; Smyl, S.; Systems, L. A hybrid residual dilated LSTM and exponential smoothing model for midterm electric load forecasting. IEEE Trans. Neural Netw. Learn. Syst. 2021, 1–13. [Google Scholar] [CrossRef]
  4. Huang, Y.; Zhu, F.; Jia, G.; Zhang, Y.; Briefs, S.I.E. A slide window variational adaptive Kalman filter. IEEE Trans. Circ. Syst. II Express Briefs 2020, 67, 3552–3556. [Google Scholar] [CrossRef]
  5. Chung, K.-L.; Cheng, J.-S.; Yang, H.-B. Effective Chroma Subsampling and Luma Modification for RGB Full-Color Images Using the Multiple Linear Regression Technique. IEEE Access 2020, 8, 118315–118323. [Google Scholar] [CrossRef]
  6. Chu, Y.; Fei, J.; Hou, S. Adaptive global sliding-mode control for dynamic systems using double hidden layer recurrent neural network structure. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 1297–1309. [Google Scholar] [CrossRef] [PubMed]
  7. Rusek, K.; Suárez-Varela, J.; Almasan, P.; Barlet-Ros, P.; Cabellos-Aparicio, A. RouteNet: Leveraging Graph Neural Networks for network modeling and optimization in SDN. IEEE J. Sel. Areas Commun. 2020, 38, 2260–2270. [Google Scholar] [CrossRef]
  8. Alamdar, F.; Mohammadi, F.S.; Amiri, A.J.I.A. Twin bounded weighted relaxed support vector machines. IEEE Access 2019, 7, 22260–22275. [Google Scholar] [CrossRef]
  9. Avolio, M.; Fuduli, A.; Systems, L. A semiproximal support vector machine approach for binary multiple instance learning. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 3566–3577. [Google Scholar] [CrossRef]
  10. Khanesar, M.; Branson, D.T.J.I.S.J. Prediction Interval Identification Using Interval Type-2 Fuzzy Logic Systems: Lake Water Level Prediction Using Remote Sensing Data. IEEE Sensor J. 2021, 21, 13815–13827. [Google Scholar] [CrossRef]
  11. Yu, F.; Xu, X. A short-term load forecasting model of natural gas based on optimized genetic algorithm and improved BP neural network. Appl. Energy 2014, 134, 102–113. [Google Scholar] [CrossRef]
  12. Li, H.-Z.; Guo, S.; Li, C.-J.; Sun, J.-Q. A hybrid annual power load forecasting model based on generalized regression neural network with fruit fly optimization algorithm. Knowl. Based Syst. 2013, 37, 378–387. [Google Scholar] [CrossRef]
  13. Singh, P.; Dwivedi, P. Integration of new evolutionary approach with artificial neural network for solving short term load forecast problem. Appl. Energy 2018, 217, 537–549. [Google Scholar] [CrossRef]
  14. Bashir, Z.; El-Hawary, M. Applying wavelets to short-term load forecasting using PSO-based neural networks. IEEE Trans. Power Syst. 2009, 24, 20–27. [Google Scholar] [CrossRef]
  15. Ge, Q.; Guo, C.; Jiang, H.; Lu, Z.; Yao, G.; Zhang, J.; Hua, Q. Industrial power load forecasting method based on reinforcement learning and PSO-LSSVM. IEEE Trans. Cybern. 2020, 1–13. [Google Scholar] [CrossRef]
  16. Zhang, G.; Liu, H.; Li, P.; Li, M.; He, Q.; Chao, H.; Zhang, J.; Hou, J.J.C. Load prediction based on hybrid model of VMD-mRMR-BPNN-LSSVM. Complexity 2020, 2020, 6940786. [Google Scholar] [CrossRef] [Green Version]
  17. Zhang, H.; Yang, Y.; Zhang, Y.; He, Z.; Yuan, W.; Yang, Y.; Qiu, W.; Li, L. A combined model based on SSA, neural networks, and LSSVM for short-term electric load and price forecasting. Neural Comput. Appl. 2021, 33, 773–788. [Google Scholar] [CrossRef]
  18. Hao, J.; Zhu, C.; Guo, X. A New CIGWO-Elman Hybrid Model for Power Load Forecasting. J. Electr. Eng. Technol. 2021, 1–15. [Google Scholar] [CrossRef]
  19. Chen, X.-D.; Hai-Yue, Y.; Wun, J.-S.; Wu, C.-H.; Wang, C.-H.; Li, L.-L. Power load forecasting in energy system based on improved extreme learning machine. Energ. Explor. Exploit. 2020, 38, 1194–1211. [Google Scholar] [CrossRef]
  20. Wang, Z.; Wang, X.; Ma, C.; Song, Z. A Power Load Forecasting Model Based on FA-CSSA-ELM. Math. Probl. Eng. 2021, 2021, 2021. [Google Scholar] [CrossRef]
  21. Wu, H.; Wang, J. A Method for Prediction of Waterlogging Economic Losses in a Subway Station Project. Mathematics 2021, 9, 1421. [Google Scholar] [CrossRef]
  22. Liu, Q.; Zhang, Y.; Li, M.; Zhang, Z.; Cao, N.; Shang, J. Multi-UAV Path Planning Based on Fusion of Sparrow Search Algorithm and Improved Bioinspired Neural Network. IEEE Access 2021, 9, 124670–124681. [Google Scholar] [CrossRef]
  23. Tang, Y.; Li, C.; Li, S.; Cao, B.; Chen, C. A Fusion Crossover Mutation Sparrow Search Algorithm. Math. Probl. Eng. 2021, 2021, 9952606. [Google Scholar] [CrossRef]
  24. Chen, X.; Huang, X.; Zhu, D.; Qiu, Y. Research on chaotic flying sparrow search algorithm. In Proceedings of the Journal of Physics: Conference Series, Sanya, China, 29–31 January 2021; p. 012044. [Google Scholar]
  25. Yuan, J.; Zhao, Z.; Liu, Y.; He, B.; Wang, L.; Xie, B.; Gao, Y.J.I.A. DMPPT control of photovoltaic microgrid based on improved sparrow search algorithm. IEEE Access 2021, 9, 16623–16629. [Google Scholar] [CrossRef]
  26. Zhang, J.; Wei, Y.-M.; Li, D.; Tan, Z.; Zhou, J. Short term electricity load forecasting using a hybrid model. Energy 2018, 158, 774–781. [Google Scholar] [CrossRef]
  27. Wu, Z.; Zhao, X.; Ma, Y.; Zhao, X. A hybrid model based on modified multi-objective cuckoo search algorithm for short-term load forecasting. Appl. Energy 2019, 237, 896–909. [Google Scholar] [CrossRef]
  28. Dong, P.; Bin, X.; Jun, M. Short-term Load Forecasting Based on EEMD-approximate Entropy and ELM. In Proceedings of the 2019 IEEE Sustainable Power and Energy Conference (iSPEC), Beijing, China, 21 November 2019; pp. 1772–1775. [Google Scholar]
  29. Li, Y.; Wang, S.; Wei, Y.; Zhu, Q. A New Hybrid VMD-ICSS-BiGRU Approach for Gold Futures Price Forecasting and Algorithmic Trading. IEEE Trans. Comput. Soc. Syst. 2021, 8, 1357–1368. [Google Scholar] [CrossRef]
  30. Zhang, X.; Ge, Z. Local parameter optimization of LSSVM for industrial soft sensing with big data and cloud implementation. IEEE Trans Ind. Inf. 2019, 16, 2917–2928. [Google Scholar] [CrossRef]
  31. Xue, J.; Shen, B. A novel swarm intelligence optimization approach: Sparrow search algorithm. Syst. Sci. Control. Eng. 2020, 8, 22–34. [Google Scholar] [CrossRef]
  32. Guo, R. Projective synchronization of a class of chaotic systems by dynamic feedback control method. Nonlinear Dynam. 2017, 90, 53–64. [Google Scholar] [CrossRef]
  33. Guo, R.; Qi, Y. Partial anti-synchronization in a class of chaotic and hyper-chaotic systems. IEEE Access 2021, 9, 46303–46312. [Google Scholar] [CrossRef]
  34. Meng, X.; Liu, Y.; Gao, X.; Zhang, H. A new bio-inspired algorithm: Chicken swarm optimization. In Proceedings of the International Conference in Swarm Intelligence, Hefei, China, 17–20 October 2014; pp. 86–94. [Google Scholar]
  35. Cheng, Z.; Wang, J.; Zhang, M.; Song, H.; Chang, T.; Bi, Y.; Sun, K. Improvement and application of adaptive hybrid cuckoo search algorithm. IEEE Access 2019, 7, 145489–145515. [Google Scholar] [CrossRef]
Figure 1. Block diagram of the SSA algorithm operation.
Figure 1. Block diagram of the SSA algorithm operation.
Mathematics 10 00028 g001
Figure 2. Four common chaotic map bifurcation graphs: (a) is the bifurcation graph of the logistic chaotic map; (b) is the bifurcation graph of the tent chaotic map; (c) is the bifurcation graph of the sine chaotic map; and (d) is the bifurcation graph of the Hénon chaotic map.
Figure 2. Four common chaotic map bifurcation graphs: (a) is the bifurcation graph of the logistic chaotic map; (b) is the bifurcation graph of the tent chaotic map; (c) is the bifurcation graph of the sine chaotic map; and (d) is the bifurcation graph of the Hénon chaotic map.
Mathematics 10 00028 g002aMathematics 10 00028 g002b
Figure 3. Simulation of a Levy flight path in the 2D plane.
Figure 3. Simulation of a Levy flight path in the 2D plane.
Mathematics 10 00028 g003
Figure 4. Flowchart of the CISSA algorithm.
Figure 4. Flowchart of the CISSA algorithm.
Mathematics 10 00028 g004
Figure 5. VMD–CISSA–LSSVM electricity load forecasting flow chart.
Figure 5. VMD–CISSA–LSSVM electricity load forecasting flow chart.
Mathematics 10 00028 g005
Figure 6. Flowchart for the transmission of electrical loads in the transmission grid.
Figure 6. Flowchart for the transmission of electrical loads in the transmission grid.
Mathematics 10 00028 g006
Figure 7. Convergence curves for eight functions: (a) indicates the convergence curve for the F1 function; (b) indicates the convergence curve for the F2 function; (c) indicates the convergence curve for the F3 function; (d) indicates the convergence curve for the F4 function; (e) indicates the convergence curve for the F5 function; (f) indicates the convergence curve for the F6 function; (g) indicates the convergence curve for the F7 function; (h) indicates the convergence curve for the F8 function.
Figure 7. Convergence curves for eight functions: (a) indicates the convergence curve for the F1 function; (b) indicates the convergence curve for the F2 function; (c) indicates the convergence curve for the F3 function; (d) indicates the convergence curve for the F4 function; (e) indicates the convergence curve for the F5 function; (f) indicates the convergence curve for the F6 function; (g) indicates the convergence curve for the F7 function; (h) indicates the convergence curve for the F8 function.
Mathematics 10 00028 g007
Figure 8. The electricity load dataset.
Figure 8. The electricity load dataset.
Mathematics 10 00028 g008
Figure 9. The VMD decomposition provided the IMF component and the Res component.
Figure 9. The VMD decomposition provided the IMF component and the Res component.
Mathematics 10 00028 g009
Figure 10. Final daily power load forecast value of different forecast models. (a) Prediction results of different prediction models in the Monday subset; (b) prediction results of different prediction models in the Tuesday subset; (c) prediction results of different prediction models in the Wednesday subset; (d) prediction results of different prediction models in the Thursday subset; (e) prediction results of different prediction models in the Friday subset; (f) prediction results of different prediction models in the Saturday subset; (g) prediction results of different prediction models in the Sunday subset.
Figure 10. Final daily power load forecast value of different forecast models. (a) Prediction results of different prediction models in the Monday subset; (b) prediction results of different prediction models in the Tuesday subset; (c) prediction results of different prediction models in the Wednesday subset; (d) prediction results of different prediction models in the Thursday subset; (e) prediction results of different prediction models in the Friday subset; (f) prediction results of different prediction models in the Saturday subset; (g) prediction results of different prediction models in the Sunday subset.
Mathematics 10 00028 g010
Figure 11. Statistics of evaluation indexes of different power load competition models. (a) Comparison of MAE values for different competition forecasting models; (b) comparison of MAPE values for different competition forecasting models; (c) comparison of MSE values for different competition forecasting models.
Figure 11. Statistics of evaluation indexes of different power load competition models. (a) Comparison of MAE values for different competition forecasting models; (b) comparison of MAPE values for different competition forecasting models; (c) comparison of MSE values for different competition forecasting models.
Mathematics 10 00028 g011
Table 1. A comprehensive analysis of the relevant literature.
Table 1. A comprehensive analysis of the relevant literature.
LiteratureTimeFieldMain ContributionsEnhancement Requirement/Disadvantage
[11]2014Short-term gas loadA genetic algorithm optimized by chaotic cat map enhances the global search ability.GA algorithm optimization strategies and the choice of data pre-processing still have to be improved.
[12]2013Power load Application of the FOA algorithm and the GRNN model to the field of power load forecastingThe structural shortcomings of the FOA algorithm itself and the influence of noise on the prediction results were not taken into account.
[13]2018Power load A new intelligent optimization algorithm (Follow the Leader, FTL) based on flock movementThe structural shortcomings of the FTL algorithm itself and the influence of the raw data on the prediction results were not taken into account.
[14]2009Power loadIntroducing intelligent optimization algorithms to the field of structural optimization of neural networksThe paper demonstrated the effectiveness of intelligent optimization algorithms for training neural networks, failing to take into account the shortcomings of the intelligent algorithms themselves.
[21]2021Economic loss It proved the superiority of the SSA algorithm and the LSSVM model.The authors only took into account the powerful global optimality-seeking capability of the SSA algorithm, ignoring the shortcomings of the algorithm itself in terms of its tendency to fall into local optima and convergence.
[15]2020Power loadThe authors proposed a prediction algorithm combining reinforcement learning particle swarm optimization and least squares support vector machines.Although the K-averaging algorithm was used to classify the production patterns of the raw load data, it failed to take into account the effect of noise on the prediction results.
[18]2021Power loadThe authors proposed a CIGWO-ELMAN electric load forecasting model. The GWO algorithm was optimized by introducing a chaotic mapping strategy and a cosine function strategy based on random distribution.The effect of noise in the raw load data on the prediction results was not taken into account.
[19]2020Power loadThe authors proposed an NCSOELM electric load forecasting model. The chicken flock algorithm was optimized by using a nonlinear dynamic inertia weighting strategy and levy variance strategy.The effect of noise in the raw load data on the prediction results was not taken into account.
literature [20]2021power loadThe authors proposed an FA–CSSA–ELM electricity load forecasting model. The sparrow algorithm was optimized using the firefly perturbation strategy and tent chaos mapping.The effect of noise in the raw load data on the prediction results was not taken into account.
[26]2018Power loadThe authors proposed a combined forecasting model based on improved empirical mode decomposition (IEMD), the autoregressive integrated moving average (ARIMA) and the FOA algorithm optimized wavelet neural network (WNN).No account was taken of the distortion caused by endpoint effects during the denoising process of the EMD algorithm.
[27]2019Power loadThe authors proposed a multi-strategy improved cuckoo algorithm based on the optimized GRNN model for electricity load forecasting and take into account seasonal factors.No account was taken of the distortion caused by endpoint effects during the denoising process of the EMD algorithm.
[28]2019Power loadThe authors proposed an electricity load forecasting model based on integrated empirical mode decomposition (EEMD), approximate entropy and the extreme learning machine (ELM).No account was taken of the distortion caused by endpoint effects during the denoising process of the EMD algorithm. The effect of parameter selection on the prediction results of neural network models was ignored.
Table 2. Test functions and their specific information.
Table 2. Test functions and their specific information.
TypeTest FunctionDimensionSectionMin
Unimodal testbench functions F 1 ( x ) = i = 1 n x i 2 30 [ 100 , 100 ] 0
F 2 ( x ) = i = 1 | x i | + i = 1 n | x i | 30 [ 10 , 10 ] 0
F 3 ( x ) = i = 1 ( j 1 i x j ) 2 30 [ 100 , 100 ] 0
F 4 ( x ) = max i { | x i | , 1 i n } 30 [ 100 , 100 ] 0
F 5 ( x ) = i = 1 n i x i 4 + r a n d o m [ 0 , 1 ) 30 [ 1.28 , 1.28 ] 0
Multimodal testbench functions F 6 ( x ) = i = 1 n [ x i 2 10 cos ( 2 π x i + 10 ) ] 30 [ 5.12 , 5.12 ] 0
F 7 ( x ) = 20 exp ( 0.2 1 n i = 1 n x i 2 ) exp ( 1 n i = 1 n cos ( 2 π x i ) ) + 20 + 30 [ 32 , 32 ] 0
F 8 ( x ) = π n { 10 sin ( π y 1 ) + i = 1 n ( y i 1 ) 2 [ 1 + 10 sin 2 ( π y i + 1 ) ] + ( y n 1 ) 2 } + i = 1 n u ( x i , 10 , 100 , 4 ) 30 [ 50 , 50 ] 0
Table 3. Parameter setting for each algorithm.
Table 3. Parameter setting for each algorithm.
AlgorithmParameters
SSApNum = 0.2 NP, sNum = 0.2 NP, R2 = 0.8
CSSApNum = 0.2 NP, sNum = 0.2 NP, R2 = 0.8
FA–CSSApNum = 0.2 NP, sNum = 0.2 NP, R2 = 0.8
CISSApNum = 0.2 NP, sNum = 0.2 NP, R2 = 0.8
PSOW = 0.729, C1 = 1.49445, C2 = 1.49445
GWOa decreased linearly from 2 to 0, r1, r2 ∈ [0, 1]
Table 4. Optimal results of different intelligent algorithms.
Table 4. Optimal results of different intelligent algorithms.
StatisticsAlgorithmF1F2F3F4F5F6F7F8
Average valuePSO9.21 × 1022.54 × 101.81 × 1031.42 × 1011.94 × 1002.14 × 1027.78 × 1008.92 × 100
GWO2.73 × 10−151.60 × 10−95.86 × 10−49.32 × 10−43.68 × 10−37.65 × 1001.40 × 10−84.83 × 10−3
SSA3.89 × 10−492.82 × 10−216.35 × 10−501.93 × 10−377.91 × 10−40.00 × 1008.88 × 10−160.00 × 100
CSSA3.10 × 10−741.13 × 10−435.32 × 10−1411.43 × 10−384.57 × 10−40.00 × 1008.88 × 10−160.00 × 100
FA–CSSA1.09 × 10−741.96 × 10−406.28 × 10−843.68 × 10−401.15 × 10−30.00 × 1008.88 × 10−160.00 × 100
CISSA2.93 × 10−1536.45 × 10−903.06 × 10−1526.77 × 10−703.16 × 10−40.00 × 1008.88 × 10−160.00 × 100
Standard deviationPSO2.95 × 1021.11 × 101.10 × 1043.28 × 1004.74 × 1002.90 × 1019.56 × 10−13.14 × 100
GWO2.75 × 10−151.00 × 10−96.11 × 10−27.52 × 10−42.36 × 10−35.63 × 1001.22 × 10−89.03 × 10−3
SSA2.12 × 10−481.55 × 10−203.48 × 10−491.06 × 10−365.02 × 10−40.00 × 1000.00 × 1000.00 × 100
CSSA1.57 × 10−736.16 × 10−432.91 × 10−1407.86 × 10−383.35 × 10−40.00 × 1000.00 × 1000.00 × 100
FA–CSSA5.98 × 10−741.07 × 10−393.44 × 10−832.01 × 10−398.21 × 10−40.00 × 10000.00 × 1000.00 × 100
CISSA1.61 × 10−1523.53 × 10−891.68 × 10−1513.71 × 10−692.30 × 10−40.00 × 1000.00 × 1000.00 × 100
Optimal fitness valuePSO4.58 × 1021.04 × 1018.93 × 1036.45 × 1006.44 × 10−21.61 × 1025.96 × 1003.83 × 100
GWO2.71 × 10−174.19 × 10−101.27 × 10−17.34 × 10−59.18 × 10−42.84 × 10−124.36 × 10−92.11 × 10−15
SSA0.00 × 1000.00 × 1000.00 × 1000.00 × 1008.58 × 10−50.00 × 1008.88 × 10−160.00 × 100
CSSA0.00 × 1000.00 × 1000.00 × 1000.00 × 1003.79 × 10−60.00 × 1008.88 × 10−168.88 × 10−16
FA–CSSA0.00 × 1000.00 × 1000.00 × 1000.00 × 1001.94 × 10−50.00 × 1008.88 × 10−160.00 × 100
CISSA0.00 × 1000.00 × 1000.00 × 1000.00 × 1002.46 × 10−50.00 × 1008.88 × 10−160.00 × 100
Table 5. Wilcoxon signed rank test p-value.
Table 5. Wilcoxon signed rank test p-value.
FunctionCISSA–SSACISSA–CSSACISSA–PSOCISSA–GWOCISSA–FACSSA
F12.9657 × 10−61.1457 × 10−47.8787 × 10−127.8787 × 10−126.3629 × 10−6
F25.6522 × 10−60.03372.9155 × 10−112.9155 × 10−116.6434 × 10−6
F32.9643 × 10−54.8701 × 10−56.4789 × 10−126.4789 × 10−124.4498 × 10−4
F44.5993 × 10−43.8739 × 10−52.9155 × 10−112.9155 × 10−118.5561 × 10−4
F52.0023 × 10−60.01703.01990 × 10−113.01990 × 10−111.1937 × 10−6
F6N/AN/A1.2118 × 10−121.2118 × 10−12N/A
F7N/AN/A1.2118 × 10−121.2118 × 10−12N/A
F8N/AN/A1.2118 × 10−121.2118 × 10−12N/A
+/=/−5/3/05/3/08/0/08/0/05/3/0
Table 6. Evaluation functions for three prediction models.
Table 6. Evaluation functions for three prediction models.
MetricsMathematical Model
MAPE M A P E = 100 % N i N | t i p i t i | ,
MSE M S E = 1 N i = 1 N ( t i p i ) 2
MAE M A E = 1 N i = 1 N | t i p i | ,
where p i represents the predicted data and t i represents the real data.
Table 7. Comparison of VMD–CISSA–LSSVM’s and CISSA–LSSVM’s predicted performance (MAPE).
Table 7. Comparison of VMD–CISSA–LSSVM’s and CISSA–LSSVM’s predicted performance (MAPE).
MondayTuesdayWednesdayThursdayFridaySaturdaySunday
CISSA–LSSVM0.74121.14820.93751.0610.68000.66890.9030
VMD–CISSA–LSSVM0.61281.00030.81000.87250.58110.58790.7942
Table 8. Evaluation index statistics of different power load competition models.
Table 8. Evaluation index statistics of different power load competition models.
DayVMD–CISSA–LSSVMCISSA1–LSSVMFA–CSSA–ELMCISSA2–LSSVMSSA–LSSVMPSO–ElmanGWO–ELMLSSVMELMELMAN
MSEMonday10.131320.616617.662818.928323.238723.84317.434341.914732.770958.157
Tuesday15.228218.873445.074149.464250.808751.526456.310435.9995140.8367128.2496
Wednesday14.307912.358416.310917.513318.489422.351021.444435.895464.191064.1859
Thursday12.752317.946614.623924.506728.276430.179928.639140.775573.646973.6273
Friday8.728811.167311.381630.546318.632318.655427.329841.8471109.4490107.0738
Saturday7.959313.040310.648415.928743.292234.188135.300440.0823118.058577.3289
Sunday14.285817.663015.356827.645938.900237.203827.283840.0775148.8233122.6396
Average11.913315.952218.722626.361931.662531.135330.534639.513198.253790.1803
MAPEMonday0.61281.00740.985441.08551.34011.23651.01652.19991.92973.3479
Tuesday1.00031.08612.7823.46323.2663.26374.00542.10729.6628.6117
Wednesday0.81000.76591.03561.15941.09811.19001.32252.0714.74214.7416
Thursday0.87251.01021.06471.44011.57441.56281.5592.44195.30735.3055
Friday0.58110.63890.68322.04160.971490.97461.85422.35157.81737.6624
Saturday0.58790.71710.63780.816532.65181.95171.84842.17418.02964.4074
Sunday0.79420.86960.82921.52712.17562.41691.58692.186410.5088.5181
Average0.75120.8707 1.1454 1.6476 1.8682 1.7995 1.8847 2.2189 6.8566 6.0849
MAEMonday61.543397.245194.5848103.8166127.4083131.331997.1697216.1774187.8975333.8536
Tuesday92.790995.6586241.5351303.5821281.4374279.6618357.8988193.4625880.3002784.9639
Wednesday59.555365.074690.1768100.674594.9683103.3297116.1901180.7395411.6198411.5823
Thursday76.961882.838687.0163125.3825132.0139133.5128136.6856204.1656448.5094448.3609
Friday58.052158.154462.7472182.625188.666288.9332166.5054224.5649717.2867703.4552
Saturday57.898364.314159.207375.3412241.1399180.5305175.6953205.9746743.8573419.7419
Sunday69.101377.752973.3964137.6229200.0022216.5977146.0501205.1353955.7187776.5759
Average67.9861 77.2912 101.2377 147.0064 166.5195 161.9854 170.8851 204.3171 620.7414 554.0762
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, G.; Wang, X.; Wang, Z.; Ma, C.; Song, Z. A VMD–CISSA–LSSVM Based Electricity Load Forecasting Model. Mathematics 2022, 10, 28. https://doi.org/10.3390/math10010028

AMA Style

Wang G, Wang X, Wang Z, Ma C, Song Z. A VMD–CISSA–LSSVM Based Electricity Load Forecasting Model. Mathematics. 2022; 10(1):28. https://doi.org/10.3390/math10010028

Chicago/Turabian Style

Wang, Guijuan, Xinheng Wang, Zuoxun Wang, Chunrui Ma, and Zengxu Song. 2022. "A VMD–CISSA–LSSVM Based Electricity Load Forecasting Model" Mathematics 10, no. 1: 28. https://doi.org/10.3390/math10010028

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop