Next Article in Journal
Performance Investigation of Single–Piston Free Piston Expander–Linear Generator with Multi–Parameter Based on Simulation Model
Previous Article in Journal
Energy Disaggregation Using Multi-Objective Genetic Algorithm Designed Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on Fault Early Warning of Wind Turbine Based on IPSO-DBN

College of Electronic Information Engineering, Hebei University, Baoding 071002, China
*
Author to whom correspondence should be addressed.
Energies 2022, 15(23), 9072; https://doi.org/10.3390/en15239072
Submission received: 11 November 2022 / Revised: 24 November 2022 / Accepted: 29 November 2022 / Published: 30 November 2022

Abstract

:
Aiming at the problem of wind turbine generator fault early warning, a wind turbine fault early warning method based on nonlinear decreasing inertia weight and exponential change learning factor particle swarm optimization is proposed to optimize the deep belief network (DBN). With the data of wind farm supervisory control and data acquisition (SCADA) as input, the weights and biases of the network are pre-trained layer by layer. Then the BP neural network is used to fine-tune the parameters of the whole network. The improved particle swarm optimization algorithm (IPSO) is used to determine the number of neurons in the hidden layer of the model, pre-training learning rate, reverse fine-tuning learning rate, pre-training times and reverse fine-tuning training times and other parameters, and the DBN predictive regression model is established. The experimental results show that the proposed model has better performance in accuracy, training time and nonlinear fitting ability than the DBN model and PSO-DBN model.

1. Introduction

In recent years, the wind power industry has developed rapidly, and the installed scale of wind power continues to expand [1,2,3]. According to the statistics of the global wind energy council (GWEC), the total installed capacity of global wind power in 2020 is 743 GW, among which China ranks first in the world [4]. There are also great challenges behind the rapid development of the wind power industry. With the increase of the running time of wind turbines and the long-term working in the natural environment of changeable forces, difficult conditions and complex working conditions, the failure rate of main components such as generator, gearbox, main bearing, yaw system and so on increased significantly [5]. Generator failure is one of the main causes of wind turbine outage [6]. The use of accurate detection technology for early fault early warning of generators can effectively reduce the failure rate of wind turbines and increase the benefits of wind farms.
Wind turbine is a system with strong internal correlation, the components influence each other, and the fault types are varied [7,8]. At present, the fault early warning methods of wind turbine mainly focus on the parameters such as temperature signal, acoustic emission signal, vibration signal, electrical signal and so on [9]. In reference [10], the generator temperature is modified by linear regression technology, and the generator condition monitoring based on temperature signal is realized. In reference [11], the spatial neighborhood coefficient of data-driven threshold is used to avoid misdiagnosis caused by strong noise, and on this basis, an empirical wavelet transform generator bearing fault diagnosis method based on measured vibration signal is proposed. In reference [12], the minimum entropy deconvolution method and envelope analysis method are combined to realize the analysis and diagnosis of electric corrosion of generator bearings by using vibration signals.
At present, with the great advantage of deep mining data abstract features and inherent rules, the deep learning algorithm has made remarkable achievements in the fields of speech recognition [13], image classification [14] and target detection [15]. Because it can solve the problem of insufficient feature extraction ability of traditional diagnosis methods, deep learning algorithm has been applied to the field of fault diagnosis [16]. In reference [17], an improved deep learning model which combines SAE and depth neural network is proposed to realize the on-line condition monitoring of generator. In reference [18], the early fault early warning of wind turbine gearbox is realized by using depth variational self-coding network.
Wang and Liu [19] proposed a data-driven wind turbine condition monitoring method based on a new multivariable state estimation technology (Multivariate State Estimation Technique, MSET), which realizes the fault early warning of wavelet transform components. Su et al. [20] proposed a wind turbine state modeling and fault identification method based on a combined model on the basis of BP neural network and nonlinear state estimation method. Sun et al. [21] established a hidden Markov state monitoring model by using the characteristic variables of the supercapacitor of the pitch control system.
To summarize, the current mainstream fault early warning methods usually have some defects, such as low sensitivity and accuracy, small predictable time range, poor real-time performance and so on. When selecting state variables, most machine learning methods rely too much on expert experience and are highly subjective. Usually, the parameters are independent of each other by default, but in practice, the parameters of the unit are often related to each other, which leads to the introduction of some redundant or useless quantities with the prediction object when determining the input parameters, which affects the results.
In view of this, aiming at the problem of wind turbine fault early warning, an IPSO -DBN wind turbine fault early warning method is proposed in this paper. Firstly, multiple restricted Boltzmann machines (RBM) and one output layer are connected to form a depth confidence network. The unsupervised greedy layer-by-layer algorithm is used to pre-train the weights and biases of the network to generate better network parameters, and then the BP back propagation algorithm is used to fine-tune the whole network. For the problems that it is difficult to determine the parameters such as the number of neurons in the hidden layer of the DBN network, the pre-training learning rate, the reverse fine-tuning learning rate, the pre-training times and the reverse fine-tuning training times, the IPSO algorithm is used to optimize it intelligently. When the generator deviates from the normal workspace, the reconstruction error sequence changes obviously and deviates seriously from the original dynamic stable state, based on which the early fault detection of the generator is realized.

2. Particle Swarm Optimization Algorithm

2.1. Basic Particle Swarm Optimization Algorithm

Particle swarm optimization (PSO) is an intelligent population optimization method to find the global optimal solution [22]. At present, it has been widely used in intelligent control [23], path planning [24], neural network training [25], image processing [26], feature selection [27] and other fields [28]. PSO randomly generates a batch of particles in the search space, which have no volume and mass, only position and speed, and update their positions according to their own memory and the information of other particles in the group. Through the competition and cooperation between particles, we continue to move closer to the historical optimal position of individuals and populations, so as to achieve the purpose of finding the optimal solution in the complex solution space. The velocity and position update formula of the particle is shown as follows [19].
{ v i d k + 1 = ω v i d k + c 1 r 1 ( p i d x i d k ) + c 2 r 2 ( p g d x i d k ) x i d k + 1 = x i d k + v i d k + 1
where v i d k is the optimal solutions of the d-dimensional speed of the i particle in the k-th iteration. x i d k is the optimal solutions of the d-dimensional position of the i particle in the k-th iteration. pid is the optimal solutions of the d-dimensional individual of the i particle in the k-th iteration. pgd is the optimal solutions of the d-dimensional population history of the i particle in the k-th iteration. v i d k + 1 and x i d k + 1 are the speed and position of the d-dimensional of the i-particle in the k + 1 iteration, respectively. ω is the inertia weight. c1 and c2 are the individual learning factor and population learning factor, respectively. r1 and r2 are a random number in the interval [0, 1].

2.2. Improved Particle Swarm Optimization Algorithm

When finding the global optimal solution of general problems, the basic PSO algorithm has a good performance, but for complex high-dimensional optimal problems, it is easy to fall into local convergence, premature convergence or even non-convergence [29,30,31]. In order to solve these problems, an improved particle swarm optimization algorithm with nonlinear decreasing inertia weight and exponential learning factor is proposed in this paper.
Inertia weight ω is a very important parameter in PSO, which controls the local and global optimization ability of PSO, and determines the optimization effect of the algorithm to a great extent. When ω is large, the global optimization ability is strong, the local optimization ability is weak, and the convergence speed is fast; when ω is small, it is just the opposite. Therefore, in order to ensure the global optimization of the population with larger inertia weights in the early stage of the iteration, the refined optimization is realized with smaller inertia weights at the end of the iteration, and the nonlinear inertia weight strategy and control factors are introduced, and the inertia weights are shown in Equation (2).
ω ( t ) = ω m a x ω m i n 1 + ( K t T ) + ω m i n
where ω(t) is the inertia weight of the t iteration. ωmax and ωmin are the maximum and minimum inertia weights, respectively. T is the maximum number of iterations. K is the weight control factor.
Learning factor is also an important parameter in PSO. The larger the individual learning factor, the stronger the global search ability; the smaller the individual learning factor, the stronger the local search ability; and the population learning factor has the opposite attribute. At the beginning of the iteration, the individual learning factor should be larger and the population learning factor should be smaller; with the continuous increase of the number of iterations, the individual learning factor of particles gradually decreases and the group learning factor increases gradually. The use of exponential learning factors can well meet this requirement, and the updated formula is shown in Equation (3).
{ c 1 = c 1 m i n + ( c 1 m a x c 1 m i n ) e 20 ( t T ) 6 c 2 = c 2 m a x ( c 2 m a x c 2 m i n ) e 20 ( t T ) 6
where c1min and c2min are the minimum of individual learning factor and population learning factor, respectively. c1max and c2max are the maximum of individual learning factor and population learning factor, respectively.
Four performance test functions such as Rastrigin, Griewank, Acklet and Rosenbrock are used to analyze and verify the performance of IPSO, and compared with the basic PSO. The maximum number of iterations T is set to 300, and other parameters are shown in Table 1.
The optimal convergence ability of PSO and IPSO in the four performance test functions is shown in Figure 1. The value of x-coordinate is the number of iterations, and the value of y-coordinate is the fitness value (the average of 20 calculations).
As can be seen from Figure 1, compared with PSO, IPSO has faster convergence speed, higher accuracy and better stability.

3. Deep Belief Network

The deep belief network proposed by Geoffrey Hinton et al. [32] overcomes the problems of traditional neural network, such as low training efficiency, lack of small sample processing ability, easy to fall into local optimal solution and so on. The structure of the DBN network is shown in Figure 2. DBN is a deep Bayesian probability generation model composed of the input layer, multiple RBM and output layer [33]. The training process of the DBN model includes two parts: pre-training and parameter tuning [34]. First of all, the RBM is pre-trained layer by layer by unsupervised greed to generate better initial network parameters. Then the BP neural network is used to reverse fine-tune the network parameters to make the model optimal.

3.1. Pre-Training of the DBN Model

The pre-training of the DBN model uses the CD-k algorithm to train the weights and biases of each RBM in the network layer by layer unsupervised [35]. RBM basic structure is shown in Figure 3. RBM consists of a visible layer and a hidden layer. There is no connection within the layer, and no full connection between layers [36].
RBM is an energy-based model. When the energy is minimum and the network reaches the best state, the construction of the network is actually the process of minimizing the energy function [37]. The definition of RBM energy function is shown in Equation (4).
E θ ( v , h ) = i = 1 n b i v i j = 1 m c j h j i = 1 n j = 1 m v i w i j h j
where θ = {wij, bi, cj}. vi is the i neuron state of the visible layer. hj is the j neuron state of the hidden layer. wij is the connection weight between the i neuron of the visible layer and the j neuron of the hidden layer. bi is the offset of the i neuron of the visible layer. cj is the bias of the j neuron of the hidden layer.
The state of neurons in the RBM network can be described by statistical probability. According to Equation (4), the joint probability distribution of state (v,h) can be obtained, as shown in Equation (5).
P θ ( v , h ) = 1 Z θ e E θ ( v , h )
where Zθ is the distribution function, as shown in Equation (6).
Z θ = v , h e E θ ( v , h )
Visible layer element distribution Pθ(v) is the key to solve practical engineering problems, which corresponds to the edge distribution of Pθ(v,h). The specific definition is shown in Equation (7).
P θ ( v ) = h e E θ ( v , h ) v , h e E θ ( v , h )
Similar probability distributions with hidden layers:
P θ ( h ) = v e E θ ( v , h ) v , h e E θ ( v , h )
In order to make the distribution under the RBM representation fit the original sample distribution as much as possible, the K-L distance in the information theory is introduced. The smaller the K-L distance, the better the fitting effect, so the real edge distribution can be transformed into the maximum likelihood estimation of some approximate distribution.
In the RBM model, the likelihood function can be expressed as:
ln L ( θ ) = ln v P θ ( v ) = v ln P θ ( v )
To require maximum likelihood, we need to derive the likelihood function.
ln P θ ( v ) θ = ln h e E θ ( v , h ) θ ln v , h e E θ ( v , h ) θ       = 1 h e E θ ( v , h ) h e E θ ( v , h ) θ 1 v , h e E θ ( v , h ) v , h e E θ ( v , h ) θ       = h e E θ ( v , h ) ( E θ ( v , h ) θ ) h e E θ ( v , h ) v , h e E θ ( v , h ) ( E θ ( v , h ) θ ) v , h e E θ ( v , h )
Bringing in the expressions of P(v,h) and P(h|v), Equation (11) is obtained:
ln P θ ( v ) θ = h ( P θ ( h | v ) ( E θ ( v , h ) θ ) ) v , h ( P θ ( v , h ) ( E θ ( v , h ) θ ) )       = E P θ ( h | v ) ( E θ ( v , h ) θ ) E P θ ( v , h ) ( E θ ( v , h ) θ )
Replacing θ with w, b, c to derive respectively, and bringing into the E(v,h) expression, we can obtain:
ln P θ ( v ) w i j = P θ ( h i = 1 | v ) v j v P θ ( v ) P θ ( h j = 1 | v ) v j
ln P θ ( v ) b i = v i v P θ ( v ) v i
ln P θ ( v ) c i = P θ ( h i = 1 | v ) v P θ ( v ) P θ ( h i = 1 | v )
Because there is no connection between RBM layers, the activation functions of RBM visible layer neurons and hidden layer neurons can be expressed as follows:
P θ ( v i = 1 | h ) = f ( b i + i = 1 n w j , i v i )
P θ ( h j = 1 | v ) = f ( c j + j = 1 m w i , j h j )
f ( x ) = 1 1 + e x
where f(x) is the activation function.
The RBM parameters are updated by the CD-k algorithm, and the update criterion of θ = {wij, bi, cj} can be obtained by combining the neuron activation functions of visible layer and hidden layer, as shown in Equation (18).
{ w i j k + 1 = w i j k + ε { v i h j P θ ( h | v ) v i h j r e c o n } b i k + 1 = b i k + ε { v i P θ ( h | v ) v i r e c o n } c j k + 1 = c j k + ε { h j P θ ( h | v ) h j r e c o n }
where ε is the learning rate. 〈·〉(h|v) is the expectation of the partial derivative under the Pθ(h|v) distribution. 〈·〉recon is the expectation of the partial derivative under the reconstructed model distribution.

3.2. Fine Tuning of the DBN Model

After the unsupervised layer-by-layer training of the DBN network using the contrast divergence algorithm, a better initial weight and bias of the model are obtained. Because the RBM of each layer is trained separately, it can only ensure that the feature extraction of the RBM visible layer of each RBM output layer is optimal but cannot guarantee the optimal feature vector mapping of the whole DBN network. Therefore, it is necessary to use the BP back propagation algorithm to fine-tune the error between the output and input of the network from top to bottom to further optimize the weights and biases of the DBN network to optimize the parameters of the whole DBN.

4. Fault Early Warning Method of Wind Turbine Generator Based on IPSO-DBN

The fault early warning process of wind turbine generator based on IPSO-DBN is shown in Figure 4. As can be seen from Figure 4, it is mainly divided into: data input processing module, IPSO training module and IPSO-DBN fault early warning module.

4.1. Selection and Preprocessing of State Parameters

In order to establish the wind turbine generator fault early warning model, it is necessary to reasonably select the SCADA variable as the input of the model. The generator state parameters selected in this paper are shown in Table 2.
During the model training, the SCADA data parameters in Table 2 generated by the wind turbine generator in a certain time range are used as training samples to train the DBN network. The input vector x can be expressed as:
x = [ v o , P , Ω , T t , T a i , T b a , T b b , T u 1 , T v 1 , T w 1 , T s s i ]
Data pre-processing is a very important step in the process of fault early warning modeling, which mainly includes two parts: elimination of outliers and data normalization. In order to eliminate the influence of attributes and dimensions between different parameters of SCADA data, the data are normalized by Formula (20).
x i * = x i x m i n x m a x x m i n
where xi and x i * are the data before and after normalization. xmax and xmin are the maximum and minimum value of the original data, respectively.

4.2. Fitness Function

When using the IPSO algorithm to optimize the DBN parameters globally, the Equation (21) is taken as the fitness function.
f f i t n e s s = 1 n i = 1 n j = 1 m ( x ^ i j x i j ) 2
where n is the number of data samples. m is the dimension of each piece of data. xij and x ^ ij are the j-dimensional data of the i original sample and the j-dimensional data of the i reconstructed sample, respectively.

4.3. Reconstruction Error

In the normal working state, the SCADA data of the wind turbine maintain relatively stable internal characteristic rules, and the reconstruction error Re is relatively stable; when the generator deviates from the normal workspace, the internal characteristic rules of SCADA data change, the DBN model constructed by the normal state data does not match with the current state, and the Re trend changes obviously. Therefore, Re can be selected as the generator condition detection index. The Re calculation method is shown in Equation (22).
R e = x ^ x 2
where x ^ and x are the reconstructed data vector and the original data vector respectively.

4.4. Reconstruction Error Threshold

  • Fixed threshold
The fault of the wind turbine generator is judged by the reconstruction error threshold. The 3σ principle is introduced to set the threshold and calculate the mean value and root mean square error of Re.
{ μ = 1 N i = 1 N R e ( i ) σ = 1 N 1 i = 1 N ( R e ( i ) μ ) 2
where Re(i) is the i reconstruction error. μ and σ are the mean and root mean square errors of Re, respectively. N is the length of Re, that is, the number of reconstruction errors.
According to the 3σ principle, the reconstruction error threshold Uth is set to:
U th = μ + 3 σ
If the reconstruction error does not exceed the threshold, the running state of the generator is normal. If the reconstruction error exceeds the threshold, the running state of the generator is abnormal, and the fault early warning information is issued at the same time.
  • Adaptive threshold based on sliding window
In addition to the operation status of wind turbines, wind speed is also an important reason that affects the variation of SCADA parameters. Sometimes, even if the operation status of wind turbines is kept within the normal workspace, the trend of reconfiguration error will have a great sudden change due to the change of wind speed. At this time, if a constant threshold is set, some poles may exceed the threshold, thus misjudging the status of the unit. Therefore, in order to prevent the disadvantages caused by the fixed threshold, on the basis of the fixed threshold, the adaptive threshold based on sliding window is introduced as the condition to judge the occurrence of wind turbine failure.
Suppose that within a certain period of time, the reconstruction error order of some part of the unit is {Re(0), Re(1), ···, Re(N), ···,}. For this sequence, take a sliding window with N width, as shown in Figure 5. First, the mean and standard deviation of continuous N reconstruction error sequences in the window are calculated by using the Formula (23), and then the first fixed threshold is calculated according to the Formula (24), which is used as the threshold of the initial reconstruction error. Starting from the position of the first sliding window, the thresholds of subsequent reconstruction errors are obtained according to the same method, and these thresholds are concatenated to form an adaptive threshold based on sliding window.

4.5. Evaluation Index

The root mean square error (RMSE), the mean absolute error (MAE) and the mean absolute percent error (MAPE) of the reconstruction error are calculated respectively to quantitatively analyze the model, and the performance of different models is compared. The calculation method is as follows:
R M S E = 1 n i = 1 n j = 1 m ( x ^ i j x i j ) 2
M A E = 1 n i = 1 n j = 1 m | x ^ i j x i j |
M A P E = 1 n i = 1 n j = 1 m | x ^ i j x i j x ^ i j |
where n and m are the number and dimension of data samples respectively. xij is the j-th dimensional data of the i original sample. x ^ ij is the j-th dimensional data of the i reconstructed sample.
The smaller the value of the above three quantitative evaluation indicators, the higher the accuracy of the model reconstruction of the original data, and the better the network performance.

5. Example Analysis

5.1. Modeling Data Statement

This paper takes the E15 wind turbine generator in a wind farm as the research object and uses the SCADA data from 00:00 on 23 December 2018 to 23:50 on 23 April 2019, as the data source to carry on the wind turbine fault early warning analysis. The data source contains 69 kinds and 17,567 pieces of information-rich 10-min sampling frequency data.

5.2. Model Parameter Setting

The IPSO-DBN model consists of four hidden layers, the optimization interval of the number of neurons in the hidden layer is [1, 100]. The optimization interval of learning rate is [0, 0.1], and the optimization interval of training times is [1, 500], pre-training 100 data per batch and reverse fine-tuning 4 data per batch. The initial parameters of IPSO are as follows: the population size is 20, ωmin = 0.45, ωmax = 0.95, cmin = 0.45, cmax = 0.95. The maximum number of iterations is 40, and the weight control factor is 8. The optimization results are shown in Table 3.

5.3. Early Warning Analysis of Generator Fault

5.3.1. Comparison of Learning Effects of Different Models

The comparison results of evaluation indicators of different models are shown in Table 4.
It can be seen from Table 4 that the IPSO-DBN wind turbine fault early warning model proposed in this paper has the best performance in RMSE, MAE and MAPE, compared with the DBN and PSO-DBN models.
The learning effect of the three models is shown in Figure 6.
From Figure 6, the reconstruction error sequence of the IPSO-DBN model is the smallest, followed by the PSO-DBN model, and the reconstruction error of the DBN model is the largest. To sum up, the IPSO-DBN model proposed in this paper has stronger nonlinear fitting ability, stronger robustness and higher accuracy, and is more suitable for early fault early warning of wind turbines.

5.3.2. Early Warning Model Testing under Normal Condition

The first 9100 groups of data of the data sample are selected as the training set. Under the normal operation state of the wind turbine, the reconstruction error sequence of the generator operation data is shown in Figure 7, and the red line in the figure is the early warning threshold. Figure 7a is a fixed threshold, and Figure 7b is an adaptive threshold based on the principle of sliding window. In this paper, the length of sliding window is set to 300.
As can be seen from Figure 7b, under normal operating conditions, because the wind speed is always fluctuating, there are fluctuations and a few extreme points in the generator reconstruction error, but it is always within the threshold range. Based on the sliding window method, the threshold changes with the change of reconstruction error, and the threshold accuracy and effectiveness are higher.

5.3.3. Early Warning Model Testing in Fault State

The fault samples of the generator in the data source are input into the trained IPSO-DBN fault early warning model, the reconstruction value is the output and the reconstruction error is calculated, and the trend change diagram of the reconstruction error sequence is obtained. Then the above fixed threshold and adaptive threshold based on sliding window are used to verify and analyze the fault early warning function of the model. Finally, the trend of residual sequence of generator related variables is analyzed to judge the possible fault types of generator.
The following 3908 data samples of the training data set are inputted into the IPSO-DBN model, and the calculated reconstruction error sequence is shown in Figure 8. Figure 8a adopts a fixed early warning threshold and Figure 8b adopts an adaptive threshold based on sliding window. In the state of generator failure, the fixed threshold is set to 0.223, and the sliding window length of the adaptive threshold is set to 300.
It can be seen from Figure 8 that the reconstruction error sequence dynamically changes within the early warning threshold for a period of time before the fault occurs (about 3000 points), which proves that the generator does not deviate from the normal workspace during this period of time. On the other hand, the reconstruction error exceeds the early warning threshold at the ① point of Figure 8a and the X point of Figure 8b (which corresponds to the data sample point 3129 in the state of generator failure), and the fault early warning is triggered.
Combined with the change of reconstruction error sequence of generator variables, the fault early warning of generator is realized by setting the strategy of fixed threshold and adaptive threshold in this paper. The sampling period of SCADA data used in this paper is too long to detect the slow change of generator-related variables, which leads to a sudden and dramatic reconstruction error exceeding the early warning threshold in Figure 8. This problem can be overcome by increasing the adoption period. For the SCADA system in wind turbine field, the data sampling period can usually reach seconds, when it can be observed that the reconstruction error rises slowly some time before the fault occurs, and then exceeds the set threshold before causing a serious fault to achieve the purpose of detecting the potential fault of the wind turbine.

5.3.4. Residual Analysis of Each Variable

According to the change of the trend of the reconstruction error sequence and the set early warning threshold, the fault can be predicted, so that the staff can check the hidden trouble in time and prevent the fault from getting worse. In order to determine the specific fault location as soon as possible and improve the maintenance efficiency, the residual trend chart can be used to analyze the SCADA parameters of each component of the wind turbine to realize the preliminary judgment of the fault type and fault location. The trend of wind speed residual series and active power residual series in the event of wind turbine generator failure is shown in Figure 9.
As can be seen from Figure 9, the wind speed and power residual series of wind turbines always fluctuate slightly around zero in the early stage, and only fluctuate in a slightly larger range in the later stage. On the whole, the fluctuation of their residuals is always within the normal range, and the larger changes in the later period must have been affected by other variables. Basically, it can be judged that the wind speed is always in a steady change for a period of time before and after the generator failure. Therefore, the generator failure caused by the natural environment can be ruled out. There is a fault caused by the generator itself or by long-term operation.
The temperature residual change trend of generator stator winding u1, v1, w1 and bearing temperature is shown in Figure 10.
It can be seen from Figure 10a that the trend of temperature residual of generator stator winding is not obviously abnormal, so the possibility of failure of generator coil winding can also be ruled out. From Figure 10b, it can be seen that the residual sequence of generator bearing temperature a (that is, generator front shaft) deviates obviously and changes sharply, which proves the previous conjecture from the positive side. This is due to the long sampling period and no slow temperature change process. Combined with the fault records of the SCADA system, the report shows that the generator has the front axle overheating. It is proved that the model proposed in this paper is effective for wind turbine fault early warning.

6. Conclusions

Aiming at the problem of wind turbine fault early warning, the deep confidence network is used to mine the characteristic information of SCADA data layer by layer under the normal operation state of the generator and intelligently learn the internal relationship among the parameters, and a multivariable DBN reconfiguration network model of wind turbine is proposed. For the problem that it is difficult to determine the parameters of the DBN network, the improved particle swarm optimization algorithm can better solve this problem. The fault early warning of generator can be realized through the reconstruction error of the IPSO-DBN model and the residual variation of each input variable. Compared with the DBN wind turbine fault model and the PSO-DBN wind turbine fault model, this model has better performance in accuracy, training time and nonlinear fitting ability, so it can better meet the needs of practical engineering.

Author Contributions

Conceptualization, Z.Z. and S.W.; methodology, Z.Z.; software, P.J.; validation, S.W. and P.J.; formal analysis, S.W.; investigation, H.Z.; resources, P.W.; data curation, Z.Z.; writing—original draft preparation, S.W.; writing—review and editing, Z.Z.; visualization, H.Z.; supervision, P.J.; project administration, P.W.; funding acquisition, P.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was financially supported by National Natural Science Foundation of China, grant number 12171135; S&T Program of Hebei, grant number 216Z2103G; Natural Science Foundation of Hebei Province, grant number A2020201021; Hebei Province High-level Talent funding Project, grant number B2020005004; Baoding Science and Technology Plan Project, grant number 2163P036.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

Acronyms and abbreviations
DBNdeep belief network
IPSOimproved particle swarm optimization
PSOparticle swarm optimization
SCADAsupervisory control and data acquisition
Indices
iindex of particle
knumber of iteration
dnumber of dimensional
Variables
bioffset of the i neuron of the visible layer
vparticle speed
xparticle position
pidoptimal solutions of the d-dimensional individual of the i particle
pgdoptimal solutions of the d-dimensional population history of the i particle
ωinertia weight
c1individual learning factor
c2population learning factor
c1minminimum of individual learning factor
c2maxpopulation learning factor
c1minminimum of individual learning factor
c2maxpopulation learning factor
cjbias of the j neuron of the hidden layer
r1random number in the interval [0, 1]
r2random number in the interval [0, 1]
ωmaxmaximum inertia weights
ωminminimum inertia weights flow of node
Tmaximum number of iterations
Kweight control factor
EθRBM energy
vii neuron state of visible layer
hjj neuron state of the hidden layer
wijconnection weight between the i neuron of the visible layer and the j neuron of the hidden layer
Pθjoint probability distribution function
Zθdistribution function
L(θ)likelihood function
f(x)activation function
εlearning rate
nnumber of neurons
vowind speed
Pgenerator power
Ωgenerator speed
Ttgenerator torque
Taiair cooling temperature of generator
Tbatemperature of generator front axle A
Tbbtemperature of generator front axle B
Tu1generator winding u1 temperature
Tv1generator winding v1 temperature
Tw1generator winding w1 temperature
Tssiside temperature of spindle impeller
xidata before normalization
xIdata after normalization
Xmaxmaximum value of the original data
xminminimum value of the original data
ffitnessfitness function
nnumber of data samples
mdimension of data samples
xijj-dimensional data of the i original sample
x ^ ijj-dimensional data of the i reconstructed sample
Rereconstruction error
x ^ reconstructed data vector
Xoriginal data vector
μmean errors of Re
σroot mean square errors of Re
Nlength of Re
Uthreconstruction error threshold

References

  1. Fang, R.; Wu, M.; Guo, X.; Shang, R.; Shao, P. Identifying early defects of wind turbine based on SCADA data and dynamical network marker. Renew. Energy 2020, 154, 625–635. [Google Scholar]
  2. Qian, P.; Zhang, D.; Tian, X.; Si, Y.; Li, L. A novel wind turbine condition monitoring method based on cloud computing. Renew. Energy 2019, 135, 390–398. [Google Scholar] [CrossRef]
  3. Jin, Z.; Huo, Z.; Xu, C.; Guo, H.; Zhou, H. Fault diagnosis for wind turbine based on random forest and XGBoost. Renew. Energy Resour. 2021, 39, 353–358. [Google Scholar]
  4. Hanumant, P.J.; Anand, K.B.; Ravinder, K.; Ahmadi, M.H.; Assad, M.E.H. RAM analysis and availability optimization of thermal power plant water circulation system using PSO. Energy Rep. 2021, 7, 1133–1153. [Google Scholar]
  5. Ma, J.; Tang, B.; Han, Y. Design of networking condition monitoring and fault diagnosis for wind turbine transmission system. J. Chongqing Univ. 2015, 38, 37–44. [Google Scholar]
  6. Zhao, H.; Liu, H.; Liu, H.; Lin, J. Condition monitoring and fault diagnosis of wind turbine generator based on stacked autoencoder network. Autom. Electr. Power Syst. 2018, 42, 102–108. [Google Scholar]
  7. Yin, S.; Hou, G.; Hu, X.; Robu, V.; Flynn, D.; Barnes, M.; Keane, J.; Nenadic, G. Fault warning and identification of front bearing of wind turbine generator. Chin. J. Sci. Instrum. 2020, 41, 242–251. [Google Scholar]
  8. Li, D.; Mao, X.; Yin, P.; Wu, F. The rear bearing temperature of early warning method in wind turbine based on NSET. Sci. Technol. Eng. 2016, 16, 205–209. [Google Scholar]
  9. Wang, Z. Research on Modeling Method of Condition Monitoring of Wind Turbine. Ph.D. Thesis, North China Electric Power University, Beijing, China, 2018. [Google Scholar]
  10. Huo, J.; Tang, G.; Liu, D.; Zhang, M. A novel monitoring method of wind turbine generator real-time reliability based on tempeture signals. Renew. Energy Resour. 2016, 34, 408–412. [Google Scholar]
  11. Chen, J.; Pan, J.; Li, Z.; Zi, Y.; Chen, X. Generator bearing fault diagnosis for wind turbine via empirical wavelet transform using measured vibration signals. Renew. Energy 2016, 89, 80–92. [Google Scholar] [CrossRef]
  12. Jiang, R.; Teng, W.; Liu, X.; Tang, S. Diagnosis of electrical corrosion fault in wind turbine generator bearing based on vibration signal analysis. Electr. Power 2019, 52, 128–133. [Google Scholar]
  13. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar]
  14. Geng, J.; Jiang, W.; Deng, X. Multi-scale deep feature learning network with bilateral filtering for SAR image classification-ScienceDirect. ISPRS J. Photogramm. Remote Sens. 2020, 167, 201–213. [Google Scholar] [CrossRef]
  15. Jeon, M.H.; Lee, Y.; Shin, Y.S.; Jang, H.; Kim, A. Underwater object detection and pose estimation using deep learning. IFAC-PapersOnLine 2019, 52, 78–81. [Google Scholar] [CrossRef]
  16. Lei, J.; Liu, C.; Jiang, D. Fault diagnosis of wind turbine based on Long Short-term memory networks. Renew. Energy 2019, 133, 422–432. [Google Scholar] [CrossRef]
  17. Jin, X.; Xu, Z.; Sun, Y.; Shan, J. Online condition monitoring for wind turbines based on SCADA data analysis and sparse auto-encoder neural network. Acta Energ. Sol. Sin. 2021, 42, 321–328. [Google Scholar]
  18. Ren, J.; Tang, B.; Yong, B.; Deng, L.; Zhang, K. Wind turbine gearbox fault warning based on depth variational autoencoders network fusion SCADA data. Acta Energ. Sol. Sin. 2021, 42, 403–408. [Google Scholar]
  19. Chen, R.; Zhang, Y.; Liu, Q.; Liu, T. Optimal design of grid structure in a DC system based on a PSO algorithm. Power Syst. Prot. Control. 2021, 49, 165–172. [Google Scholar]
  20. Zhang, Y.; Wang, L.; Zhou, H.; Zhao, H. Design and application of particle swarm optimization based on competitive learning. Comput. Meas. Control. 2021, 29, 182–189. [Google Scholar]
  21. Liu, M.; Dong, M.; Jing, C. Scheduled completion learing based multi-objective particle swarm optimization algorithm. J. Comput. Appl. 2019, 39, 330–335. [Google Scholar]
  22. Wang, S.; Liu, G. A nonlinear dynamic adaptive inertial weight particle swarm optimization. Comput. Simul. 2021, 38, 249–253. [Google Scholar]
  23. Li, X.; Xiao, J.; Pan, Y.; Liu, S.; Deng, X. Research on leveling control system of fuzzy PID vehicle-mounted platform based on PSO. Mod. Manuf. Eng. 2021, 2, 58–65. [Google Scholar]
  24. Hui, S.L.; Fan, S.; Chin, C.; Chai, S.; Bose, N.; Kim, E. Constrained path planning of autonomous underwater vehicle using selectively-hybridized particle swarm optimization algorithms. IFAC-PapersOnLine 2019, 52, 315–322. [Google Scholar]
  25. Rui, J.; Zhang, H.; Zhang, D.; Han, F.; Guo, Q. Total organic carbon content prediction based on support-vector-regression machine with particle swarm optimization. J. Pet. Sci. Eng. 2019, 180, 699–706. [Google Scholar] [CrossRef]
  26. Zhu, L.; Shao, S.; Jing, W.; Liu, L. PSO-optimized-fast segmentation algorithm for 3D Otsu forest canopy image. J. Harbin Univ. Sci. Technol. 2019, 24, 128–133. [Google Scholar]
  27. Engelbrecht, A.P.; Grobler, J.; Langeveld, J. Set based particle swarm optimization for the feature selection problem. Eng. Appl. Artif. Intell. 2019, 85, 324–336. [Google Scholar] [CrossRef]
  28. Li, Q. Power big data anomaly detection method based on an improved PSO-PFCM clustering algorithm. Power Syst. Prot. Control. 2021, 49, 161–166. [Google Scholar]
  29. Li, Z.; Liang, C.; Wang, M. Short-term power generation output prediction based on a PSO-DBN neural network. Power Syst. Prot. Control. 2020, 48, 149–154. [Google Scholar]
  30. Hua, Y.; Wang, S.; Bai, G.; Li, B. An improved particle swarm optimization algorithm based on nonlinear decreasing inertial weights. J. Chongqing Technol. Bus. Univ. (Nat. Sci. Ed.) 2021, 38, 1–9. [Google Scholar]
  31. Sheng, H.; Yang, J.; Zhang, G.; Yang, Y.; Geng, H.; Lu, L. Optimal configuration of micro energy network based on improved PSO algorithm. Power Syst. Clean Energy 2021, 37, 23–31. [Google Scholar]
  32. Khosravi, A.; Nunes, R.O.; Assad, M.E.H.; Machado, L. Comparison of artificial intelligence methods in estimation of daily global solar radiation. J. Clean. Prod. 2018, 194, 342–358. [Google Scholar] [CrossRef]
  33. Dai, J. Research on Transformer State Assessment Based on Deep Learning. Ph.D. Thesis, Shanghai Jiao Tong University, Shanghai, China, 2018. [Google Scholar]
  34. Yang, W.; Yin, K.; Bao, Y.; Yi, X.; Xu, B. Fault types identification of power grid based on deep belief network. Electr. Power Eng. Technol. 2012, 40, 113–117. [Google Scholar]
  35. Ding, S.; Li, X.; Hang, J.; Wang, Y.J.; Wang, Q.J. Deep learning theory and its application to fault diagnosis of an electric machine. Power Syst. Prot. Control. 2020, 48, 172–187. [Google Scholar]
  36. Peng, B.S.; Xia, H.; Liu, Y.K.; Yang, B.; Guo, D.; Zhu, S.M. Research on intelligent fault diagnosis method for nuclear power plant based on correlation analysis and deep belief network. Prog. Nucl. Energy 2018, 108, 419–427. [Google Scholar] [CrossRef]
  37. Li, B.; Zeng, H. Application of improved deep belief network in electric spindle fault diagnosis. Mech. Sci. Technol. Aerosp. Eng. 2021, 40, 1051–1057. [Google Scholar]
Figure 1. PSO and IPSO optimization results. (a) Fitness value under Rastrigin test function. (b) Fitness value under Griewank test function. (c) Fitness value under Acklet test function. (d) Fitness value under Rosenbrock test function.
Figure 1. PSO and IPSO optimization results. (a) Fitness value under Rastrigin test function. (b) Fitness value under Griewank test function. (c) Fitness value under Acklet test function. (d) Fitness value under Rosenbrock test function.
Energies 15 09072 g001
Figure 2. DBN network structure.
Figure 2. DBN network structure.
Energies 15 09072 g002
Figure 3. RBM basic structure.
Figure 3. RBM basic structure.
Energies 15 09072 g003
Figure 4. Forecasting results between before and after rough set correction.
Figure 4. Forecasting results between before and after rough set correction.
Energies 15 09072 g004
Figure 5. Reconstruction error sliding window.
Figure 5. Reconstruction error sliding window.
Energies 15 09072 g005
Figure 6. Reconstruction error results of different models.
Figure 6. Reconstruction error results of different models.
Energies 15 09072 g006
Figure 7. Training set reconstruction error. (a) Fixed threshold. (b) Adaptive threshold based on sliding window.
Figure 7. Training set reconstruction error. (a) Fixed threshold. (b) Adaptive threshold based on sliding window.
Energies 15 09072 g007
Figure 8. Validation set reconstruction error. (a) Variation trend of residual wind speed. (b) Change trend of power residual.
Figure 8. Validation set reconstruction error. (a) Variation trend of residual wind speed. (b) Change trend of power residual.
Energies 15 09072 g008
Figure 9. Test set reconstruction error. (a) Fixed threshold. (b) Adaptive threshold based on sliding window.
Figure 9. Test set reconstruction error. (a) Fixed threshold. (b) Adaptive threshold based on sliding window.
Energies 15 09072 g009
Figure 10. Test set reconstruction error. (a) Variation trend of temperature residual of generator stator windings. (b) Variation trend of temperature residual of generator bearing.
Figure 10. Test set reconstruction error. (a) Variation trend of temperature residual of generator stator windings. (b) Variation trend of temperature residual of generator bearing.
Energies 15 09072 g010
Table 1. PSO and IPSO parameter setting.
Table 1. PSO and IPSO parameter setting.
ArithmeticPopulation Size (N)Dimension (D)ωc1c2
PSO100300.82.02.0
IPSO100300.9~0.42.5~0.50.5~2.5
Table 2. Description of generator condition variables.
Table 2. Description of generator condition variables.
Variable NameNumerical ValueUnit
Wind speedvom/s
Generator powerPkW
Generator speedΩr/min
Generator torqueTtN·m
Air cooling temperature of generatorTai°C
Temperature of generator front axle ATba°C
Temperature of generator front axle BTbb°C
Generator winding u1 temperatureTu1°C
Generator winding v1 temperatureTv1°C
Generator winding w1 temperatureTw1°C
Side temperature of spindle impellerTssi°C
Table 3. IPSO optimization results.
Table 3. IPSO optimization results.
ParametersNumerical Value
The number of neurons in the first hidden layer65
The number of neurons in the second hidden layer68
The number of neurons in the third hidden layer21
The number of neurons in the fourth hidden layer98
Pre-training learning rate0.0185
Reverse fine-tuning learning rate0.0456
Number of pre-training27
Number of reverse fine-tuning431
Table 4. Hidden layer neuron selection results.
Table 4. Hidden layer neuron selection results.
ModelRMSEMAEMAPE
DBN0.12560.26388.0291
PSO-DBN0.08070.168516.0129
IPSO-DBN0.06230.13154.9700
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, Z.; Wang, S.; Wang, P.; Jiang, P.; Zhou, H. Research on Fault Early Warning of Wind Turbine Based on IPSO-DBN. Energies 2022, 15, 9072. https://doi.org/10.3390/en15239072

AMA Style

Zhang Z, Wang S, Wang P, Jiang P, Zhou H. Research on Fault Early Warning of Wind Turbine Based on IPSO-DBN. Energies. 2022; 15(23):9072. https://doi.org/10.3390/en15239072

Chicago/Turabian Style

Zhang, Zhaoyan, Shaoke Wang, Peiguang Wang, Ping Jiang, and Hang Zhou. 2022. "Research on Fault Early Warning of Wind Turbine Based on IPSO-DBN" Energies 15, no. 23: 9072. https://doi.org/10.3390/en15239072

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop