Next Article in Journal
Prediction of the Comprehensive Error Field in the Machining Space of the Five-Axis Machine Tool Based on the “S”-Shaped Specimen Family
Previous Article in Journal
Modelling Particle Motion during High-Voltage Electrostatic Abrasive Implantation Based on Multi-Physics Field Coupling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Wind Turbine Fault Classification Model Using Broad Learning System Optimized by Improved Pelican Optimization Algorithm

1
College of Water Conservancy and HydroPower Engineering, HoHai University, Nanjing 210098, China
2
College of Hydraulic and Civil Engineering, Xinjiang Agricultural University, Urumqi 830052, China
3
College of Energy and Electrical Engineering, HoHai University, Nanjing 210098, China
4
School of Mechanical Engineering, Nanchang Institute of Technology, Nanchang 330099, China
*
Author to whom correspondence should be addressed.
Machines 2022, 10(5), 407; https://doi.org/10.3390/machines10050407
Submission received: 30 March 2022 / Revised: 12 May 2022 / Accepted: 17 May 2022 / Published: 23 May 2022
(This article belongs to the Section Turbomachinery)

Abstract

:
As a classification model, a broad learning system is widely used in wind turbine fault diagnosis. However, the setting of hyperparameters for the models directly affects the classification accuracy of the models and it generally relies on practical experience and prior knowledge. In order to effectively solve the problem, the parameters of the broad learning system such as the number of feature nodes, the number of enhancement nodes, and the number of mapped features layer were optimized by the improved pelican optimization algorithm, and a classification model was built based on the broad learning system optimized by the improved pelican optimization algorithm. The classification accuracy of the proposed model was the highest and reached 98.75%. It is further shown that compared with the support vector machine, deep belief networks, and broad learning system models optimized by particle swarm optimization algorithm, the proposed model effectively improves the accuracy of wind turbine fault diagnosing.

1. Introduction

As a clean renewable energy source, wind energy does not produce pollutants [1], and WT is an important contributor toward energy production free of CO2 [2]. According to the most recent report of the WWEA, the global wind power capacity reached 840 Gigawatt in 2021 [3]. Wind farms are generally located in remote areas such as plateaus or coastal areas, with inconvenient transportation, a poor working environment, and susceptibility to extreme weather conditions, which bring many difficulties to the operation and maintenance of wind turbines [4]. With the increase in operation time, various types of failures may occur in the electrical system, transmission system, and control system of WT, leading to the WT’s abnormal operation, and high operation and maintenance costs [5].
Fault diagnosis technology plays an important role in the running safety of WT, and maintenance staff use this technology to detect abnormality in time, handle it accordingly, and to extend the lifetime of WT [6]. An analysis of the series data obtained by a SCADA system is usually used to diagnose a WT’s condition [7]. A large number of variables related to WT operating characteristics, such as wind speed, output power, temperature, current, and voltage are collected in the SCADA system, which can provide a rich source of data for WT diagnosis [8].
Artificial intelligence (AI) methods such as machine learning (ML), artificial neural network (ANN), and deep learning (DL) are commonly used in the WT fault diagnosis, because of their robustness and self-adaptive capabilities. ML is to improve computer intelligence by independently learning the patterns that exist in large amounts of data and gaining new experience and knowledge to have a human-like decision-making ability. The ML algorithms that are most widely used in fault diagnosis are mainly K-nearest neighbor (KNN) [9], SVM [10], random forest (RF) [11], and extreme learning machine (ELM) [12]. ANN is a complex network that simulates the structure of the biological nervous system, with a large number of simple interconnected neurons. It performs the parallel processing of information and non-linear transformation by imitating the way a human brain neural processes information [13]. In the past decades, it has been developed rapidly and widely used in fault diagnosis of WT, for example, ANN methods have been used to diagnose faults in gearboxes [14,15,16,17] bearings [18,19], and generators [20,21].
Although ANN and ML have been broadly applied for fault diagnosis in recent decades, their diagnostic performance largely relies on feature extraction and selection. Instead, the DL algorithm has gradually become the main method for WT fault diagnosis due to its powerful feature extraction ability. At present, DL methods such as convolutional neural networks (CNN) [22,23], stacked automatic encoders (SAE), recurrent neural networks (RNN) [24], and deep belief networks (DBN) [25] are used in the fault diagnosis of rotating machinery.
Due to the complex deep structure of DL, which involves a large number of hyperparameters, the DL network is trained slowly and easily falls into a locally optimal solution. To overcome these problems, Chen et al. [26] proposed the BLS, which not only has a simple structure, fast training speed, and high accuracy rate but also has the advantage of enhanced learning. After BLS was proposed, researchers have successively proposed many improvements of BLS [27,28] and applied them to image classification [29], pattern recognition [30], numerical regression [31], automatic control [32], and other fields.
As a novel learning method, BLS solves the problem that the traditional neural network training methods are not applicable to multi-layer network training, effectively prevents the network from falling into local optimum, and improves the effectiveness of network convergence. In order to build a BLS model that satisfies the requirements, it is necessary to adapt the relevant parameters, and researchers often set the parameters according to their practical experience and a priori knowledge. For different problems, it may be necessary to repeatedly manually tune the relevant parameters.
To solve such problems, the parameters of BLS are optimized using the IPOA algorithm, hence, constructing the IPOA-BLS classification model. The experimental results indicate that the proposed model can better classify the faults of WT. The main contributions are highlighted as follows:
(1)
The experimental SCAD data were preprocessed and fault-related features were selected based on the principal component analysis (PCA) method [33,34];
(2)
The original POA algorithm was improved and was tested with benchmark functions;
(3)
The parameters of the BLS such as the number of feature nodes N f , number of enhancement nodes N e , and number of mapped feature layers N f l were optimized by the IPOA algorithm;
(4)
The processed WT fault data were used as the experimental sample with the SVM, DBN, BLS, PSO-BLS, POA-BLS, and IPOA-BLS models to classify and compare performance.
The rest of this paper is organized as follows: BLS, POA, and IPOA algorithms are introduced in Section 2. Section 3 describes SCADA data preprocessing, the IPOA-BLS classification model, and relevant parameter settings. Some graphical results with an analysis are given in Section 4. Finally, Section 5 concludes this paper.

2. Description of BLS, POA, and IPOA

2.1. BLS

BLS is a novel learning method proposed on the basis of a random vector functional-link neural network (RVFLNN) [35,36]. It does not require a time-consuming training process and has a strong function approximation capability. The structure diagram of RVFLNN and BLS is shown in Figure 1.
As can be seen from Figure 1, RVFLNN consists of three parts, namely the input layer, the enhancement node, and the output layer, and the output layer connects both the input layer and the enhancement nodes (EN). In contrast to RVFLNN, the BLS maps to the mapped feature (MF) first before the input layer data is mapped to EN, and the output layer connects both MF and EN.
In terms of network structure, the network of BLS is horizontally expanded and vertically fixed, which is very different from deep neural networks, and the number of layers is greatly reduced, and its structure is shown in Figure 2. A feature node Z i , maps the original data to obtain the following formula:
Z i = θ i X W f i + B f i , i = 1 , 2 , , n
where X is the input layer data; W f i and B f i denote the weight and bias of the input layer X to Z i , respectively, which is generally fine-tuned by a sparse autoencoder to produce the optimal value; θ i is a linear or nonlinear activation function; Z i is the i-th group of feature nodes containing p neurons.
All feature nodes can be represented as Z n = Z 1 , Z 2 , , Z i , , Z n , after obtaining the feature node Z n , it can be used to calculate the enhancement node H m .
H j = ξ j Z n W h j + B h j , j = 1 , 2 , , m H m = H 1 , H 2 , , H j , , H m
where H j is the j-th group of enhancement nodes containing q neurons; ξ j is the nonlinear activation function; W h j and B h j represent weights and biases, respectively. The model of the BLS can be obtained by further calculations:
Y = Z n ξ Z n W h 1 + B h 1 , , ξ Z n W h m + B h m W m = Z n H 1 , , H m W m = Z n H m W m if A = Z n H m Y = A W m
where W m represents the weight between the output layer and the hidden layer formed by the feature nodes and enhancement nodes. In the BLS, parameters such as W f i , B f i , W h j , B h j are randomly generated and fine-tuned by a sparse autoencoder, and remain constant during the training process; the metrics learned in the network are only W m . The objective function of the BLS is:
a r g   m i n W m : A W m Y 2 2 + λ W m 2 2 W m = A T A + λ I 1 A T Y
where A W m Y 2 2 is used to control the minimization of training error; λ W m 2 2 is used to prevent overfitting of the model; λ is the regularization factor; Y is the output; A T is the transpose matrix of A ; I is the unit matrix.

2.2. POA

The POA is a new meta-heuristic optimization algorithm [37] inspired by pelican hunting behaviors. It has the advantages of few adjustment parameters, fast convergence speed, and simple calculation. Pelicans are found in the warm waters of the world and live mainly in lakes, rivers, coasts, and swamps [38]. Pelicans generally live in flocks; they are not only good at flying but also good at swimming [39]. They have sharp eyesight in flight, as well as excellent observation skills, and they mainly feed on fish. After determining the location of the prey, the pelicans rush towards the prey from a height of 10–20 m and dive straight into the water to hunt [40]. If pelicans find schools of fish, they will arrange themselves in a line or U-shape to swoop down from the sky towards the fish in the water and use their wings to flap the water, forcing fishes to move upwards, and then collect the fishes in their throat pouch. Depending on the above explanation, the mathematical model of the POA algorithm was built.
(1)
Initialization: Assuming that there are N pelicans in an M dimensional space, the position of the i-th pelican in the M dimensional space is P i = [ p i 1 ,   p i 2 ,     , p i m , ,   p i M ] , the position P of the N pelicans is expressed as follows:
P = P 1 P 2 P i P N = p 11 p 12 p 1 m p 1 M p 21 p 22 p 2 m p 2 M p i 1 p i 2 p i m p i M p N 1 p N 2 p N m p N M , i = 1 , 2 , , N
where p i m denotes the position of the i-th pelican in the m-th dimension. At the initialization stage, pelicans are randomly distributed within a certain range, and the position update of the pelican is described as
P i m = l o w m + r a n d o m · u p m l o w m , i = 1 , 2 , , N ; m = 1 , 2 , , M ;
where l o w m , u p m is the search range of the pelican; r a n d o m is a random number between (0, 1).
(2)
Moving towards prey: In this phase, the pelican identifies the prey’s location and then rushes to the prey from a high altitude as shown in Figure 3. The random distribution of the prey in the search space increases the exploration ability of the pelican, and the update of the pelican’s location during each iteration is described as
P i m t + 1 = P i m t + r a n d ( S m t λ P i m t ) , F ( P S ) < F ( P i ) P i m t + r a n d ( P i m t S m t ) , F ( P S ) F ( P i )
where t is a current iteration number; P i m t denotes the position of the i-th pelican in the m-th dimension; S m t is the position of the prey in the m-th dimension; λ is randomly equal to 1 or 2; F ( P s ) is the objective function value; F ( P i ) denotes the value of the fitness function of i-th pelican in the m-th dimension.
(3)
Winging on the water surface: After the pelicans reach the surface of the water, they spread their wings on the surface of the water to move the fish upwards, then collect the prey in their throat pouch. This behavior of pelicans during hunting is mathematically simulated
P i m t + 1 = P i m t + γ T t T 2 r a n d o m 1 P i m t
where t is a current number of iterations; T is a maximum iteration number; γ T t T is the neighborhood radius of P i m t , and it represents the radius of the neighborhood of the population members to search locally near each member to converge to a better solution; r a n d o m is a random number between (0, 1).

2.3. IPOA

After improving the original POA algorithm, the optimization efficiency could be further improved. The specific improvement strategy is as follows.
(1)
Initialization strategy: The Tent chaotic map [41] is used to replace the randomly generated method in the original POA to initialize the pelicans after the Tent chaotic mapping is introduced, and the Equation (6) can be rewritten as follows:
p i m = l o w m + T e n t · u p m l o w m , i = 1 , 2 , , N ; m = 1 , 2 , , M ;
T e n t t + 1 = T e n t t z , T e n t t [ 0 , z ) ( 1 T e n t t ) ( 1 z ) , T e n t t [ z , 1 ]
where t is a current number of iterations; T is a maximum iteration number; where z ( 0 , 1 ) , T e n t t [ 0 , 1 ] , t = 1 , 2 , , T .
At this stage, the position of the pelicans is initialized using the Tent chaotic map, which helps to improve the global search performance of the POA algorithm.
(2)
Moving towards prey: At this stage, the dynamic weight factor [42] θ helps the pelican to constantly update its position. At the beginning of the iteration, θ has a large value, when the pelican is able to perform a better global search, and at the end of the iteration θ decreases adaptively, and this time the pelican is able to perform a better local search while increasing the convergence speed. the Equation (7) can be rewritten as follows:
P i m t + 1 = θ = e 2 ( 1 t / T ) e 2 ( 1 t / T ) e 2 ( 1 t / T ) + e 2 ( 1 t / T ) P i m t + r a n d ( S m t P i m t ) θ , F ( P S ) < F ( P i ) P i m t + r a n d ( P i m t S m t ) θ , F ( P S ) F ( P i )
To prove the correctness and validity of the IPOA algorithm, ten benchmark functions were tested using PSO, GWO, WOA, POA, and IPOA. The value of each algorithm parameter is shown in Table 1.
Descriptions of the benchmark functions are shown in Table 2. In this table, f 1–f 4 are the unimodal test functions, f 5–f 7 are the multimodal test functions, and f 8–f 10 are the composite benchmark functions.
Table 3, Table 4 and Table 5 report the relevant statistical experiment results from running each algorithm 30 times in the benchmark function independently. In these tables, “Best”, “Worst”, “Average”, and “Std.” represent the best value, the worst value, the average value, and the standard deviation of the algorithm, respectively.
The convergence curve of the f 1f 10 functions is shown in Figure 4: the red line denotes the IPOA algorithm, and the blue line represents the POA algorithm. It implies that the fitness value of the IPOA algorithm is lower and the convergence speed faster than other algorithms.

3. Wind Turbine Fault Classification

3.1. SCADA Data Preprocessing

In this study, the information from the SCADA system recorded from 38 Vestas 1.5 MW wind turbines in a wind farm in Liu He from 1 January 2018 to 31 December 2018 was used as experimental data and the time resolution of the data was ten minutes. After data cleaning, the fault categories were coded in the experimental data set such as the gearbox overheated (F1), yaw system fault (F2), generator overheated (F3), inverter system fault (F4), and pitch system fault (F5). Each fault includes 640 training samples and 160 test samples. A total of 4000 samples were selected, among them 3200 samples for training and 800 samples for testing.
In this experimental SCADA data, there were 52 observation feature parameters of the Vestas wind turbine. After feature extraction with the PCA method and comparison with Pearson correlation coefficient values, 35 characteristic parameters were chosen as the observation parameters of fault diagnosis. The relevant feature codes and their description are specified in Table 6. The feature code and its Pearson correlation coefficient of each fault are shown in Table 7.

3.2. IPOA-BLS Classification Model

As shown in Figure 5, the classification process of the proposed model is described as follows:
  • Step 1: The relevant data of WT were collected by the SCADA system.
  • Step 2: PCA was used to reduce the dimensionality of the preprocessed SCADA data, and fault-related features were selected. After cleaning and dimensionality reduction, 4000 groups were taken as experimental samples, and each sample had 35 features.
  • Step 3: Training samples and testing samples were taken from the experimental samples. In Figure 5, each of the five fault types included 640 training samples and 160 testing samples. X i , j represents the j-th feature value of the i-th sample, and Y i , j represents the j-th fault type of the i-th sample.
  • Step 4: The IPOA-BLS model was used for training and testing, respectively, and in this step:
    (1)
    The BLS { N f , N e , N f l } parameter and the IPOA algorithm parameters, including the pelican’s population N , the maximum iterations T , the upper limit of argument u p , the lower limit of argument l o w , the dimension M , and the sample d a t a were initialized.
    (2)
    The data of the fitness value were calculated. The classification accuracy obtained by the BLS was used as the fitness value, within the iteration range. Equations (9), (11), and (8) were used to calculate the location of the pelican. If the current new location was better, the old location was updated.
    (3)
    According to the optimal parameter combination { N f , N e , N f l } , the BLS was trained, and testing samples were used to classify and save the result.
  • Step 5: The results from Step 4 were anti-normalized.
  • Step 6: The final classification results were obtained and outputted with graphical.
  • Step 7: The relevant evaluation metrics were calculated.

3.3. Parameters Setting

In this paper, algorithms such as PSO, POA, and IPOA were used to optimize the hyperparameters of the BLS, and the relevant parameter values are shown in Table 8.

4. Experimental Results and Analysis

In order to state the advantages of the IPOA-BLS model in fault diagnosis, the classification results were compared with SVM, DBN, BLS, PSO-BLS, and POA-BLS diagnostic models. The classification results of each model are shown in Figure 6. In Figure 6, the blue circle (○) and red star (✲) mean the actual and predicted classification of the testing set samples, respectively. The blue circles and red stars overlap each other when the actual and predicted values of the classification are equal. If they are different, the blue circle and the red star do not overlap. The more overlapping means the fault resolution of the model is higher. In Figure 6a, the blue circle and red star overlap less, and the classification effect of the SVM model plays worse. In Figure 6f, the red star and the blue circle overlap each other the most times, which indicates that the IPOA-BLS model has the best classification effect.
To further visually reflect the fault identification capabilities of the model, the confusion matrix was used to visualize the fault recognition ability of SVM, DBN, BLS, PSO-BLS, POA-BLS, and IPOA-BLS models. The confusion matrix of each diagnostic model was calculated and plotted separately, and the results are shown in Figure 7. The diagonal elements of the confusion matrix indicate the number of samples that can be accurately classified, and the greater the number of diagonal elements, the better the classification performance of the model. The off-diagonal elements of the confusion matrix represent the number of samples that can be wrongly classified, and the smaller number of off-diagonal elements, the better the classification efficiency of the model.
In Figure 7a, the SVM model can only accurately classify fault-F3. In total, 14 samples originally belonging to fault-F1 were misclassified as fault-F2. Two samples originally belonged to fault-F2; however, one of them was misclassified as fault-F1, and another one was misclassified as fault-F4, respectively. Altogether, 45 samples originally belonged to fault-F4; however, 1 of them was misclassified as fault-F5, 12 of them were misclassified as fault-F1, and 32 of them were misclassified as fault-F1, respectively. In total, 37 samples originally belonged to fault-F5, but 1 of them was misclassified as fault-F4, and 36 of them were misclassified as fault-F2, respectively. In Figure 7f, the IPOA-BLS model almost accurately classified fault-F1, fault-F2, fault-F3, and fault-F5. Only nine samples originally belonging to fault-F4 were misclassified as fault-F1.
It can be seen from Figure 7 that among the 800 test samples, the SVM model accurately classified 702 samples with an accuracy of 87.75%. The DBN model accurately classified 730 samples, and its accuracy was 91.25%. The BLS model exactly categorized 734 samples with an accuracy of 91.75%. The PSO-BLS model accurately classified 759 samples with an accuracy of 94.875%. The POA-BLS model rightly classified 774 samples, and its accuracy was 96.75%. The IPOA-BLS model exactly classified 790 samples with an accuracy of 98.75%. An observation can be made that the IPOA-BLS model was the most accurate compared with other models.
In order to better assess the performance of the proposed classification model, accuracy, precision, recall, and F1-score were used in the present study. The specific expressions are, respectively, as follows.
(1)
The accuracy represents the number of correctly predicted samples as a percentage of the total number of samples.
a c c u r a c y = n u m b e r   o f   c o r r e c t l y   c l a s s i f i e d   s a m p l e s t o t a l   n u m b e r   o f   s a m p l e s
(2)
The precision reflects the proportion of samples with a positive prediction that are really positive.
p r e c i s i o n = T P T P + F P
(3)
The recall is the proportion of the number of correctly classified positive samples to the number of true positive samples.
r e c a l l = T P T P + F N
(4)
The F1-score is the weighted harmonic mean of precision and recall.
F 1 - s c o r e = 2 · p r e c i s i o n · r e c a l l p r e c i s o n + r e c a l l
In the above equations, TP represents true positives; FP is false positives; and FN is false negatives.
In each fault, the classification performance metrics are shown in Table 9. It can be seen from Table 9 that the proposed model had the higher evaluation indicators no matter which evaluation standard was used.
The IPOA-BLS model has many advantages. Compared with the other models, there was a significant improvement in the classification accuracy in terms of WT fault diagnosis. The PCA method played an important role in dimension reduction and feature extraction, and it was used to reduce the number of features of the experimental SCADA data from 52 to 35, which helped to decrease the calculation time of the model. The IPOA algorithm was evaluated through some different benchmark functions and was applied to the parameter’s estimation of BLS. As can be seen from Figure 6 and Figure 7, compared with the standard pelican optimization algorithm, the improved pelican optimization algorithm had a positive effect on BLS’s parameters optimization. After optimizing the related parameters of BLS with the improved pelican optimization algorithm, the number of rightly classified samples increased by about 16, and the accuracy improved by 2%.

5. Conclusions

In this paper, in addition to the chaotic map strategy, the dynamic weight factor strategy was also used to improve the location of the standard pelican optimization algorithm in the initialization and moving towards prey stage; this helps to enhance the exploration of the improved pelican optimization algorithm technique. To improve the classification accuracy of the broad learning system model, the experimental data were normalized and the improved pelican optimization algorithm was used to optimize hyperparameters such as the number of feature nodes N f , number of enhancement nodes N e , and number of mapped feature layers N f l . Performance indexes, classification accuracy, a classification results diagram, and confusion matrices fully demonstrate the excellent performance of the proposed classification model and verify the classification effect. The proposed model is very suitable for wind turbine fault classification in future practical applications.
The classification accuracy rate of the proposed model reaches a high level for typical wind turbine faults. Consequently, this model possesses great potential in wind turbine fault detection. In future studies, the development of the algorithms for fault feature extraction purposes will be investigated, and classification models based on deep learning will be analyzed.

Author Contributions

Data curation, M.H., L.G., Z.C.; methodology, C.X.; investigation and writing, W.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology of Peoples Republic of China (Grant No. 2019YFE0104800), the Special Training Plan for Minority Science and Technology Talents of Natural Science Foundation of Xinjiang Uyghur Autonomous Region (Grant No. 2020D03004), the Fundamental Research Funds for the Central Universities (Grant No. B210201018), and the National Natural Science Foundation of China (Grant No. 52106238).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support this manuscript are available from Chang Xu upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

AIArtificial intelligenceGWOGrey wolf optimizer
ANNArtificial neural networkKNNK-nearest neighbor
AverageAverage value of the algorithmMFMapped feature
BestBest value of the algorithmMLMachine learning
BLSBroad learning systemPCAPrincipal component analysis
CNNConvolutional neural networksPOAPelican optimization algorithm
DBNDeep belief networkPSOParticle swarm optimization
DLDeep learningRFRandom forest
ENEnhancement nodesRNNRecurrent neural networks
F1Fault 1: the gearbox overheatedSAEStacked automatic encoders
F2Fault 2: the yaw system motor faultStd.Standard deviation of the algorithm
F3Fault 3: generator overheatedSVMSupport vector machine
F4Fault 4: invertor system faultWOAWhale optimization algorithm
F5Fault 5: pitch system faultWorstWorst value of the algorithm
IPOAImproved pelican optimization algorithmWTWind turbine
SCADASupervisory control and data acquisitionWWEAWorld wind energy association
RVFLNNRandom vector functional-link neural network
PSO-BLSBroad learning system model optimized by particle swarm optimization
POA-BLSBroad learning system model optimized by pelican optimization algorithm
IPOA-BLSBroad learning system model optimized by improved pelican optimization algorithm

References

  1. Tian, Z.; Chen, H. Multi-step short-term wind speed prediction based on integrated multi-model fusion. Appl. Energy 2021, 298, 117248. [Google Scholar]
  2. Freire, N.M.A.; Cardoso, A.J.M. Fault Detection and Condition Monitoring of PMSGs in Offshore Wind Turbines. Machines 2021, 9, 260. [Google Scholar] [CrossRef]
  3. WWEA. World Market for Wind Power Saw Another Record Year in 2021: 97,3 Gigawatt of New Capacity Added, Total Global Capacity Reached 840 Gigawatt. Available online: https://wwindea.org/world-market-for-wind-power-saw-another-record-year-in-2021-973-gigawatt-of-new-capacity-added/ (accessed on 18 March 2022).
  4. Xue, F.; Duan, H.; Xu, C.; Han, X.; Shangguan, Y.; Li, T.; Fen, Z. Research on the Power Capture and Wake Characteristics of a Wind Turbine Based on a Modified Actuator Line Model. Energies 2022, 15, 282. [Google Scholar] [CrossRef]
  5. Meyer, A. Vibration Fault Diagnosis in Wind Turbines Based on Automated Feature Learning. Energies 2022, 15, 1514. [Google Scholar] [CrossRef]
  6. Dao, P.B. Condition monitoring and fault diagnosis of wind turbines based on structural break detection in SCADA data. Renew. Energy 2022, 185, 641–654. [Google Scholar] [CrossRef]
  7. Zhang, S.; Lang, Z. SCADA-data-based wind turbine fault detection: A dynamic model sensor method. Control Eng. Pract. 2020, 102, 104546. [Google Scholar] [CrossRef]
  8. Velandia-Cardenas, C.; Vidal, Y.; Pozo, F. Wind Turbine Fault Detection Using Highly Imbalanced Real SCADA Data. Energies 2021, 14, 1728. [Google Scholar] [CrossRef]
  9. Chen, X.; Li, L.; Liu, R.; Yang, B.; Deng, Z.; Li, Z. Fault Diagnosis Method of Wind Turbine Pitch Angle Based on PCA-KNN Fusion Algorithm. Electr. Power 2021, 54, 190–198. [Google Scholar]
  10. Tuerxun, W.; Chang, X.; Hongyu, G.; Zhijie, J.; Huajian, Z. Fault Diagnosis of Wind Turbines Based on a Support Vector Machine Optimized by the Sparrow Search Algorithm. IEEE Access 2021, 9, 69307–69315. [Google Scholar] [CrossRef]
  11. Zhang, L.; Liu, K.; Wang, Y.; Omariba, Z. Ice Detection Model of Wind Turbine Blades Based on Random Forest Classifier. Energies 2018, 11, 2548. [Google Scholar] [CrossRef] [Green Version]
  12. Marti-Puig, P.; Blanco-M, A.; Serra-Serra, M.; Solé-Casals, J. Wind Turbine Prognosis Models Based on SCADA Data and Extreme Learning Machines. Appl. Sci. 2021, 11, 590. [Google Scholar] [CrossRef]
  13. Samarasinghe, S.S. Artificial Neural Network Modelling; Springer: Cham, Switzerland, 2016; pp. 1–14. [Google Scholar]
  14. Huang, Q.; Jiang, D.; Hong, L.; Ding, Y. Application of Wavelet Neural Networks on Vibration Fault Diagnosis for Wind Turbine Gearbox. In International Symposium on Neural Networks; Springer: Berlin/Heidelberg, Germany, 2008; pp. 313–320. [Google Scholar] [CrossRef]
  15. Strączkiewicz, M.; Barszcz, T. Application of Artificial Neural Network for Damage Detection in Planetary Gearbox of Wind Turbine. Shock Vib. 2016, 2016, 4086324. [Google Scholar] [CrossRef] [Green Version]
  16. Yang, S.; Li, W.; Wang, C. The intelligent fault diagnosis of wind turbine gearbox based on artificial neural network. In Proceedings of the 2008 International Conference on Condition Monitoring and Diagnosis, Beijing, China, 21–24 April 2008. [Google Scholar] [CrossRef]
  17. Feng, Z.; Liang, M.; Zhang, Y.; Hou, S. Fault diagnosis for wind turbine planetary gearboxes via demodulation analysis based on ensemble empirical mode decomposition and energy separation. Renew. Energy 2012, 47, 112–126. [Google Scholar] [CrossRef]
  18. Li, Z.; Yu, J.; Hu, C.; Peng, Z. Multi-Dimensional Variational Mode Decomposition Applied to Intrinsic Vibration Mode Extraction for Bearing Crack Detection in Wind Turbines with Large Speed Variation. In Proceedings of the 35th Wind Energy Symposium, Grapevine, TX, USA, 9–13 January 2017. [Google Scholar] [CrossRef]
  19. Zimroz, R.; Bartelmus, W.; Barszcz, T.; Urbanek, J. Diagnostics of bearings in presence of strong operating conditions non-stationarity—A procedure of load-dependent features processing with application to wind turbine bearings. Mech. Syst. Signal Process. 2014, 46, 16–27. [Google Scholar] [CrossRef]
  20. Santolamazza, A.; Dadi, D.; Introna, V. A Data-Mining Approach for Wind Turbine Fault Detection Based on SCADA Data Analysis Using Artificial Neural Networks. Energies 2021, 14, 1845. [Google Scholar] [CrossRef]
  21. Attya, A.B.; Hartkopf, T. Penetration impact of wind farms equipped with frequency variations ride through algorithm on power system frequency response. Int. J. Electr. Power Energy Syst. 2012, 40, 94–103. [Google Scholar] [CrossRef]
  22. Rahimilarki, R.; Gao, Z.; Jin, N.; Zhang, A. Convolutional neural network fault classification based on time-series analysis for benchmark wind turbine machine. Renew. Energy 2022, 185, 916–931. [Google Scholar]
  23. Xiang, L.; Wang, P.; Yang, X.; Hu, A.; Su, H. Fault detection of wind turbine based on SCADA data analysis using CNN and LSTM with attention mechanism. Measurement 2021, 175, 109094. [Google Scholar] [CrossRef]
  24. Liu, H.; Zhou, J.; Zheng, Y.; Jiang, W.; Zhang, Y. Fault diagnosis of rolling bearings with recurrent neural network-based autoencoders. ISA Trans. 2018, 77, 167–178. [Google Scholar] [CrossRef]
  25. Shao, H.; Jiang, H.; Zhang, H.; Duan, W.; Liang, T.; Wu, S. Rolling bearing fault feature learning using improved convolutional deep belief network with compressed sensing. Mech. Syst. Signal Process. 2018, 100, 743–765. [Google Scholar] [CrossRef]
  26. Chen, C.P. Broad Learning System and its Structural Variations. In Proceedings of the 2018 IEEE 16th International Symposium on Intelligent Systems and Informatics (SISY), Subotica, Serbia, 13–15 September 2018. [Google Scholar] [CrossRef]
  27. Liu, Z.; Huang, S.; Jin, W.; Mu, Y. Broad learning system for semi-supervised learning. Neurocomputing 2021, 444, 38–47. [Google Scholar] [CrossRef]
  28. Ding, W.; Tian, Y.; Han, S.; Yuan, H. Greedy Broad Learning System. IEEE Access 2021, 9, 79307–79315. [Google Scholar] [CrossRef]
  29. Liu, R.; Liu, Y.; Zhao, Y.; Chen, X.; Cui, S.; Wang, F.; Yi, L. Multi-Feature Broad Learning System for Image Classification. Int. J. Pattern Recogn. 2021, 35, 2150033. [Google Scholar] [CrossRef]
  30. Cheng, L.; Wang, X.; Zhang, D.; Jiang, L. Face Recognition Algorithm Based on Broad Learning System. In Proceedings of the 2021 5th Asian Conference on Artificial Intelligence Technology (ACAIT), Haikou, China, 29–31 October 2021; pp. 564–569. [Google Scholar] [CrossRef]
  31. Chew, M.; Wu, P.; Chen, Y. Operon Prediction Using Broad Learning System. In Proceedings of the 2020 International Conference on Electrical, Communication, and Computer Engineering (ICECCE), Istanbul, Turkey, 12–13 June 2020; pp. 1–5. [Google Scholar] [CrossRef]
  32. Liu, Y.; Wang, Y.; Chen, L.; Zhao, J.; Liu, Q. Incremental Bayesian broad learning system and its industrial application. Artif. Intell. Rev. 2021, 54, 1–21. [Google Scholar] [CrossRef]
  33. Pozo, F.; Vidal, Y.; Salgado, Ó. Wind Turbine Condition Monitoring Strategy through Multiway PCA and Multivariate Inference. Energies 2018, 11, 749. [Google Scholar] [CrossRef] [Green Version]
  34. Wang, Y.; Ma, X.; Qian, P. Wind Turbine Fault Detection and Identification Through PCA-Based Optimal Variable Selection. IEEE Trans. Sustain. Energy 2018, 9, 1627–1635. [Google Scholar] [CrossRef] [Green Version]
  35. Pao, Y.H.; Takefuji, Y. Functional-link net computing: Theory, system architecture, and functionalities. Computer 1992, 25, 76–79. [Google Scholar] [CrossRef]
  36. Pao, Y.; Park, G.; Sobajic, D.J. Learning and generalization characteristics of the random vector functional-link net. Neurocomputing 1994, 6, 163–180. [Google Scholar] [CrossRef]
  37. Trojovský, P.; Dehghani, M. Pelican Optimization Algorithm: A Novel Nature-Inspired Algorithm for Engineering Applications. Sensors 2022, 22, 855. [Google Scholar] [CrossRef]
  38. Santoso, N.; Sutopo. Habitat Conservation of Australian Pelican (Pelecanus conspicillatus Temminick 1824) in Mangrove Ecosystem of Bengawan Solo Estuary, Gresik Regency—East Java Province. IOP Conf. Ser. Earth Environ. Sci. 2019, 394, 12031. [Google Scholar] [CrossRef]
  39. John, G.T.A. Foraging Behavior of the American White Pelican (Pelecanus erythrorhyncos) in Western Nevada. Colonial Waterbirds 1991, 14, 166–172. [Google Scholar] [CrossRef]
  40. O’Malley, J.; Evans, R. Kleptoparasitism and Associated Foraging Behaviors in American White Pelicans. Colonial Waterbirds 1983, 6, 126–129. [Google Scholar] [CrossRef]
  41. Li, Y.; Han, M.; Guo, Q. Modified Whale Optimization Algorithm Based on Tent Chaotic Mapping and Its Application in Structural Optimization. KSCE J. Civ. Eng. 2020, 24, 3703–3713. [Google Scholar] [CrossRef]
  42. Liu, J.S.; Yuan, M.M.; Zuo, F. Global search-oriented adaptive leader salp swarm algorithm. Control Decis. 2021, 9, 2152–2160. [Google Scholar] [CrossRef]
Figure 1. Structure of RVFLNN and BLS.
Figure 1. Structure of RVFLNN and BLS.
Machines 10 00407 g001
Figure 2. Structure of BLS.
Figure 2. Structure of BLS.
Machines 10 00407 g002
Figure 3. Diagram of the pelicans moving towards prey.
Figure 3. Diagram of the pelicans moving towards prey.
Machines 10 00407 g003
Figure 4. The convergence curves of benchmark functions.
Figure 4. The convergence curves of benchmark functions.
Machines 10 00407 g004aMachines 10 00407 g004b
Figure 5. Classification process of the IPOA-BLS model.
Figure 5. Classification process of the IPOA-BLS model.
Machines 10 00407 g005
Figure 6. Classification result of the different models.
Figure 6. Classification result of the different models.
Machines 10 00407 g006aMachines 10 00407 g006b
Figure 7. Confusion matrices of the different model.
Figure 7. Confusion matrices of the different model.
Machines 10 00407 g007aMachines 10 00407 g007b
Table 1. Parameter setting of PSO, GWO, WOA, POA, and IPOA.
Table 1. Parameter setting of PSO, GWO, WOA, POA, and IPOA.
AlgorithmParameter Setting
Common settingMaximum iteration T = 1000
Population number: N = 30
Runs: r = 30
PSOC1 = 2
C2 = 2
W = 0.7
GWOm1 = 0.3
WOAL1 = 3
L2 = 5
POAΛ= round (random (1,2))
γ = 0.2
IPOA λ = θ
γ = 0.2
Table 2. Benchmark functions.
Table 2. Benchmark functions.
Benchmark FunctionFunction NameDRangefopt
(1) Unimodal test functions
f 1 x = Σ i = 1 n x i 2 Sphere30[−100,100]0
f 2 x = i = 1 D x i + Π i 1 D x i Schwefel 2.2230[−10,10]0
f 3 x = i = 1 D ( Σ j 1 D x i ) 2 Schwefel 1.230[−100,100]0
f 4 x = m a x i X i , 1 i D Schwefel 2.2130[−100,100]0
(2) Multimodal test functions
f 5 x = i = 1 D x i 2 10 cos 2 π x i + 10 sin 2 π x i Rastrigin30[−5.12,5.12]0
f 6 x = 20 + e 20 exp 20 1 / D Σ i = 1 D x i 2 ) exp ( ( 1 / D ) i = 1 D cos 2 π x i Ackley30[−32,32]8.8818 × 10−16
f 7 x = 1 / 4000 i = 1 D x i 2 i = 1 D cos x i / i + 1 Griewank30[−600,600]0
(3) composite benchmark functions
f 8 x = 1 / 500 + j = 1 25 1 / j + i = 1 2 x i a i j 6 1 Foxholes2[−65.53,65.53]0.998004
f 9 x = i = 1 11 a i x 1 b i 2 + b i x 2 / b i 2 + b i x 3 + x 4 1 Six hump4[−5,5]0.0003075
f 10 x = i = 1 10 X a i X a i T + c i 1 Langerman 104[0,10]−10.5364
Table 3. Statistical results for unimodal benchmark functions.
Table 3. Statistical results for unimodal benchmark functions.
FunctionStatisticsPSOGWOWOAPOAIPOA
f 1Best8.0606 × 10+11.3824 × 10−581.3904 × 10−1308.1414 × 10−2070
Worst2.1442 × 10+46.7429 × 10+47.2836 × 10+47.1909 × 10+41.1224 × 10+4
Average1.4696 × 10+33.3884 × 10+24.0963 × 10+21.4462 × 10+21.0947 × 10+2
Std.2.6760 × 10+33.4928 × 10+34.1509 × 10+32.5477 × 10+31.3498 × 10+3
f 2Best2.7469 × 10+18.8984 × 10−353.7388 × 10−952.3633 × 10−1070
Worst5.3174 × 10+127.3390 × 10+127.2151 × 10+133.6333 × 10+135.9060 × 10+12
Average1.2499 × 10+107.3536 × 10+97.3708 × 10+103.6522 × 10+105.9060 × 10+8
Std.2.0235 × 10+112.3208 × 10+112.2820 × 10+121.1490 × 10+121.8676 × 10+11
f 3Best9.8883 × 10+36.1321 × 10−143.0153 × 10+41.8928 × 10−2040
Worst6.4994 × 10+41.4551 × 10+51.6016 × 10+51.6112 × 10+51.9157 × 10+4
Average1.5945 × 10+41.6766 × 10+36.6527 × 10+45.8836 × 10+22.1205 × 10+2
Std.6.8243 × 10+39.3954 × 10+42.9123 × 10+47.8845 × 10+36.0947 × 10+3
f 4Best4.54081.5182 × 10−143.8109 × 10+12.1908 × 10−1052.6256 × 10−308
Worst5.9368 × 10+18.8048 × 10+18.8640 × 10+18.7096 × 10+19.5583 × 10+1
Average1.2826 × 10+11.84485.0235 × 10+13.5377 × 10−11.1376 × 10−1
Std.9.57049.55341.4559 ×10+13.92123.0484
Table 4. Statistical results for multimodal benchmark functions.
Table 4. Statistical results for multimodal benchmark functions.
FunctionStatisticsPSOGWOWOAPOAIPOA
f 5Best1.9909 × 10+22.3932 × 10−1000
Worst4.4057 × 10+24.4237 × 10+24.4746 × 10+24.5905 × 10+25.6918 × 10+2
Average2.4516 × 10+21.4587 × 10+11.3139 × 10+14.44178.4287 × 10−1
Std.4.0864 × 10+15.2324 × 10+14.9471 × 10+13.0706 × 10+11.8966 × 10+1
f 6Best8.07181.6046 × 10−143.9672 × 10−154.2040 × 10−158.8818 × 10−16
Worst2.0233 × 10+12.0699 × 10+12.0725 × 10+12.0740 × 10+12.1099 × 10+1
Average1.1201 × 10+13.9652 × 10−14.3575 × 10−11.6615 × 10−13.6062 × 10−2
Std.2.78732.26552.16931.40607.2128 × 10−01
f 7Best1.76942.7510 × 10−36.1431 × 10−300
Worst1.4914 × 10+26.2027 × 10+26.3457 × 10+26.5070 × 10+21.0007 × 10+3
Average1.2120 × 10+13.15993.90321.44821.0140
Std.1.8168 × 10+13.2284 × 10+13.7483 × 10+12.3843 × 10+13.1645 × 10+1
Table 5. Statistical results for composite benchmark functions.
Table 5. Statistical results for composite benchmark functions.
FunctionStatisticsPSOGWOWOAPOAIPOA
f 8Best9.9800 × 10−14.71603.67341.03111.5871
Worst4.4687 × 10+11.8978 × 10+22.3138 × 10+21.7590 × 10+21.4355 × 10+2
Average1.25975.08144.32971.39771.0184
Std.1.69816.11797.74065.92954.6540
f 9Best9.8558 × 10−35.7168 × 10−36.9056 × 10−41.7361 × 10−34.3471 × 10−4
Worst2.0059 × 10−11.9738 × 10−13.7275 × 10−18.3830 × 10−18.4090 × 10−1
Average1.0510 × 10−26.1706 × 10−31.7122 × 10−33.4990 × 10−31.5492 × 10−3
Std.7.2294 × 10−36.3856 × 10−31.3101 × 10−23.0524 × 10−22.6717 × 10−2
f 10Best−1.0177 × 10+1−1.0357 × 10+1−7.9845−9.6350−1.0520 × 10+1
Worst−7.7636 × 10−1−7.4272 × 10−1−7.2858 × 10−1−7.6370 × 10−1−6.1054 × 10−1
Average−9.1388−7.8546−7.2752−9.3844−9.6476
Std.1.66802.17161.17878.9320 × 10−11.5835 × 10−1
Table 6. Feature code and its description.
Table 6. Feature code and its description.
Feature CodeFeature DescriptionFeature CodeFeature Description
X1impeller speed (rpm)X19pitch bearings temperature (°C)
X2generator active power(kVA)X20pitch oil pressure (mPa)
X3average reactive power (kVA)X21pitch system motor speed (rpm)
X4average blade angle (deg)X22nacelle temperature (°C)
X5bias of blade (m)X23converter cabinet temperature (°C)
X6blade temperature (°C)X24converter cabinet fan temperature (°C)
X7paddle motor temperature (°C)X25yaw bearing radial force (kN)
X8paddle system temperature (°C)X26yaw drive temperature (°C)
X9ambient temperature (°C)X27yaw bearing axial force (kN)
X10generator stator winding temperature u (°C)X28average wind speed (m/s)
X11generator stator winding temperature v (°C)X29yaw inverter cabinet temperature (°C)
X12generator stator winding temperature w (°C)X30yaw bearings temperature (°C)
X13stator temperature (°C)X31yaw control cabinet temperature (°C)
X14rotor temperature (°C)X32wind direction angle
X15rotor speed (rpm)X33transformer temperature (°C)
X16gearbox oil temperature (°C)X34generator torque (kN·m)
X17gearbox bearing temperature (°C)X35tower temperature (°C)
X18gearbox oil pressure (mPa)
Table 7. Feature code (Fc) and its Pearson correlation coefficient (Pcc) of each fault.
Table 7. Feature code (Fc) and its Pearson correlation coefficient (Pcc) of each fault.
FcPccFcPccFcPccFcPccFcPcc
gearbox overheated (F1)
X20.3872X90.1736X100.4436X110.4507X120.4523
X130.3154X140.3511X150.4625X160.7831X170.8865
X180.3221X220.1023X280.4178
yaw system fault (F2)
X20.3661X250.3622X260.6134X270.3531X280.3164
X290.5171X300.5731X310.5463X320.5273
generator overheated (F3)
X20.4175X30.4336X100.6734X110.6804X120.6779
X130.7863X140.7162X150.5153X220.1227X340.5271
converter system fault (F4)
X140.4267X220.1103X230.6832X240.6054X330.3762
X350.1272
pitch system fault (F5)
X10.3874X20.3642X250.4623X260.6723X270.4167
X280.3465X290.5732X300.6264X310.6114X320.4178
Table 8. The parameter values of PSO-BLS, POA-BLS, and IPOA-BLS.
Table 8. The parameter values of PSO-BLS, POA-BLS, and IPOA-BLS.
Model NameParameter NameParameter Value
PSO-BLSIterations N30
Population P8
C14
C24
W0.9
Lower band lb[10 10 10]
Upper band ub[100 100 100]
POA-BLSIterations N30
Population P8
Lower band lb[10 10 10]
Upper band ub[100 100 100]
Λround (random (1,2))
γ 0.2
IPOA-BLSIterations N30
Population P8
Lower band lb[10 10 10]
Upper band ub [100 100 100]
Λ θ (dynamic weight factor)
γ 0.2
Table 9. Performance indicators of fault classification model.
Table 9. Performance indicators of fault classification model.
Classification ModelEvaluation IndexFault Types
F1F2F3F4F5
SVMPrecision81.56%71.82%100%98.29%99.19%
Recall91.25%98.75%100%71.88%76.88%
F1-score86.14%83.16%100%83.03%86.62%
Accuracy87.75%
DBNPrecision87.43%89.33%100%85.81%95.81%
Recall95.63%99.38%78.12%83.13%100%
F1-score91.34%94.08%87.72%84.44%97.86%
Accuracy91.25%
BLSPrecision89.41%84.28%98.16%97.89%89.76%
Recall95.00%83.75%100%86.88%93.13%
F1-score92.12%84.04%99.07%92.05%91.41%
Accuracy91.75%
PSO-BLSPrecision83.25%95.15%100%99.22%100%
Recall99.38%98.12%100%80.00%96.88%
F1-score90.60%96.62%100%88.58%98.41%
Accuracy94.87%
POA-BLSPrecision97.53%95.51%100%92.40%98.68%
Recall98.75%93.13%100%98.75%9313%
F1-score98.14%94.30%100%95.47%95.82%
Accuracy96.75%
IPOA-BLSPrecision94.14%100%100%100%100%
Recall100%99.38%100%94.37%100%
F1-score96.97%99.69%100%97.11%100%
Accuracy98.75%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tuerxun, W.; Xu, C.; Haderbieke, M.; Guo, L.; Cheng, Z. A Wind Turbine Fault Classification Model Using Broad Learning System Optimized by Improved Pelican Optimization Algorithm. Machines 2022, 10, 407. https://doi.org/10.3390/machines10050407

AMA Style

Tuerxun W, Xu C, Haderbieke M, Guo L, Cheng Z. A Wind Turbine Fault Classification Model Using Broad Learning System Optimized by Improved Pelican Optimization Algorithm. Machines. 2022; 10(5):407. https://doi.org/10.3390/machines10050407

Chicago/Turabian Style

Tuerxun, Wumaier, Chang Xu, Muhaxi Haderbieke, Lei Guo, and Zhiming Cheng. 2022. "A Wind Turbine Fault Classification Model Using Broad Learning System Optimized by Improved Pelican Optimization Algorithm" Machines 10, no. 5: 407. https://doi.org/10.3390/machines10050407

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop