Next Article in Journal
Nonlinear Static Analysis for Seismic Evaluation of Existing RC Hospital Building
Next Article in Special Issue
Examining Gait Characteristics in People with Osteoporosis Utilizing a Non-Wheeled Smart Walker through Spatiotemporal Analysis
Previous Article in Journal
The Chinese Aviation Network: An Empirical Temporal Analysis on Its Structural Properties and Robustness
Previous Article in Special Issue
Dynamic Anomaly Detection in Gantry Transactions Using Graph Convolutional Network-Gate Recurrent Unit with Adaptive Attention
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Wind Turbine Blade Cracking Detection under Imbalanced Data Using a Novel Roundtrip Auto-Encoder Approach

1
Henan Key Laboratory of Intelligent Manufacturing of Mechanical Equipment, Zhengzhou University of Light Industry, Zhengzhou 450003, China
2
CITIC Heavy Industries Co., Ltd., Luoyang 471039, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(21), 11628; https://doi.org/10.3390/app132111628
Submission received: 12 September 2023 / Revised: 13 October 2023 / Accepted: 13 October 2023 / Published: 24 October 2023
(This article belongs to the Special Issue Intelligent Manufacturing and Medical-Engineering Integration)

Abstract

:
Imbalanced data cause low recognition of wind turbine blade cracking. Existing data-level augmentation methods, including sampling and generative strategies, may yield lots of high-confidence but low-value samples, which fail to improve the detection of blade cracking. Therefore, this paper designs a novel RTAE (roundtrip auto-encoder) method. Based on the idea of the roundtrip approach, we design two generator networks and two discriminator networks to ensure the cycle mapping between cracking samples and latent variables. Further, by leveraging cycle consistency loss, generated samples fit the distribution of historical cracking samples well. Thus, these generated samples effectively realize data augmentation and improve recognition of blade cracking. Additionally, we apply an auto-encoder method to reduce the dimension of historical samples and thus the complexity of the generator network and discriminator network. Through the analysis of real wind turbine blade cracking data, the recognition of cracking samples is improved by 19.8%, 23.8% and 22.7% for precision, recall and F1-score.

1. Introduction

Due to the insufficient bending strength, tensile strength and other parameters of wind turbine blade material, as well as their overly large dynamic load, blade cracking is a common cause of failure, accounting for approximately 30% of all downtime accidents [1,2,3]. However, due to constantly changing speeds and transient impacts, modeling the failure mechanism of blade cracking proves very challenging. Furthermore, since blades account for roughly 15–20% of total equipment costs, blade cracking can lead to potentially enormous maintenance expenses and safety hazards. Given its importance and costs, timely and accurate approaches to detecting blade cracking are attracting widespread research attention [4,5,6].
Data-based detection approaches work independently of the dynamics and kinematics of physical systems, and are thus effective for blade cracking detection. However, data are generally imbalanced, with very few cracking samples in real scenarios, leading to detection approaches that support normal samples and rarely identify cracking samples in the early stage (theoretical analysis in Section 2.1). Aiming at such problems, current data-level solutions mainly include sampling and generative strategies.
Sampling strategies, e.g., SMOTE (synthetic minority oversampling technique) and ADASYN (adaptive synthetic), utilize prior knowledge about the space distribution of collected cracking samples to synthesize new samples. Ge [7] adopted 29 sensor features as original input and used SMOTE to synthesize virtual cracking samples via linear interpolation between two random real cracking samples. Meanwhile, the density and concentration area are introduced to increase the confidence of virtual cracking samples. Jiang [8] designed a synthetic and dependent wild-bootstrapped oversampling technique for wind turbine fault diagnosis which is a modification of SMOTE and mimics the temporal dependence of time-series data. In the data analysis part, two datasets collected from two wind farms of northeast and northwest China are used, with 70 attributes in total, including wind speed, x-axis nacelle vibration, yaw angle, etc. Cristian [9] dealt with imbalanced wind turbine data using the random oversampling technique, which directly removed some random normal samples to achieve amount balance. To verify the proposed method, 19 real variables covering 19 different attributes were used, including generated power, R-phase voltage, wind speed, etc. Yi [10] designed a minority clustering SMOTE approach for wind turbine fault detection, which achieved different clusters of minority samples and solved the problem of the uneven distribution of fault samples. Chen [11] adopted ADASYN to alleviate the critical imbalance and convolutional LSTM-GRU to recognize blade icing status. The dataset for verification was collected over 341.88 h with 26 sensor values, including wind speed, yaw position, vertical acceleration, etc.
Generative strategies estimate the probability density of cracking samples and then sample new data points to supplement cracking samples. Chen [12] designed a deep convolutional generative adversarial network (GAN) to produce a threshold for a condition monitoring scheme of wind turbines. The employed signal was the frequency spectrum transformed by fast Fourier transform. Liu [13] used a generative adversarial network to transform normal data into rough fault data, and furthermore, a refiner developed using a GAN was adopted to make them much similar to real fault data. For verification, the input data contained 28 variables, such as pitch angle, hub angle and generator torque. Liu [14] introduced sparse dictionary learning into an adversarial variational auto-encoder to generate virtual data and determine the posterior distribution from six sensor variables, including wind speed, active power, generator speed, three-phase current and voltage. Wang [15] designed a least-square GAN to determine the distribution of health data from 15 selected sensor variables and realize the data augmentation. Yang [16] used wavelet packet transform to generate time–frequency data of wind turbines and a GAN to compensate for the imbalance level. Ogaili [17] published a wind turbine fault diagnosis dataset considering vibration under different wind speeds for fault diagnosis under both imbalanced and balanced conditions. Jiang [18] took 28 sensor variables (e.g., yaw position, yaw speed) as input of a GAN to generate virtual blade icing samples. Zou [19] combined a convolution neural network and GAN to detect wind turbine blade surface damage using a small number of damage images.
Although plenty of approaches have been developed, there are still two critical problems that deserve in-depth research based on above studies. (1) Both sampling strategies and generative strategies easily produce virtual cracking samples with a high confidence but low usage value. The reason for this is that sampling strategies more often pick samples in a dense area as a basis, leading to many synthesized samples lying in the middle area. Similarly, the generator in a GAN yields virtual cracking samples in the middle area (with high confidence) to easily cheat the discriminator. However, virtual samples near the decision border rather than the middle area may be more helpful for supporting good classification results. (2) Generative strategies focus on generating real, similar cracking samples but ignore the overall space distribution of real samples. Furthermore, if real samples are distributed over more than one cluster, with some clusters owning many data points, generative strategies may only produce samples for these clusters. However, these clusters with few data points may be more important for improving classification accuracy.
Aiming at the above key points, this work designs a novel blade-cracking detection solution. Considering the case that the amount of redundant features will bring a high number of weights to be optimized in GAN, we curtail the input dimension using an unsupervised auto-encoder. Further, a virtual cracking sample generative strategy-based roundtrip framework is designed to achieve bidirectional mapping between virtual and real samples. By inverse mapping, it catches the overall space distribution and avoids the generated virtual samples falling into a dense area. Through verification on the benchmark dataset and real wind turbine blade cracking, the results show the effectiveness.

2. Motivation and Preliminary

2.1. Motivation

Wind turbine blade-cracking detection is typically a binary classification problem. In order to explain the effects of imbalanced data, this section takes two common binary classification methods for analysis, namely decision tree (C4.5) and support vector machine (SVM). For the former, a splitting node is selected according to the information gain ratio:
G a i n _ r a t i o = i = 1 N D i D D i + D i log 2 D i + D i D i D i log 2 D i D i D + D log 2 D + D D D log 2 D + D s . t .   D = D + + D , D i = D i + + D i D + D , D i + D i
where  D  denotes the total sample amount;  D +  denotes the normal sample amount;   D  denotes the blade cracking sample amount.  N  is the total number of attribute values for a specific attribute.  D i  denotes the sample amount for  i -th attribute value.  D i +  and  D i  are the normal sample amount and cracking sample amount for  i -th attribute value.
Since
lim x 0 + x l o g 2 x = lim x 0 + x ln x ln 2 = 1 ln 2 lim x 0 + x ln x                                                   = 1 ln 2 lim x 0 + ln x 1 x = 1 ln 2 l i m x 0 + 1 x 1 x 2 = 1 ln 2 l i m x 0 + x = 0
Supposing there are very few cracking samples, we can infer that
Gain _ ratio   i = 1 N   D i D D i + D i l o g 2 D i + D i D + D l o g 2 D + D = i = 1 N   D i + D + l o g 2 D i + D i log 2 D + D
From Equation (3), it is clear that gain ratio is irrelevant with cracking samples. That is, building the decision tree only depends on normal samples, causing a high recognition uncertainty for cracking samples.
For SVM, we take the linear separable case for explanation. If the data are balanced, the decision surface is obtained by minimizing Equation (4):
arg m i n w   Loss   = W + C i = 1 D   ξ i ,   s . t .   y i W x i + b 1 ξ i , ξ i 0
where  W  decides the position and direction of decision surface;  C  is a penalty coefficient;  ξ i   is a relaxation variable;  x i ,   y i  denotes  i -th sample. If some cracking samples are removed and surface shifts distance d towards cracking samples, the variation in Loss is computed by Equation (5):
Δ L o s s = C i = 1 D +   ξ i d s i n θ + C j = 1 D   ξ j + d s i n θ C i = 1 D +   ξ i + C j = 1 D   ξ j                         = C d s i n θ D D + < 0
where  θ  represents the angle between the x-axis and decision surface. Obviously,  Δ L o s s  is less than zero, implying a better support for normal samples, but it may bring many misclassifications of these unseen cracking samples near the former decision surface.
From the above analysis, the root cause for low recognition of cracking samples is the imbalanced sample amount. Current solutions mainly include two aspects, namely sampling strategies and generative strategies. However, there are still critical problems that deserve deep research. As illustrated in Figure 1a, sampling strategies (e.g., SMOTE and its variants) may pick the basis sample from a middle or dense area, leading to synthesized samples lying far from border. As shown in Figure 1b,c, generative strategies (such as GAN) tend to generate highly confident samples for easily cheating the discriminator. However, samples on the border with low confidence may be better for improving the decision surface. Additionally, in the case of several clusters (as Figure 1d), generative strategies may ignore cluster 1 because of less cracking samples, since it mainly fits the probability distribution of cluster 2. Instead, cluster 1 may be more important for improving classification accuracy.

2.2. Preliminary

Roundtrip

Roundtrip is a method based on cycle GAN and its framework is shown in Figure 2 [20]. It contains four parts: generator network G and H, discriminators DX and DZ. Z is a latent variable drawn from normal distribution; X denotes historical samples;  Z ~  and  X ~  denote generated latent variables and generated virtual samples. H transforms latent variables to virtual samples and G transforms historical samples to latent variables. Both of them are nonlinear transforms. DX is employed to discriminate real and virtual samples. Likewise, discriminator DZ discriminates real and virtual latent variables. DX and DZ determines their authenticity (real 0, fake 1). By minimizing the following six objective functions, G, H, DX, and DZ achieve the optimal configuration.
Z ~ = H ( X )
X ~ = G ( Z )
m i n D X   L 1 = E X P ( X ) D X 2 ( X ) + E Z P ( Z ) 1 D X ( X ~ ) 2 m i n D Z   L 2 = E X P ( X ) 1 D Z ( Z ~ ) 2 + E Z P ( Z ) D Z 2 ( Z ) m i n G   L 3 = E Z P ( Z ) D X 2 ( X ~ ) m i n H   L 4 = E X P ( X ) D z 2 ( Z ~ ) m i n G , H   L 5 = E Z P ( Z ) Z H ( G ( Z ) ) 2 2 m i n G , H   L 6 = E X P ( X ) X G ( H ( X ) ) 2 2
where  L 1  and  L 2  aim to obtain well-performed discriminators;  L 3  and  L 4  supervise  G  and  H  to generate high-quality data, namely  Z ~  and  X ~ ;   L 5  and  L 6  measure the cycle consistency. It is noted that when  L 3  and  L 4  decrease, the generated data will move away from the decision boundary and stay in high-density areas, which will produce data with high confidence but of low quality.

3. RTAE Method

3.1. Intrinsic Feature Extracted by Auto-Encoder

The collected wind turbine blade cracking data are usually high dimensional (75-dimension in the case study). Directly inputting such data into roundtrip will bring big H and G networks, that is, a large number of hidden layers and hidden neurons. Generally, it is difficult to design such effective networks and configure them. Therefore, reducing the dimensionality before inputting into roundtrip is urgently necessary.
Auto-encoder (AE) is an outstanding dimension reduction method and effective for nonlinear problems [21]. It consists of an encoder and a decoder. The encoder encodes input into a feature representation and decoder reconstructs the input. Given the observations  Y R n , the encoder computes low dimensional representation by
Y ~ = f ( W E Y + b E )
where  f ·  is activation function;  W E R m × n  is an encoder matrix and  b E R  is a bias;  n  is the dimension of original sample. Decoder reconstructs it by:
Y = f ( W D Y ~ + b D )
where  W D R n × m  is a decoder matrix;  b D R  is a bias;  m  is the low dimension. AE learns these matrixes and biases by
a r g m i n θ   1 2 K i = 1 K   Y ~ i Y i 2 2 , θ = W E , b E , W D , b D , i = 1,2 , , K
where  K  denotes the sample amount.

3.2. Implementation of Roundtrip Density Estimation

For roundtrip, H, G, DX and DZ are the key components. Based on the dataset used in the data analysis part (each sample contains 75 attributes), we designed these four models as listed in Table 1. As shown, all of these four models contain five layers, with an alternating fully connected layer and batch normalization.
For the fully connected layer, the computation formula is as follows:
X o u t = W X i n + b
where  X i n R I P  denotes input for the current layer;  X o u t R O P  denotes output of the current layer;  W R O P × I P  represents a transformation matrix;  b  is a bias;  I P  and  O P  are shown in Table 1.
For batch normalization, the computation formula is:
μ 1 m i = 1 K   X i j σ 2 1 K i = 1 K   X i j μ 2 X ~ i j X i j μ σ 2 + ϵ X ^ i j γ X ~ i j + β
where  X  denotes the input batch;  X i j  denotes the  j th dimension of  i th sample;  μ  denotes the average value of  j th dimension over all the samples in a batch;  σ 2  denotes the variance for the  j th dimension;  ϵ  is a small value avoiding denominator to be zero;  γ  and  β  are scale and translation factors.  X ^ i j  is the output of batch normalization for the  j th dimension.

3.3. Alternating Gradient Descent-Based Training Process

To train H, G, DX and DZ, we adopt an alternating gradient descent method. First, for DX and DZ, we train them by the following formula:
θ D X = θ D X l r L 1 θ D X
θ D Z = θ D Z l r L 2 θ D Z
where  θ D X  and  θ D Z  are, respectively, the parameters from the network in Table 1. Specifically, they are the weights and bias from (12);  l r  is the learning rate;  L 1  and  L 2  are the loss function in Equation (8).
Second, to optimize G and H synchronously, we rewrite the loss function in Equation (8) as:
min G , H L = L 3 + L 4 + α ( L 5 + L 6 )
where  L  is the overall loss function;  α  is a pre-specified coefficient. Finally, the weights and bias in  G  and  H  are updated by the following equation:
θ G , H = θ G , H l r L θ G , H

3.4. Overall Framework

The overall framework of the proposed RTAE method is shown in Figure 3. The specific steps are:
(1)
Obtain normal and cracking samples;
(2)
Input both cracking and normal samples into auto-encoder for dimension reduction;
(3)
Input low-dimensional representation cracking samples into roundtrip for data augmentation;
(4)
Generate cracking samples and fuse them with these original samples (after dimension reduction);
(5)
Implement classification result analysis.
The detailed steps are listed in Table 2. It is noted that we can compare the results ‘before data augmentation’ and ‘after data augmentation’ to verify the effectiveness of RTAE.

4. Engineering Applications

4.1. Description of Blading Cracking Dataset

The datasets used for analysis were collected from the CCF Big Data & Computing Intelligence Contest (https://www.datafountain.cn/competitions/302/datasets (accessed on 28 June 2018)). These data were collected by supervisory control and data acquisition system under running status. Samples are collected every 10 min, with each sample containing 75 attributes, such as blade angle, over-speed sensor value, x-axis vibration, inverter power voltage, etc. Details can be found in Table 3. There are in total 48,340 samples collected during approximately one year with a majority of normal samples. To verify the effectiveness under a different imbalance ratio, we constructed 10 datasets as listed in Table 4. We randomly selected 2000 normal samples and varying number of cracking samples. The sign ‘#’ in Table 4 denotes the sample amount. The imbalance ratio varies from 20:1 to 2:1.
Figure 4 shows two normal samples as well as two cracking samples. For each status, these two samples are collected from two different time points. Obviously, it is difficult to distinguish blade status with only one or just a few attributes, since samples from different time points present different value distributions. From such results, we can infer that samples from the same class distribute over more than one cluster. Applying SMOTE or GAN to enhance the cracking sample amount may fail.

4.2. Sample Balanced by RTAE

Due to different units, the attribute ranges vary significantly. Normalization is used to scale the range of each attribute to [0, 1]. The normalization method is presented below:
A ~ = A A m i n A m a x A m i n
where  A ~  is the normalized attribute value;  A m i n  and  A m a x  are, respectively, the minimum and maximum values among all the samples.
Auto-encoder is adopted to extract intrinsic features and reduce the dimension. Wind_50 dataset is used to train an AE model, with its structure [75, 20, 10]. The encoder include two hidden layers, with twenty units for the first layer and ten units for the second layer. Sigmoid is taken as the activation function. For the decoder, it has the inverse structure, namely, as shown in [10, 20, 75]. Figure 5 shows four samples and their reconstructed results. Obviously, the reconstructed data approximate original samples, implying the intrinsic feature can adequately represent the key information.
Roundtrip is employed to generate virtual cracking samples. Table 5 lists detailed hyper-parameter settings. These parameters are determined by trial and error. To observe the results of roundtrip, loss values during 5000 iterations are shown in Figure 6. L1 and L2 in Equation (8) are relevant to the generator, thus they are added together. Both the losses of the generator and discriminator go down very fast and converge after 2000 iterations. In the last 100 iterations (see the inset window), both of them fluctuate to a small extent. When the red line goes down, blue line will go up, implying adversarial learning. To observe generated data, TSNE [22] is adopted to map the intrinsic features into two-dimension space. The results are shown in Figure 7. The generated data well fit the original data distribution and many samples lie on the border, which is helpful for improving classification accuracy.

4.3. Effectiveness of Data Augmentation

Table 6 lists the comparisons between, with and without data augmentation. Decision tree is adopted to classify normal and cracking samples [23]. Three metrics are adopted to evaluate the performance, namely precision, recall and F1-score. Detailed metric computations can be found in the literature [24]. For normal samples, the three metrics with data augmentation are slightly lower than or equal to those without data augmentation. However, for cracking samples, these metrics are higher. For instance, ‘Wind_35′ achieves precision (0.73), recall (0.76) and F1-score (0.75) with data augmentation, which are slightly lower than that without data augmentation (precision: 0.74, recall: 0.78, F1-score: 0.76). The recall value improves from 0.22 to 0.26 and F1-score increases from 0.23 to 0.25, implying the effectiveness of data augmentation. With the data augmentation approach, the average value of the cracking samples are 0.133, 0.135 and 0.135, meaning 19.8%, 23.8% and 22.7% improvement on these three metrics.
We draw the results with data augmentation in histograms, as shown in Figure 8. For normal samples, these three metrics draw down. This is because too few cracking samples will be overwhelmed and nearly all these samples are classified into normal samples. Therefore, from Wind_5 to Wind_35, the recall value is bigger than precision (note that the F1-score is determined by the former two metrics). With an increase in the amount of cracking samples, some normal samples are misclassified into cracking samples, causing a small recall value. For cracking samples, with the imbalance ratio alleviating, all the three metrics increase. It can be concluded that RTAE is effective for improving the recognition of wind blade cracking samples.

4.4. Comparison with Other Imbalance Learning Methods

TRAE is compared to other two representative methods, GAN and SMOTE. Likewise, three metrics are used to evaluate the performance. For SMOTE, the number of the neighbor is set as 3, which is a commonly used value. For GAN, generator network and discriminator network are set as the same structure as in the literature [25]. The comparisons are listed in Table 7. RTAE outperforms GAN and SMOTE under all the 10 datasets. In particular, for Wind_35, Wind_40 and Wind_45, RTAE is far better than GAN and SMOTE. The main reason is that GAN focuses on ‘real’ or ‘fake’, which generates samples with high confidence, but low usage for improving the recognition of cracking samples. SMOTE cannot handle multi clusters. Therefore, we use a small neighbor number (i.e., 3) to alleviate such a situation. However, such a small value may make it pick the basis sample from the middle or a dense area, leading to many synthesized samples lying far from border.

5. Conclusions and Future Work

Aiming at wind turbine blade cracking recognition under imbalanced data, this paper designs a novel roundtrip auto-encoder method. Two generator networks and two discriminator networks are designed to ensure the generated samples well fit the distribution of historical cracking samples. Auto-encoder method is applied to reduce the dimension of historical samples and thus the complexity of generator and discriminator. From the reconstructed results, it is concluded that auto-encoder is effective for extracting low-dimensional intrinsic features. From the results under different imbalance levels, the detection performance shows significant improvement with the RTAE method. When the imbalance level is above 4:1, all the cracking samples cannot be identified without data augmentation, which shows the importance of considering the imbalance problem in real application.
Analysis of the real wind turbine blade cracking data is carried out, the recognition of cracking samples improves by 19.8%, 23.8% and 22.7% on precision, recall and F1-score. The cracking detection under imbalanced data and the comparisons show that: (1) compared to the popular data-level methods, RTAE is superior under the influence of highly imbalanced data; (2) through integrating auto-encoder and roundtrip model, the framework provides a possibility for solving imbalance problem with high dimension data.
As is known, a wind turbine blade contains many types of cracks in real industrial applications. Currently, subject to roundtrip, RTAE can only deal with the binary-class imbalanced data problems. Multi-class cracking problems are more complex as they may contain more than one minority classes. In the future work, we will extend the application of RTAE and apply it to multi-class problem. Meanwhile, we will explore more effective roundtrip structures and apply them to high-dimensional time-series data.

Author Contributions

Conceptualization, Y.Z. (Yuyan Zhang); methodology, Y.Z. (Yafeng Zhang); software, L.Y.; vali-dation, Y.Z. (Yuyan Zhang) and Y.Z. (Yafeng Zhang); formal analysis, Y.Z. (Yuyan Zhang) and H.W.; investigation, Y.Z. (Yuyan Zhang) and H.W.; resources, H.L.; data curation, H.L.; writ-ing—original draft preparation, Y.Z. (Yuyan Zhang); writing—review and editing, Y.Z. (Yafeng Zhang); visualization, X.W.; supervision, H.L.; project administration, X.W.; funding acquisition, Y.Z. (Yuyan Zhang). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (52105536); the Guangdong Basic and Applied Basic Research Fund (2022A1515140066); the Henan Provincial Key R&D and Promotion Special Project (Science and Technology Research) (232102221009).

Data Availability Statement

Data are contained within the article. The data presented in this study are available in Section 4.1.

Conflicts of Interest

Author Lingdi Yan was employed by the company CITIC Heavy Industries Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Din, I.U.; Hao, P.; Franz, G.; Panier, S. Elastoplastic CDM model based on Puck’s theory for the prediction of mechanical behavior of Fiber Reinforced Polymer (FRP) composites. Compos. Struct. 2018, 201, 291–302. [Google Scholar] [CrossRef]
  2. Din, I.U.; Panier, S.; Hao, P.; Franz, G.; Bijwe, J.; Hui, L. Finite element modeling of indentation and adhesive wear in sliding of carbon fiber reinforced thermoplastic polymer against metallic counterpart. Tribol. Int. 2019, 135, 200–212. [Google Scholar] [CrossRef]
  3. Cao, Z.; Li, S.; Li, C.; Li, P.; Ko, T.J. Formation mechanism and detection and evaluation methods as well as repair technology of crack damage in fiber-reinforced composite wind turbine blade: A review. Int. J. Adv. Manuf. Technol. 2022, 120, 5649–5672. [Google Scholar] [CrossRef]
  4. Lu, Q.; Ye, W.; Yin, L. ResDenIncepNet-CBAM with principal component analysis for wind turbine blade cracking fault prediction with only short time scale SCADA data. Measurement 2023, 212, 112696. [Google Scholar] [CrossRef]
  5. Katsaprakakis, D.A.; Papadakis, N.; Ntintakis, I. A comprehensive analysis of wind turbine blade damage. Energies 2021, 14, 5974. [Google Scholar] [CrossRef]
  6. Reddy, A.; Indragandhi, V.; Ravi, L.; Subramaniyaswamy, V. Detection of Cracks and damage in wind turbine blades using artificial intelligence-based image analytics. Measurement 2019, 147, 106823. [Google Scholar] [CrossRef]
  7. Ge, Y.; Yue, D.; Chen, L. Prediction of wind turbine blades icing based on MBK-SMOTE and random forest in imbalanced data set. In Proceedings of the 2017 IEEE Conference on Energy Internet and Energy System Integration (EI2), Beijing, China, 26–28 November 2017. [Google Scholar]
  8. Jiang, N.; Li, N. A wind turbine frequent principal fault detection and localization approach with imbalanced data using an improved synthetic oversampling technique. Int. J. Electr. Power Energy Syst. 2021, 126, 106595. [Google Scholar] [CrossRef]
  9. Velandia-Cardenas, C.; Vidal, Y.; Pozo, F. Wind turbine fault detection using highly imbalanced real SCADA data. Energies 2021, 14, 1728. [Google Scholar] [CrossRef]
  10. Yi, H.; Jiang, Q.; Yan, X.; Wang, B. Imbalanced classification based on minority clustering synthetic minority oversampling technique with wind turbine fault detection application. IEEE Trans. Ind. Inform. 2020, 17, 5867–5875. [Google Scholar] [CrossRef]
  11. Chen, W.; Zhou, H.; Cheng, L.; Xia, M. Wind Turbine Blade Icing Diagnosis Using Convolutional LSTM-GRU With Improved African Vultures Optimization. IEEE Open J. Instrum. Meas. 2022, 1, 1–9. [Google Scholar] [CrossRef]
  12. Chen, P.; Li, Y.; Wang, K.; Zuo, M.J.; Heyns, P.S.; Baggeröhr, S. A threshold self-setting condition monitoring scheme for wind turbine generator bearings based on deep convolutional generative adversarial networks. Measurement 2021, 167, 108234. [Google Scholar] [CrossRef]
  13. Liu, J.; Qu, F.; Hong, X.; Zhang, H. A small-sample wind turbine fault detection method with synthetic fault data using generative adversarial nets. IEEE Trans. Ind. Inform. 2018, 15, 3877–3888. [Google Scholar] [CrossRef]
  14. Liu, X.; Teng, W.; Wu, S.; Wu, X.; Liu, Y.; Ma, Z. Sparse dictionary learning based adversarial variational auto-encoders for fault identification of wind turbines. Measurement 2021, 183, 109810. [Google Scholar] [CrossRef]
  15. Wang, A.; Qian, Z.; Pei, Y.; Jing, B. A de-ambiguous condition monitoring scheme for wind turbines using least squares generative adversarial networks. Renew. Energy 2022, 185, 267–279. [Google Scholar] [CrossRef]
  16. Yang, S.; Zhou, Y.; Chen, X.; Deng, C.; Li, C. Fault diagnosis of wind turbines with generative adversarial network-based oversampling method. Meas. Sci. Technol. 2023, 34, 044004. [Google Scholar] [CrossRef]
  17. Ogaili AA, F.; Jaber, A.A.; Hamzah, M.N. Wind turbine blades fault diagnosis based on vibration dataset analysis. Data Brief 2023, 49, 109414. [Google Scholar] [CrossRef] [PubMed]
  18. Jiang, W.; Jin, J. Intelligent icing detection model of wind turbine blades based on scada data. arXiv 2021, arXiv:2101.07914. [Google Scholar]
  19. Zou, L.; Cheng, H. Research on wind turbine blade surface damage identification based on improved convolution neural network. Appl. Sci. 2022, 12, 9338. [Google Scholar] [CrossRef]
  20. Liu, Q.; Xu, J.; Jiang, R.; Wong, W.H. Density estimation using deep generative neural networks. Proc. Natl. Acad. Sci. USA 2021, 118, e2101344118. [Google Scholar] [CrossRef]
  21. Zhang, Y.; Li, X.; Gao, L.; Chen, W.; Li, P. Ensemble deep contractive auto-encoders for intelligent fault diagnosis of machines under noisy environment. Knowl.-Based Syst. 2020, 196, 105764. [Google Scholar] [CrossRef]
  22. Anowar, F.; Sadaoui, S.; Selim, B. Conceptual and empirical comparison of dimensionality reduction algorithms (pca, kpca, lda, mds, svd, lle, isomap, le, ica, t-sne). Comput. Sci. Rev. 2021, 40, 100378. [Google Scholar] [CrossRef]
  23. Kherif, O.; Benmahamed, Y.; Teguar, M.; Boubakeur, A.; Ghoneim, S.S.M. Accuracy improvement of power transformer faults diagnostic using KNN classifier with decision tree principle. IEEE Access 2021, 9, 81693–81701. [Google Scholar] [CrossRef]
  24. Zhang, Y.; Li, X.; Gao, L.; Wang, L.; Wen, L. Imbalanced data fault diagnosis of rotating machinery using synthetic oversampling and feature learning. J. Manuf. Syst. 2018, 48, 34–50. [Google Scholar] [CrossRef]
  25. Zhang, Y.; Zhang, Y.; Sun, C.; Wang, H.; Wen, X.; Qiao, D.; Yan, X.; Li, H. Identification method of cracking state of wind turbine blade based on GAN under imbalanced samples. Comput. Integr. Manuf. Syst. 2023, 29, 532–543. [Google Scholar]
Figure 1. Illustration of problem about sampling strategy and generative strategy.
Figure 1. Illustration of problem about sampling strategy and generative strategy.
Applsci 13 11628 g001
Figure 2. Overall framework of roundtrip.
Figure 2. Overall framework of roundtrip.
Applsci 13 11628 g002
Figure 3. Overall framework of RTAE.
Figure 3. Overall framework of RTAE.
Applsci 13 11628 g003
Figure 4. Normal sample (first row) and cracking sample (second row).
Figure 4. Normal sample (first row) and cracking sample (second row).
Applsci 13 11628 g004
Figure 5. Four original samples and their reconstructed result.
Figure 5. Four original samples and their reconstructed result.
Applsci 13 11628 g005
Figure 6. Loss value change during the iteration.
Figure 6. Loss value change during the iteration.
Applsci 13 11628 g006
Figure 7. Four original samples and their reconstructed results.
Figure 7. Four original samples and their reconstructed results.
Applsci 13 11628 g007
Figure 8. Result with data augmentation. (a) Normal Sample. (b) Cracking Sample.
Figure 8. Result with data augmentation. (a) Normal Sample. (b) Cracking Sample.
Applsci 13 11628 g008
Table 1. Details of H, G, DX and DZ.
Table 1. Details of H, G, DX and DZ.
PartLayer TypeInput Dimension (IP)Output Dimension (OP)
HFully Connected LayerDetermined by AE32
Batch Normalization3232
Fully Connected Layer3264
Batch Normalization6464
Fully Connected Layer64Dimension of Latent Variable
GFully Connected LayerDimension of Latent Variable64
Batch Normalization6464
Fully Connected Layer6432
Batch Normalization3232
Fully Connected Layer32Determined by AE
DXFully Connected LayerDetermined by AE32
Batch Normalization3232
Fully Connected Layer3216
Batch Normalization1616
Fully Connected Layer162
DZFully Connected LayerDimension of Latent Variable20
Batch Normalization2020
Fully Connected Layer2010
Batch Normalization1010
Fully Connected Layer102
Table 2. Detailed steps of RTAE method.
Table 2. Detailed steps of RTAE method.
Input: Original   Samples   Y = Y 1 , , Y K  Contain Both Normal and Cracking Samples
Step1: Minimize  a r g m i n θ   1 2 K i = 1 K   Y ~ i Y i 2 2  to obtain  W E , b E  
Step2: According to  Y ~ = f ( W E Y + b E ) , compute low-dimensional representation.
Step3: For cracking and normal samples, we name them as  Y ~ m i n  and  Y ~ m a j ,  respectively.
Step4: Replace  X  in Figure 2 by  Y ~ m i n Z  is set as the same dimension as  X . Each dimension of  Z  is a random variable drawn from normal distribution with mean value 0.5, standard deviation 0.05.
Step5: Train H, G, DX and DZ by (14), (15) and (17).
Step6: Compute  X ~ = G ( H ( X ) ) .
Step7: Implement 5-fold cross-validation and compute the average validation result.
Step8: Take the result only on  Y ~ m i n , Y ~ m a j  as ‘before data augmentation’. For each fold train process, adding  X ~  to the train set and take the result as ‘after data augmentation’.
Table 3. Detailed attributes of the used data.
Table 3. Detailed attributes of the used data.
No.AttributeNo.AttributeNo.Attribute
1Hub speed26Inverter inlet temperature51Variable motor 1 Power estimation
2Hub angle27Converter outlet temperature52Variable motor 2 Power estimation
3Blade 1 angle28Inverter inlet pressure53Variable motor 3 Power estimation
4Blade 2 angle29Converter outlet pressure54Current status of the blade
5Blade 3 angle30Generator power limiting value55Current hub status value
6Variable pitch motor1 Current31Reactive power set value56Yaw status value
7Variable pitch motor2 Current32Rated hub speed57Yaw requirement value
8Variable pitch motor3 Current33Ambient temperature of the wind tower58Blade 1 Battery box temperature
9Detection value of the overspeed sensor34Generator stator temperature 159Blade 2 Battery box temperature
10Average of 5 s yaw against wind35Generator stator temperature 260Blade 3 Battery box temperature
11Vibration value in x direction36Generator stator temperature 361Blade 1 Variable motor temperature
12Vibration value in y direction37Generator stator temperature 462Blade 2 Variable motor temperature
13Hydraulic brake pressure38Generator stator temperature 563Blade 3 Variable motor temperature
14Cabin weather station wind speed39Generator stator temperature 664Blade 1 Inverter box temperature
15Absolute wind direction40Generator air temperature 165Blade 2 Inverter box temperature
16Atmospheric pressure41Generator air temperature 266Blade 3 Inverter box temperature
17Reactive power control status42Main bearing temperature 167Blade 1 Super-capacitor voltage
18Inverter power grid side current43Main bearing temperature 268Blade 2 Super-capacitor voltage
19Inverter power grid side voltage44Hub temperature69Blade 3 Super-capacitor voltage
20Active power of inverter power grid side45Hub control cabinet temperature70Drive 1 Thyristor temperature
21Inverter grid side reactive power46Cabin temperature71Drive 2 Thyristor temperature
22Inverter generator side power47Control cabinet temperature72Drive 3 Thyristor temperature
23Generator operating frequency48Inverter INU temperature73Drive 1 Output torque
24Generator current49Inverter ISU temperature74Drive 2 Output torque
25Generator torque50Inverter INU RMIO temperature75Drive 3 Output torque
Table 4. Details of 10 datasets.
Table 4. Details of 10 datasets.
No.Dataset# Normal# CrackingImbalance Ratio
1Wind_5200010020:1
2Wind_10200020010:1
3Wind_1520003006.67:1
4Wind_2020004005:1
5Wind_2520005004:1
6Wind_3020006003.33:1
7Wind_3520007002.86:1
8Wind_4020008002.5:1
9Wind_4520009002.22:1
10Wind_50200010002:1
Table 5. Details of hyper-parameter values.
Table 5. Details of hyper-parameter values.
Hyper ParameterSymbolValue
Factor in (13)   ε 0.001
Scale factor in (13)   γ 1
Translation factor in (13)   β 0
Learning rate in (14), (15) and (17)lr0.0002
Weight in (16)α10
Iteration Number-5000
Table 6. Comparison between, with and without data augmentation.
Table 6. Comparison between, with and without data augmentation.
DatasetWithout Data AugmentationWith Data Augmentation
PrecisionRecallF1-ScorePrecisionRecallF1-Score
Wind_5Normal0.950.970.960.950.960.96
Cracking0000.020.020.02
Wind_10Normal0.900.930.910.900.910.91
Cracking0000.020.020.02
Wind_15Normal0.860.890.870.850.850.85
Cracking0000.020.020.02
Wind_20Normal0.810.850.830.810.850.83
Cracking0000.040.030.04
Wind_25Normal0.770.830.800.770.820.79
Cracking0000.050.040.05
Wind_30Normal0.790.810.780.760.770.76
Cracking0.210.170.190.180.170.18
Wind_35Normal0.740.780.760.730.760.75
Cracking0.250.220.230.240.260.25
Wind_40Normal0.690.690.690.710.710.71
Cracking0.210.210.210.270.260.26
Wind_45Normal0.660.620.640.660.640.65
Cracking0.250.290.270.260.280.27
Wind_50Normal0.590.580.580.610.590.60
Cracking0.190.200.200.230.250.24
AverageCracking0.1110.1090.110.1330.1350.135
Table 7. Comparison with GAN and SOMTE.
Table 7. Comparison with GAN and SOMTE.
DatasetRTAEGANSMOTE
PrecisionRecallF1PrecisionRecallF1PrecisionRecallF1
Wind_5Cracking0.020.020.02000000
Wind_10Cracking0.020.020.02000000
Wind_15Cracking0.020.020.02000000
Wind_20Cracking0.040.030.04000000
Wind_25Cracking0.050.040.05000000
Wind_30Cracking0.180.170.180.160.170.160.150.170.16
Wind_35Cracking0.240.260.250.220.220.220.220.220.22
Wind_40Cracking0.270.260.260.210.220.210.210.210.21
Wind_45Cracking0.260.280.270.240.270.250.230.250.24
Wind_50Cracking0.230.250.240.180.190.180.190.190.19
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Zhang, Y.; Li, H.; Yan, L.; Wen, X.; Wang, H. Wind Turbine Blade Cracking Detection under Imbalanced Data Using a Novel Roundtrip Auto-Encoder Approach. Appl. Sci. 2023, 13, 11628. https://doi.org/10.3390/app132111628

AMA Style

Zhang Y, Zhang Y, Li H, Yan L, Wen X, Wang H. Wind Turbine Blade Cracking Detection under Imbalanced Data Using a Novel Roundtrip Auto-Encoder Approach. Applied Sciences. 2023; 13(21):11628. https://doi.org/10.3390/app132111628

Chicago/Turabian Style

Zhang, Yuyan, Yafeng Zhang, Hao Li, Lingdi Yan, Xiaoyu Wen, and Haoqi Wang. 2023. "Wind Turbine Blade Cracking Detection under Imbalanced Data Using a Novel Roundtrip Auto-Encoder Approach" Applied Sciences 13, no. 21: 11628. https://doi.org/10.3390/app132111628

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop