Next Article in Journal
Research on Economic Evaluation Methods of Offshore Oil Multi-Platform Interconnected Power System Considering Petroleum Production Characteristics
Next Article in Special Issue
Integrated Algorithm for Selecting the Location and Control of Energy Storage Units to Improve the Voltage Level in Distribution Grids
Previous Article in Journal
Method for Building Information Modeling Supported Project Control of Nearly Zero-Energy Building Delivery
Previous Article in Special Issue
Optimal Sizing of a Real Remote Japanese Microgrid with Sea Water Electrolysis Plant Under Time-Based Demand Response Programs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved Weighted k-Nearest Neighbor Based on PSO for Wind Power System State Recognition

Department of Electrical Engineering, Chung Yuan Christian University, No. 200, Zhongbei Road, Zhongli District, Taoyuan City 320, Taiwan
*
Author to whom correspondence should be addressed.
Energies 2020, 13(20), 5520; https://doi.org/10.3390/en13205520
Submission received: 27 August 2020 / Revised: 12 October 2020 / Accepted: 19 October 2020 / Published: 21 October 2020
(This article belongs to the Special Issue Electronic Power and Energy Systems)

Abstract

:
In this paper, we propose using particle swarm optimization (PSO) which can improve weighted k-nearest neighbors (PWKNN) to diagnose the failure of a wind power system. PWKNN adjusts weight to correctly reflect the importance of features and uses the distance judgment strategy to figure out the identical probability of multi-label classification. The PSO optimizes the weight and parameter k of PWKNN. This testing is based on four classified conditions of the 300 W wind generator which include healthy, loss of lubrication in the gearbox, angular misaligned rotor, and bearing fault. Current signals are used to measure the conditions. This testing tends to establish a feature database that makes up or trains classifiers through feature extraction. Not lowering the classification accuracy, the correlation coefficient of feature selection is applied to eliminate irrelevant features and to diminish the runtime of classifiers. A comparison with other traditional classifiers, i.e., backpropagation neural network (BPNN), k-nearest neighbor (k-NN), and radial basis function network (RBFN) shows that PWKNN has a higher classification accuracy. The feature selection can diminish the average features from 16 to 2.8 and can reduce the runtime by 61%. This testing can classify these four conditions accurately without being affected by noise and it can reach an accuracy of 83% in the condition of signal-to-noise ratio (SNR) is 20dB. The results show that the PWKNN approach is capable of diagnosing the failure of a wind power system.

1. Introduction

With the emergence of green energy, wind power plays a principle role in energy. However, wind power is a form of intermittent energy without a stable output. In addition, the malfunction of wind generators is a great concern. Downtime leads to heavy cost losses in peak-wind seasons [1,2,3]. Furthermore, a large number of wind generators in wind farms increases the probability of failure. Without sufficient personnel, an automatic inspection of the wind generator is indispensable [4]. A central system is employed to monitor the operation of wind generators, which can effectively reduce personnel cost. According to [5,6], the most common fault condition was bearing damage, accounting for 45%; followed by stator failure, accounting for 35%; and finally, rotor failure, accounting for 10%; and other damages are 10%. When the motor fails, the running cost increases. Therefore, fault identification has always been one of the topics in industrial applications. Generally, the automatic identification system includes the following three parts: the signal analysis, feature extraction, and condition classification. The signal analysis has been carefully researched in the past, in which the feature extraction and condition classification are crucial factors [7,8]. Feature extraction involves extracting specific and representative features, providing few data to classifiers in classification. However, the features extracted are not always the effective ones. The valid features are likely to become invalid features if the signals include noise. This research emphasizes how to gain effective features and raise the classification accuracy [9]. Artificial neural network (ANN) has been widely used in different fields, such as machine operation control, automatic fault detection, and image recognition [6,10,11,12,13,14]. Since the first neural network was proposed, different types of networks have been developed one after another, such as feedforward neural network, probabilistic neural network (PNN) [10,11], and backpropagation neural network (BPNN) [6,12,13,14]. The use of different classifiers for various types of issues have different performance, architecture, and classification effects. BPNN is a supervised learning network, and its recognition ability is higher than the unsupervised learning network. Furthermore, classifiers offer a method of automatic classification. A well-organized classifier can provide a higher classification accuracy, which takes less time for computing, increases computing efficiency, and favors a large quantity, real-time system.
There have been considerable studies on the signal analysis of rotating machines, of which the vibration signal analysis was the focus [1,15]. Vibration signals can show the machine operation clearly. While measuring vibration signals, an accelerometer is a compulsory supplement. Additionally, one accelerometer can merely measure a single position’s vibration. Recently, more and more researches have used the fault diagnosis of electric currents and voltage [16,17,18]. Electric currents and voltage can lessen the number and cost of sensors, complexity of measurement system, and failure rate.
The signal analysis can detect the unusual signals; the traditional signal analysis is Fourier transforms (FT) [19] which cannot obtain transient messages. However, rotating machines often have intermittent messages. A multi-resolution analysis (MRA) can be a good method of analyzing unusual signals [20]. Features extracted from signals in automatic classification can reduce the input number of classifiers and computing time. Feature extraction probably includes invalid features and its number affects computing. In addition, invalid features lower the classification accuracy of classifiers [21,22,23].
Although k-nearest neighbor (k-NN) is a simple and fast classifier, its procedure and structure need some improvement. The greedy method, used in [24], is to adjust k-NN with weight features. In [25], the k parameters should be adjusted according to different categories of training sets. Additionally, the setting of parameter k and identical probability of classification remain to be solved.
In this paper, we propose one approach to recognize the operational conditions of power generators. The approach is based on structure and consists of the following two parts: First, the feature extraction technique is used, where the wavelet transform method extracts the features from the 300 W power system. Second, the correlation coefficients are used to optimize the feature set. Finally, the proposed method is used to recognize the operation of the wind power system. The proposed method combines k-NN and PSO which can improve the accuracy, namely PWKNN. Experiments show that the performance of the PWKNN model is better than the other classification methods, and the average classification accuracy of PWKNN is higher than that of other classifiers.
This paper is organized as follows: In Section 2, we introduce the signal analysis and feature extraction approach; in Section 3, we present the procedure and structure of improved weight k-NN based on particle swarm optimization (PWKNN); in Section 4, we discuss the types of malfunctions of a wind power system, the feature selection; in Section 5, we present the simulation results; and in Section 6, we provide the conclusions.

2. Signal Processing and Feature Extraction

2.1. Wavelet Transform

The French geophysicist Jean Morlet invented the wavelet transform for analyzing local properties of seismic waves [26]. He found that the fast Fourier transform (FFT) could not meet the requirements, therefore, he introduced a wavelet into the signal analysis and decomposed signals. The wavelet transform of signals can be observed at different resolutions. Therefore, local properties can be examined effectively. It is a signal transform analysis between time and frequency domains.

Continuous Wavelet Transform

The scale parameter and translation parameter can adjust the mother wavelet ψ ( t ) shown in Equation (1), where a is the scale parameter and b is the translation parameter.
ψ ( t ) = 1 a ψ ( t b a )
The function f ( t ) of continuous wavelet transform is defined in Equation (2), where the asterisk represents the operation of complex conjugate.
C W T ( a , b ) = 1 a f ( t ) ψ * ( t b a ) d t

2.2. Multi-Resolution Analysis

The function f ( t ) is decomposed by the scaling function and the mother wavelet function through the scaling change and wavelet transform. The scaling function is considered to be a low-pass filter and the mother wavelet function is considered to be a high-pass filter. The structure shown in Figure 1 decomposes the signal f ( t ) into a high- and low-frequency signal. The low-frequency signal is decomposed for the second time and the decomposed signals,   d 1 ,   ,   d n and a n can be yielded as follows:
f ( t ) = d 1 + d 2 + + d n + a n
The procedure mentioned above is an MRA [26]. Through the viewpoint of spectral spectrum, the MRA can process the detailed analysis of low-frequency signals shown in Figure 2. Therefore, more decomposition implies higher resolution.

2.3. Feature Extraction

Scales of wavelet coefficients can show the signals in Equation (3). By using Parseval’s theorem, Equation (3) can be reconstructed as follows:
n = | f ( t ) | 2 d t = l = | a j ( l ) | 2 + j = 0 z = | d j ( z ) | 2
The first component refers to the average power of a signal. The latter component is the sum of the average power of coefficients after wavelet transformation. The simplified formula is shown as follows:
E f = E a j + j = 0 j 1 E d j
The energy spectrum indicates the energy for each frequency band of signal. The different distributions of energy spectrum can easily show operational conditions of wind generator. Hence, the energy spectrums are considered to be the features in this paper.

3. PWKNN and Feature Selection

3.1. PWKNN

The k-NN is a supervised learning method and also among the simplest of machine learning algorithms [27]. The method classifies an unknown sample class according to the nearest k samples of training datasets. The nearest k samples are called k-nearest neighbors which can be obtained through the distance computation. The class having the maximum number is considered to be the unknown category. The classification decision is as follows:
D ( x i ) = a r g m a x m t j k N N y ( t j , c m )
The equation means that the predictor will be the class that has the largest number of members in the k-nearest neighbors. For example, the left part of Figure 3 indicates three classes of samples (square, circle, and triangle) and an unknown sample (rhombus). The nearest four samples are selected. The numeric means the order of the distance 1 is the nearest and 4 is the farthest. The figure shows that Class A is a strong probability. Therefore, the unknown sample is classified to Class A.
K-NN employs a distance measure, usually Euclidean distance, to show the similarity. If X = ( x 1 , x 2 ,   ,   x n )   and   Y = ( y 1 , y 2 ,   ,   y n ) are n-dimensions vectors, the following shows how Euclidean distance computes:
d i s t ( X , Y ) = ( x 1 y 1 ) 2 + ( x 2 y 2 ) 2 + + ( x n y n ) 2
However, k-NN has some weakness as follows: (1) All feature weights are equal. Not all features to the classified results have the positive correlation. The invalid features lead to the error classified results. (2) Identical classified probabilities probably occur. As shown on the left side of the Figure 3, if a nearest neighbor is added to the classification decisions, the probabilities of Class A, Class B, and Class C are 40%, 40%, and 20%, respectively. Class A and Class B have the same probability.
Refers to the k-NN weakness (1), the classification result of k-NN is sensitive to features. Invalid features reduce the classification accuracy. The weighted k-NN (WKNN) is implemented to modify this drawback. Different weights are used according to different features. The weight, W = [ w 1 , w 2 ,   ,   w n ] , is taken into account in (x) to compute the weighted distance as follows:
d i s t ( X , Y ) = w 1 ( x 1 y 1 ) 2 + w 2 ( x 2 y 2 ) 2 + + w n ( x n y n ) 2
From Figure 3, the weight is W = [ 1 , 1 ] in the left diagram and the weight is W = [ 1 , 0.5 ] in the right diagram. The equation mentioned above shows that the weight change can influence the classification results. Because, initially, Class A is a high probability, the probability of Class B will be stronger than that of Class A after the changed weight. Feature 2 has less effect on the classification results. The weight adjustment can alter the effect of different features on classification results and eliminate certain invalid features to raise the classification accuracy.
Concerning the classified result of k-NN producing identical probability, this research suggests the distance decision to determine the classification. In the left diagram of Figure 3, Class A and Class B have the identical probability. The method computes the distance from Class A and Class B to each two test points of their k-nearest neighbors. The distances from Class A to its test points of k-nearest neighbors are d1 and d3. The distances from Class B to its test points of k-nearest neighbors are d4 and d5. If D is the decision result, the following shows that the sum of shorter distances to the class is considered as the predictor in classification:
{ D = Class   A   if   d 1 + d 3 < d 4 + d 5 D = Class   B   if   d 1 + d 3 > d 4 + d 5
This method can overcome the identical probability and enhance the correctness of classification.
PWKNN is a classifier of upgrading weighted k-nearest neighbors based on particle swarm optimization (PSO). The PSO optimizes weights and k value and estimates the predictive classification accuracy of PWKNN with leave one out cross-validation (LOOCV). In the LOOCV, each sample data is a class. One class is a testing sample in each computation, others are training samples. All classes are measured, and the sum of classified accuracy are computed as predictive classification accuracy p c a C V . If the number of the correct classification is N C o r r e c t l y and the number of total training samples is N T o t a l , p c a C V is shown as follows:
p c a C V = N C o r r e c t l y N T o t a l × 100 %

3.2. Particle Swarm Optimization (PSO)

PSO is a computational method on optimizing a problem presented by R. Eberhart and J. Kennedy [28]. PSO has become a popular optimization algorithm and widely used in practical problem solving by researchers [29,30,31]. PSO generates particles randomly from the search space. The movement of a particle is influenced by its best position and better positions founded by other particles, which optimizes a solution in the space. The optimal result is that all particles are found in the same position of the search space. This position is the best solution.
Unlike the genetic algorithm (GA), the procedure of PSO works according to a few simple formula and conceptions [28,32]. The object function or fitness function F i t ( ) must be maximized. To avoid the solution converging toward the local best, the suitable number of particles is set, and the position is initialized randomly. Then, the velocity of particle is determined by:
v i ( t + 1 ) = w ( t ) v i ( t ) + ϕ 1 r 1 ( p b e s t i x i ) + ϕ 2 r 2 ( g b e s t x i )
In addition, the right choice of inertial weight provides a balance between global exploration and local exploitation. In general, the inertial weight, w, can be set to change depending on the current number of iterations, t, as follows:
w   =   w m a x     w m a x     w m i n T   ×   t
The equation is influenced by the following three factors: the previous particle’s velocity v i ( t ) , self-best position of particle p b e s t i and the all particle’s best position g b e s t i . Next, update the particle’s position, shown as follows:
x i ( t + 1 ) = x i ( t ) + v i ( t + 1 )
Additionally, the testing judges whether or not the output of fitness function is optimal to replace the particle self-best position. If the output is greater than the previous one, the new position replaces the particle’s best position. If not, the best position of particle remains in the same position. The equation is shown as follows:
p b e s t i ( t + 1 ) = { p b e s t ( t )   if   F i t ( x i ( t + 1 ) ) F i t ( p b e s t i ( t ) ) x i ( t + 1 )   if   F i t ( x i ( t + 1 ) ) > F i t ( p b e s t i ( t ) )
Moreover, the best position of all particles can be found to select the greatest output of fitness function. Finally, the position is considered to be the best solution, and shown in the following equation:
g b e s t ( t + 1 ) = { p b e s t ( t + 1 )   if   max ( F i t ( p b e s t ( t + 1 ) ) ) F i t ( g b e s t ( t ) ) g b e s t ( t )   if   max ( F i t ( p b e s t ( t + 1 ) ) ) > F i t ( g b e s t ( t ) )
The equation of particle movement is to decide the ability of search solution and convergence. The formula mentioned above is commonly used after improvement.
To ensure exploration ability at the start of the algorithm, and then to ensure exploitation ability in the search area. Therefore, the inertia weight factor in the studies is adjusted to linearly decrease. The values wmax and wmin are the upper and lower limits of the inertial weight and are set at 0.9 and 0.4 in (12), respectively. The maximum number of iterations, T, is set in 1000. Additionally, by adjusting the learning coefficients φ 1 and φ 2 , are set as 2.0, the particle self-best position and all particle’s best positions can influence particle movement. The random number is to increase the interference and avoid the solution getting stuck into local best in the optimization process [33].
In this paper, weight W = [ w 1 , w 2 ,   ,   w n ] and k value are taken for particle position x i = [ w 1 , w 2 ,   , w n , k ] . The predictive accuracy of PWKNN P C V is regarded as the fitness function F i t ( x i ) = P C V . The optimization process is shown in Figure 4. A set of optimum parameters [ w 1 , w 2 ,   , w n , k ] b e s t can generate after PSO optimization, which PWKNN can use to classify categories to gain the highest classified correction accuracy.

3.3. Feature Selection

Wavelet coefficients of different features can be taken for the features of generator operation and the input of classifier computing. In addition, correlation coefficients are used to select features to reach optimization and reduce computation. The formula of correlation coefficient is shown as follows:
R = i = 1 n ( x i x ¯ ) ( y i y ¯ ) i = 1 n ( x i x ¯ ) 2 i = 1 n ( y i y ¯ ) 2
This paper determines the feature selection according to correlation coefficients. The predictive classification accuracy of PWKNN is considered to be the stop criterion when all features are selected. Features with minor correlation coefficients are removed first until none is left out. The flowchart of feature selection is shown in Figure 5. The following is the procedure of the feature selection:
Step (1)
Suppose the number of all features is N , select all features and use the cross one out cross-validation (LOOCV) to compute the classification accuracy p c a C V max , N of PWKNN, which is the criterion of raising the classification accuracy. W N is the weight after the PWKNN optimization.
Step (2)
After removing the lowest correlation coefficient, the feature number is N 1 . Use W N to recompute the classification accuracy p c a C V .
Step (3)
If p c a C V p c a C V max , N , go back to Step 2, otherwise, features after removal can gain the optimum weight W N 1 and classification accuracy p c a C V max , N 1 through weight optimization.
Step (4)
If p c a C V max , N 1 p c a C V max , N , go back to Step 2. Record W N 1 and this set of features is optimum; otherwise, come to an end of this procedure.

4. Types of Malfunctions of a Wind Power System

This testing employs a wind power simulation platform, load, data acquisition (DAQ), and personal computer, as shown in Figure 6, as well as a 300 W rated three-phase generator and gearbox with ratio of 1 to 10 driven by a drive motor to simulate wind power. The output of wind power generators links to the load. The load of wye connected single phase is 20 Ω. One phase of current signals can be derived from the NI-PXI 1033 chassis and NI-PXI 4071 digital multimeter (manufactured by National Instruments, Austin, Texas, USA) to the personal computer. The derived signal is segmented with proper length. This research is based on the time of 0.5 s as the length of segment, shown in Figure 7. Finally, we use the computer to analyze and automatically detect the signals.

4.1. Loss of Lubrication in the Gearbox

Lubrication leakage leading to malfunction occurs when a gearbox leaks. Usually, the gearbox is filled with eight-tenths lubrication. However, the vibration of power generators is more serious in real scenarios. Vibration is likely to cause the leak valve to loosen and leakage of lubrication. A gearbox is damaged by deficient lubrication for long periods. This research is to examine the leakage of an empty gearbox causing malfunction. The output current of the power generator is shown in Figure 7b.

4.2. Rotor Angular Misalignment

Rotor angular misalignment occurs when the rotor angles of the power generator and gearbox misalign. Normally, two rotor angles are at the same level, as shown in Figure 8. Nevertheless, a large wind turbine is at high altitudes and easily influenced by an external force. The base probably misaligns due to lengthy external vibration, resulting indirectly in rotor angular misalignment. In this research, rotor angular misalignment occurs when the height of the gearbox is lowered, shown in Figure 9. This testing is to remove the thin plain washer to lower the height. After putting the thick and thin plain washers under the gearbox, its rotor angle reaches the same level as that of the power generator. The height of the thick plain washer is 2.96 mm and the thin plain washer is 2.00 mm. The output current of the power generator is shown in Figure 7c.

4.3. Bearing Fault

The ball bearing, shown in Figure 10a, involves an inner race, an outer race, steel balls, and a retainer. Normally, the steel ball rolls smoothly in the inner raceway and outer raceway. However, the external objects can damage the raceway or steel balls. In addition, the ball bearing can suffer abrasion by operating over a long period of time. This research suggests that the damage of the inner raceway and outer raceway is the malfunction situation. Since the bearing material is steel, electrical discharge machining (EDM) is employed to avoid the excessive abrasion. Figure 10b shows the outer raceway after damage. Figure 10c shows the damaged bearing and a sample of this testing. The apertures of the holes in the inner and outer raceway are 1 mm. The output current of the power generator is shown in Figure 7d.

5. Simulation Results

5.1. Classified Dataset

According to the four operating conditions of the generator, data is measured and recorded, sequentially. Each operating condition can generate 200 samples; the first 160 samples are taken for a training set and the latter 40 samples are taken as a testing set. The samples total 800, including the training set of 640 samples and testing data of 160 samples, shown in Table 1. Each data is a signal lasting for 0.5 s. According to Equation (5), E f does not have unit, and the energy spectrums of four operational conditions of each feature are shown in Figure 11.
Classification can be influenced by different levels of noise; therefore, various white Gaussian noises are involved in the testing. The signal-to-noise ratio (SNR) is used to measure the level of noise. SNRs of 40, 30, and 20 dB are included to compare the difference. The SNR is defined as follows:
SNR   =   10   log 10 ( P s i g n a l P n o i s e )   dB
where P s i g n a l is the power of original signal and P n o i s e is noise.

5.2. Simulation Results

This paper uses the PWKNN as a classifier of malfunction diagnosis and compares it with the back propagation neural network (BPNN), the k-NN, and the radial basis function network (RBFN). The iteration number of the PWKNN training is 400 times. The BPNN uses 40 neurons. The learning rate is 0.02; the iteration number is 400 times. The k parameter of k-NN is 3. The iteration number of RBFN training is 400 times; the spread is set as 0.1.
Without feature selection and noise disturbance, PWKNN, BPNN, and RBFN show correct classified accuracy. However, k-NN has 97.5% classified accuracy, as shown in Table 2. However, SNRs of 40, 30, and 20 dB show different results. If the SNR is 30 dB, PWKNN has 94% classification accuracy. The classification accuracy of BPNN and k-NN is comparatively less, i.e., 79% and 71%, respectively. The RBFN shows 65% classification accuracy. Regarding the computing time, the training time of the PWKNN and the BPNN is 140 and 102 s, respectively. The k-NN does not need to train. The k-NN and RBFN take 1.67 × 104 and 1.78 × 104 s, respectively, to complete the output of classifiers, whereas PWKNN takes 2.19 × 104 s and BPNN takes the most time to complete. Although k-NN and RBFN are better than PWKNN in time, PWKNN and BPNN are superior to k-NN and RBFN in classification accuracy. Furthermore, PWKNN is better than BPNN in average computing time or classification accuracy. Although PWKNN with a SNR of 20 dB is inferior to BPNN, PWKNN with a SNR of 30 dB is better than BPNN in accuracy.
Correlation coefficients for the classification results of all features should be computed first in feature extraction. The features of smallest correlation coefficients are the first to be removed. Table 3 shows the correlation coefficients for the classification of all features; d 15 , d 14 and so forth are removed. PWKNN, k-NN and RBFN use the predictive classification accuracy as the criterion of feature selection; BPNN uses the mean square error as the criterion. The classification results and number after feature selection are shown in Table 4. Most classifiers enhance classification accuracy. When the SNR is 30 dB, RBFN raises the classification accuracy from 65% to 91%. The percentage of accuracy rises about 40%. The feature selection increases the training time but reduces the output time. In Table 4, while BPNN needs the longest training time, RBFN needs less training time than PWKNN, and k-NN needs the least training time. PWKNN, BPNN, k-NN, and RBFN enhance the output time by 61%, 6%, 42%, and 17%, respectively. The feature number is the average number of repetitive computations for five times. The average features of PWKNN, BPNN, k-NN, and RBFN are 2.8, 10.1, 2.1, and 6.5, respectively. PWKNN and k-NN have the highest efficiency of feature selection. The results show that the feature number has less influence on the output time because BPNN is limited to the network structure and neuron numbers. PWKNN has more efficient feature selection and can improve the output time.
Although PWKNN does not have the least training and output time, the average classification accuracy is higher than other classifiers. BPNN has the equivalent classification accuracy but needs longer training time. Nevertheless, the output time is the highest. k-NN needs the least time; however, its classification accuracy is not higher than that of PWKNN. In addition, features influence RBFN significantly. Therefore, feature selection is an important factor. Thus, PWKNN is efficient for recognizing the operating conditions of wind generators. With the feature selection, PWKNN can reduce the feature numbers effectively and shorten the output time of classifiers.

6. Conclusions

This paper proposes the PWKNN to recognize the operational conditions of power generators. The PWKNN solves the identical probability of multi-label classification. The weight after adjustment can influence the features of classification through the parameter optimization of PSO and raise the classification accuracy effectively. Feature selection can eliminate the input time of classifiers and remove features effectively without lowering the classification accuracy. The research results show that whether there is noise or not, the average classification accuracy of the PWKNN is higher than that of other classifiers. If the SNR is 30 dB, the PWKNN has 93% classification accuracy. With the feature selection, the average feature numbers decrease from 16 to 2.8. The output time lowers 61%. If the SNR is 20 dB, the PWKNN increases classification accuracy from 72% to 82% through feature selection, which is better than other traditional classifiers. The PWKNN’s ability to efficiently eliminate redundant features and strong classifications in noise conditions significantly improves the predictability of wind turbine failures as compared with other methods. In addition, the PWKNN is capable of diagnosing faults in wind power systems and can also consider applying in fault diagnosis of rotating machines.

Author Contributions

Conceptualization, C.-Y.L., K.-Y.H., Y.-X.S., Y.-C.L.; Methodology, C.-Y.L., K.-Y.H., Y.-X.S., Y.-C.L.; Software C.-Y.L., K.-Y.H., Y.-X.S., Y.-C.L.; Validation, C.-Y.L., K.-Y.H., Y.-X.S., Y.-C.L.; Formal Analysis, C.-Y.L., K.-Y.H., Y.-X.S., Y.-C.L.; Investigation, C.-Y.L., K.-Y.H., Y.-X.S., Y.-C.L.; Resources, C C.-Y.L., K.-Y.H., Y.-X.S., Y.-C.L.; Data Curation, C.-Y.L., K.-Y.H., Y.-X.S., Y.-C.L.; Writing—Original Draft Preparation, C.-Y.L., K.-Y.H., Y.-X.S., Y.-C.L.; Writing—Review & Editing, C.-Y.L., K.-Y.H., Y.-X.S., Y.-C.L.; Visualization, C.-Y.L., K.-Y.H., Y.-X.S., Y.-C.L.; Supervision, C.-Y.L.; Project Administration, C.-Y.L.; Funding Acquisition, C.-Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

a The scale parameter
a n Approximation coefficient
b Translation parameter
C W T ( a , b ) Continuous wavelet transform
dB Decibel
d 1 d n Detail coefficients
d 1 d n Euclidean distance between unknown point and class
d i s t ( X , Y ) Euclidean distance computes of n-dimensions vectors
D Decision result
E f Energy spectrum
E a j Energy of approximation coefficient
E d j Energy of detail coefficients
f ( t ) Original signal
F i t ( ) Fitness function
g b e s t i The best position of all particles
j Scale
k Nearest neighbor
l and z Time
N Number of all features
N C o r r e c t l y Number of the correct classification
N T o t a l Number of total training samples
P C V Predictive accuracy of PWKNN
p c a C V Cross-validation predictive classification accuracy
p c a C V Classification accuracy
p c a C V max , N Classification accuracy of PWKNN
p c a C V max , N 1 classification accuracy through weight optimization
P s i g n a l Power of original signal
P n o i s e Noise
p b e s t i The best position of particle i
R Correlation coefficient,
r 1 and r 2 Random number
t j One of the neighbors in the training set
x Input
x i The current position of particle i
x ¯ Mean of input,
X   and   Y n-Dimensions vectors
y Output
y ¯ Mean of output
y ( t j , c m ) Indicates whether t j belongs to class c m
v i The velocity of particle i
w Inertia weight factor
W Weight
W N Weight after the PWKNN optimization
W N 1 Features after removal can gain the optimum weight
ψ ( t ) Mother wavelet
φ 1 Particle self-learning coefficient
φ 2 Particle swarm learning coefficient

References

  1. Turnbull, A.; Carroll, J.; Koukoura, S.; McDonald, A. Prediction of wind turbine generator bearing failure through analysis of high-frequency vibration data and the application of support vector machine algorithms. J. Eng. 2019, 4965–4969. [Google Scholar] [CrossRef]
  2. Artigao, E.; Sapena-Bano, A.; Honrubia-Escribano, A.; Martinez-Roman, J.; Puche-Panadero, R.; Gómez-Lázaro, E. Long-term operational data analysis of an in-service wind turbine DFIG. IEEE Access 2019, 7, 17896–17906. [Google Scholar] [CrossRef]
  3. Kabir, M.J.; Oo, A.M.T.; Rabbani, M. A brief review on offshore wind turbine fault detection and recent development in condition monitoring based maintenance system. In Proceedings of the 2015 Australasian Universities Power Engineering Conference (AUPEC), Wollongong, Australia, 27–30 September 2015; pp. 1–7. [Google Scholar]
  4. Rezamand, M.; Carriveau, R.; Ting, D.S.; Davison, M.; Davis, J.J. Aggregate reliability analysis of wind turbine generators. IET Renew. Power Gener. 2019, 13, 1902–1910. [Google Scholar] [CrossRef]
  5. Tobon-Mejia, D.A.; Medjaher, K.; Zerhouni, N. CNC machine tool’s wear diagnostic and prognostic by using dynamic Bayesian networks. Mech. Syst. Signal Process. 2012, 28, 167–182. [Google Scholar] [CrossRef] [Green Version]
  6. Lee, C.Y.; Lin, W.C. Induction motor fault classification based on FCBF-PSO feature selection method. Appl. Sci. 2020, 10, 5383. [Google Scholar] [CrossRef]
  7. Watson, S.J.; Xiang, B.J.; Yang, W.; Tavner, P.J.; Crabtree, C.J. Condition monitoring of the power output of wind turbine generators using wavelets. IEEE Trans. Energy Convers. 2010, 25, 715–721. [Google Scholar] [CrossRef] [Green Version]
  8. Tavner, P.J. Review of condition monitoring of rotating electrical machines. IET Electr. Power Appl. 2008, 2, 215–247. [Google Scholar] [CrossRef]
  9. Lee, C.Y.; Shen, Y.X. Optimal feature selection for power-quality disturbances classification. IEEE Trans. Power Deliv. 2011, 26, 2342–2351. [Google Scholar] [CrossRef]
  10. Jin, N.; Liu, D.-R. Wavelet basis function neural networks for sequential learning. IEEE Trans. Neural Netw. 2008, 19, 523–528. [Google Scholar] [CrossRef]
  11. Perera, N.; Rajapakse, A.D. Recognition of fault transients using a probabilistic neural-network classifier. IEEE Trans. Power Deliv. 2010, 26, 410–419. [Google Scholar] [CrossRef]
  12. Tripathy, M.; Maheshwari, R.; Verma, H. Power transformer differential protection based on optimal probabilistic neural network. IEEE Trans. Power Deliv. 2009, 25, 102–112. [Google Scholar] [CrossRef]
  13. Ying, S.; Jianguo, Q. A method of arc priority determination Based on Back-Propagation Neural Network. In Proceedings of the 2017 4th International Conference on Information Science and Control Engineering (ICISCE), Changsha, China, 21–23 July 2017; pp. 38–41. [Google Scholar]
  14. Wu, W.; Feng, G.; Li, Z.; Xu, Y. Deterministic convergence of an online gradient method for BP neural networks. IEEE Trans. Neural Netw. 2005, 16, 533–540. [Google Scholar] [CrossRef] [PubMed]
  15. Wang, D.; Tsui, K.; Miao, Q. Prognostics and health management: A review of vibration based bearing and gear health indicators. IEEE Access 2017, 6, 665–676. [Google Scholar] [CrossRef]
  16. Wang, J.; Peng, Y.; Qiao, W. Current-aided order tracking of vibration signals for bearing fault diagnosis of direct-drive wind turbines. IEEE Trans. Ind. Electron. 2016, 63, 6336–6346. [Google Scholar] [CrossRef]
  17. Gong, X.; Qiao, W. Current-based mechanical fault detection for direct-drive wind turbines via synchronous sampling and impulse detection. IEEE Trans. Ind. Electron. 2015, 62, 1693–1702. [Google Scholar] [CrossRef]
  18. Cheng, F.; Qu, L.; Qiao, W.; Wei, C.; Hao, L. Fault diagnosis of wind turbine gearboxes based on DIFG stator current envelope analysis. IEEE Trans. Sustain. Energy 2019, 10, 1044–1053. [Google Scholar] [CrossRef]
  19. Sapena-Bañó, A.; Pineda-Sanchez, M.; Puche-Panadero, R.; Martinez-Roman, J.; Matić, D. Fault diagnosis of rotating electrical machines in transient regime using a single stator current’s FFT. IEEE Trans. Instrum. Meas. 2015, 64, 3137–3146. [Google Scholar] [CrossRef] [Green Version]
  20. Bayram, D.; Şeker, S. Redundancy-based predictive fault detection on electric motors by stationary wavelet transform. IEEE Trans. Ind. App. 2017, 53, 2997–3004. [Google Scholar] [CrossRef]
  21. Ambusaidi, M.A.; He, X.; Nanda, P.; Tan, Z. Building an intrusion detection system using a filter-based feature selection algorithm. IEEE Trans. Comp. 2016, 65, 2986–2998. [Google Scholar] [CrossRef] [Green Version]
  22. Chkirbene, Z.; Erbad, A.; Hamila, R.; Mohamed, A.; Guizani, M.; Hamdi, M. TIDCS: A dynamic intrusion detection and classification system based feature selection. IEEE Access 2020, 8, 95864–95877. [Google Scholar] [CrossRef]
  23. Alhakami, W.; Alharbi, A.; Bourouis, S.; Alroobaea, R.; Bouguila, N. Network anomaly intrusion detection using a nonparametric bayesian approach and feature selection. IEEE Access 2019, 7, 52181–52190. [Google Scholar] [CrossRef]
  24. Sarvari, S.; Mohd Sani, N.F.; Mohd Hanapi, Z.; Abdullah, M.T. An efficient anomaly intrusion detection method with feature selection and evolutionary neural network. IEEE Access 2020, 8, 70651–70663. [Google Scholar] [CrossRef]
  25. Han, E.H.; Karypis, G.; Kumar, V. Text categorization using weight adjusted k-Nearest Neighbor classification. In Pacific-Asia Conference on Knowledge Discovery and Data Mining; Springer: Berlin/Heidelberg, Germany, 2001; pp. 53–65. [Google Scholar]
  26. Mullick, S.S.; Datta, S.; Das, S. Adaptive learning-based k-nearest neighbor classifiers with resilience to class imbalance. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 5713–5725. [Google Scholar] [CrossRef] [PubMed]
  27. Mallat, S. A theory for multiresolution signal decomposition: The wavelet representation. IEEE Trans. Pattern Anal. Mach. Intell. 1989, 11, 674–693. [Google Scholar] [CrossRef] [Green Version]
  28. Zhang, S.; Li, X.; Zong, M.; Zhu, X.; Wang, R. Efficient kNN classification with different numbers of nearest neighbors. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 1774–1785. [Google Scholar] [CrossRef]
  29. Kennedy, J.; Eberhart, R. Particle swarm optimization. Proc. IEEE Int. Conf. Neural Netw. 1995, 6, 1942–1948. [Google Scholar]
  30. Lee, C.Y.; Tuegeh, M. Optimal optimisation-based microgrid scheduling considering impacts of unexpected forecast errors due to the uncertainty of renewable generation and loads fluctuation. IET Renew. Power Gen. 2020, 14, 321–331. [Google Scholar] [CrossRef]
  31. Lee, C.Y.; Tuegeh, M. An Optimal solution for smooth and non-smooth cost functions-based economic dispatch problem. Energies 2020, 13, 3721. [Google Scholar] [CrossRef]
  32. Shi, Y.; Eberhart, R. Parameter selection in particle swarm optimization. In Proceedings of the Seventh Annual Conference on Evolutionary Programming, San Diego, CA, USA, 25–27 March 1998; pp. 591–600. [Google Scholar]
  33. Shi, Y.; Eberhart, R. Empirical study of particle swarm optimization. Proc. IEEE Int. Congr. Evol. Comp. 1999, 3, 101–106. [Google Scholar]
Figure 1. The structure of the multi-resolution analysis (MRA).
Figure 1. The structure of the multi-resolution analysis (MRA).
Energies 13 05520 g001
Figure 2. The spectral distribution of wavelet coefficient.
Figure 2. The spectral distribution of wavelet coefficient.
Energies 13 05520 g002
Figure 3. The diagram of classification by weighted k-NN (WKNN).
Figure 3. The diagram of classification by weighted k-NN (WKNN).
Energies 13 05520 g003
Figure 4. Flowchart of particle swarm with WKNN (PWKNN).
Figure 4. Flowchart of particle swarm with WKNN (PWKNN).
Energies 13 05520 g004
Figure 5. Flowchart of feature selection.
Figure 5. Flowchart of feature selection.
Energies 13 05520 g005
Figure 6. The block diagram of wind power simulation platform.
Figure 6. The block diagram of wind power simulation platform.
Energies 13 05520 g006
Figure 7. Generator output current: (a) Healthy; (b) Loss of lubrication; (c) Rotor angular misalignment; (d) Bearing fault.
Figure 7. Generator output current: (a) Healthy; (b) Loss of lubrication; (c) Rotor angular misalignment; (d) Bearing fault.
Energies 13 05520 g007
Figure 8. Operation of healthy condition.
Figure 8. Operation of healthy condition.
Energies 13 05520 g008
Figure 9. Operation of rotor angular misalignment.
Figure 9. Operation of rotor angular misalignment.
Energies 13 05520 g009
Figure 10. The bearing: (a) Healthy; (b) Outer raceway after damage; (c) Damaged bearing.
Figure 10. The bearing: (a) Healthy; (b) Outer raceway after damage; (c) Damaged bearing.
Energies 13 05520 g010
Figure 11. Energy spectrum: (a) Healthy; (b) Gearbox of loss lubrication; (c) Rotor angular misalignment; (d) Bearing fault.
Figure 11. Energy spectrum: (a) Healthy; (b) Gearbox of loss lubrication; (c) Rotor angular misalignment; (d) Bearing fault.
Energies 13 05520 g011
Table 1. The state distribution of training set and test set.
Table 1. The state distribution of training set and test set.
Operation ConditionTraining SetTest Set
Healthy16040
Gearbox of loss lubrication16040
Rotor angular misalignment16040
Rearing fault16040
Total640160
Table 2. The classification results with feature selection.
Table 2. The classification results with feature selection.
Without Feature SelectionClassification Accuracy (%)
PWKNNBPNNk-NNRBFN
No noise10010097.5100
40 dB98.893.179.483.1
30 dB93.579.471.365.0
20 dB71.675.060.043.1
Training time (s)104120--36
Output time (10−4 s)2.192.521.781.67
Feature number16
Table 3. The correlation of each feature.
Table 3. The correlation of each feature.
Correlation Coefficient
d 1 d 5 d 4 d 3 d 8 d 2 a 15 d 7
4.533.183.163.081.030.890.470.42
d 6 d 11 d 9 d 13 d 12 d 10 d 14 d 15
0.330.320.320.280.270.210.200.20
Table 4. The classification results with feature selection.
Table 4. The classification results with feature selection.
With Feature SelectionClassification Accuracy (%)
PWKNNBPNNk-NNRBFN
No noise100100100100
40 dB99.490.881.9100
30 dB93.181.383.190.6
20 dB82.575.580.655.0
Training time (s)46793743046
Output time (10−4 s)1.362.371.031.39
Feature number (Average)2.810.12.56.5
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, C.-Y.; Huang, K.-Y.; Shen, Y.-X.; Lee, Y.-C. Improved Weighted k-Nearest Neighbor Based on PSO for Wind Power System State Recognition. Energies 2020, 13, 5520. https://doi.org/10.3390/en13205520

AMA Style

Lee C-Y, Huang K-Y, Shen Y-X, Lee Y-C. Improved Weighted k-Nearest Neighbor Based on PSO for Wind Power System State Recognition. Energies. 2020; 13(20):5520. https://doi.org/10.3390/en13205520

Chicago/Turabian Style

Lee, Chun-Yao, Kuan-Yu Huang, Yi-Xing Shen, and Yao-Chen Lee. 2020. "Improved Weighted k-Nearest Neighbor Based on PSO for Wind Power System State Recognition" Energies 13, no. 20: 5520. https://doi.org/10.3390/en13205520

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop