Next Article in Journal
Demosaicing of CFA 3.0 with Applications to Low Lighting Images
Next Article in Special Issue
Construction of a Sensitive and Speed Invariant Gearbox Fault Diagnosis Model Using an Incorporated Utilizing Adaptive Noise Control and a Stacked Sparse Autoencoder-Based Deep Neural Network
Previous Article in Journal
Multi-Mode Love-Wave SAW Magnetic-Field Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bearing Fault Diagnosis Using a Particle Swarm Optimization-Least Squares Wavelet Support Vector Machine Classifier

1
Centre for Intelligent and Autonomous Manufacturing Systems, and School of Electronics, Electrical Engineering and Computer Science, Queen’s University Belfast, Belfast BT7 1NN, UK
2
Department of Electrical Engineering, University of Ulsan, Ulsan 44610, Korea
3
School of Electrical Engineering, University of Ulsan, Ulsan 44610, Korea
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(12), 3422; https://doi.org/10.3390/s20123422
Submission received: 12 May 2020 / Revised: 11 June 2020 / Accepted: 14 June 2020 / Published: 17 June 2020

Abstract

:
Bearing is one of the key components of a rotating machine. Hence, monitoring health condition of the bearing is of paramount importace. This paper develops a novel particle swarm optimization (PSO)-least squares wavelet support vector machine (PSO-LSWSVM) classifier, which is designed based on a combination between a PSO, a least squares procedure, and a new wavelet kernel function-based support vector machine (SVM), for bearing fault diagnosis. In this work, bearing fault classification is transformed into a pattern recognition problem, which consists of three stages of data processing. Firstly, a rich information dataset is built by extracting the features from the signals, which are decomposed by the nonlocal means (NLM) and empirical mode decomposition (EMD). Secondly, a minimum-redundancy maximum-relevance (mRMR) method is employed to determine a subset of feature that can provide an optimal performance. Thirdly, a novel classifier, namely LSWSVM, is proposed with the aid of a PSO, to provide higher classification accuracy. The key innovative science of this work is to propropose a new classifier with the aid of an new wavelet kernel type to increase the classification precision of bearing fault diagnosis. The merit features of the proposed approach are demonstrated based on a benchmark bearing dataset and a comprehensive comparison procedure.

1. Introduction

Since bearing is a crucial component in the machine, its failure will hugely affect to the disruption of the machine. Therefore, condition monitoring for rolling bearings has become more and more important to detect early damage and increase safe of the operating systems. In the literature, two approaches can be applied to detect the bearing defects: (1) acoustic signal analysis, where the acoustic signal is acquired to obtain bearing characteristic information, and (2) vibration signal analysis, where the vibration signal is acquired. Among them, using vibration signal usually provides better defect detecting accuracy becuase it contains rich information of the bearing characteristics and less measurement noise [1].
Bearing defects can be detected by either analyzing the fault frequency spectrum [2] or pattern recognition [3]. However, the analysis in [4] shown that the pattern recognition can give higher accuracy compared to the spectrum approach. In the approach of traditional pattern recognition, the system will include three major components: feature extraction, feature selection and feature classification. The goal of the feature extraction task is to get as much information about the condition of the system as good. For this purpose, we employ the NLM-EMD method, which has been developed in our previous work [5] and proved its effectiveness, to extract a rich bearing feature set.
Feature extraction usually results in a large feature set. Unfortunately, the large feature set does not neccessarily provide higher classification accuracy as it possibly contains irrelevant and redundant features. Thus, it is signiticant to eliminate the irrelevant and redundant features before it is fed back to a classifier. To obtain an optimal feature subset, a minimum-redundancy maximum-relevance (mRMR) feature selection method has been developed [6]. The mRMR tries to search an outstanding combination of candidate features for minimum redundancy and maximum relevance. Due to the merits of the mRMR, it is employed in this paper to select the effective features.
Once the salient features are selected, they are fed into a classifier to identify the system condition. Due to its high performance classification and less requirement on sample data input, the support vector machine (SVM) proposed by Cortes and Vapnik [7] has been successfully applied to signal processing [8], regression analysis [9], pattern recognition [10], and bearing fault diagnosis [11]. However, the original SVM classifier provides high computational burden due to the method used to solve the quadratic programming problem in the SVM [12]. In order to reduce this, many methods have been developed, for example the SVM light decomposition algorithm [13], sequential minimal optimization (SMO) algorithm [14], neighbor algorithm [15], and least squares SVM (LSSVM) [16]. Among them, the LSSVM is commonly applied in real applications due to its simplicity in implementation and efficiency in classification and computation [17].
In the SVM classifier, a kernel function is used to transform the data from the lower dimension space to a high dimension space. Hence, the prior selection of the kernel will decide the way of classification of the SVM [18]. Several kind of kernels have been developed for SVM, for example, polynomial, dot product, and radial basis function (RBF) kernels. Among them, RBF kernel has shown to be more effective because it has good capacity to approximate nonlinear functions. Recently, wavelet kernel has been developed as an effective method for nonlinear approximation and mapping [4,19]. In [20], Zhang et al. has employed the wavelet kernel for the SVM classifier, and a wavelet SVM (WSVM) classifier has been proposed as a result. Since the wavelet transform provides better approximation capacity than the RBF, the WSVM classifier provides higher accuracy than the SVM with RBF kernel. Since then, the WSVM have been employed in many real applications, such as in the medical field [21], and machine fault diagnosis [22]. Due to the merits of the LSSVM classifier and the approximation capability of the wavelet kernel, a new least squares wavelet support vector machine (LSWSVM) is proposed first time in this paper to improve both computational efficiency and classification accuracy. However, the generalization performance of the LSWSVM is affected by its parameters. Thus, it is necessary to optimize the parameters to obtain a better performance. In the literature, Particle swarm optimization (PSO) [23] has been developed as an effective optimization technique to optimize parameters of a process. Compared with other optimization methods, PSO have many advantages, such as simple implementation, few parameters, parallel computation ability, and quickly converge [24]. The PSO had proved its optimization capacity when applying for many practical applications, such as for optimizing the parameters of SVMs [25] and other optimization problems [26,27]. Therefore, the PSO is used in this paper to effectively select the parameters of the LSWSVM, leading to a new PSO-LSWSVM classifier, which addresses all difficulties in the use of the SVM classifier.
In summary, the novelties and main contributions of this paper can be listed as follows:
  • A new methodology for bearing fault diagnosis is developed by combining between feature extration based on a NLM-EMD method, a feature selection based on a mRMR and a new PSO-LSWSVM classifier.
  • To improve the generalization performance of the SVM, a novel PSO-LSWSVM classifier, which combines between a least squares procedure, a new wavelet kernel function and the PSO, is proposed.

2. Feature Extraction

In this paper, we employ the NLM-EMD method, which has been developed in our previous work [5] and proved its effectiveness, to extract a rich bearing feature set. For the merit features of the NLM-EMD and its detail description, the interested readers can refer to the previous work [5].

2.1. Nonlocal Mean (NLM) De-Noising

Consider a noise signal has a form as y = u + n , where u is the true signal and n is an additive noise. The noise component can be eliminated using a NLM as below:
u ^ ( i ) = 1 M ( i ) j Ω i ω ( i , j ) y ( i )
The parameters used in (1) can be designed as in [5]. For more detail description of the NLM denoising, the interested readers can refer to our previous paper [5].

2.2. Empirical Mode Decomposition

Consider an original signal x ( t ) , a number of IMFs C ( t ) can be obtained from the original signal using EMD method as [28]
x ( t ) = j = 1 n C j ( t ) + r n ( t )
where the high frequency is decreased from C 1 ( t ) , C 2 ( t ) , C 3 ( t ) , , C n ( t ) , and r n ( t ) contains no meaningful information. Generally, fault information is distributed significantly on the high and mid-frequency components [4,19]. Thus, the first five IMFs are used in this work for bearing fault analysis since they represent the mid- and high frequency components of the original signal.

2.3. Energy Feature Extraction

In the previous section, the EMD has been employed to decompose the original signal into a number of IMF components with different frequency bands. On the other hand, the frequency band can be referenced of the energy of fault vibration signal. Hence, in order to capture the effects of faults on the change of the energy of the vibration signal, IMF energy features are employed.
Each IMF component C j ( t ) possesses an energy E j ( t ) , which can be calculated as:
E j = | C j ( t ) | 2 d t
Then, a normalization procedure can be applied for each E j ( t ) :
T j = E j T
where T is the total energy of the first five IMF components:
T = j = 1 5 | E j | 2 1 2

2.4. Time-Domain Feature Extraction

Time-domain features usually provide rich information to distinguish normal condition and fault condition. In this paper, the nine time-domain dimensionless parameters defined in Table 1 is used to extract fault information from the de-noised signal and the first five IMFs to obtain rich information of bearing faults.
Finally, a set of features, which includes 5 + 9 × 6 = 59 fetures, is obtained to represent a bearing condition.

3. Minimum Redundancy Maximum Relevance (MRMR) Feature Selection

Let F be the initial feature set and | S | be the cardinality in seeking feature subset S. The following criterion is developed for minimal redundancy:
min S F 1 | S | 2 i , j S I ( f i , f j )
and the maximum relevance criterion is defined as:
max S F 1 | S | i S I ( C , f i )
where I ( f i , f j ) is the mutual information of two features, f i and f j ; and I ( C , f i ) quantifies the relevance of the feature, f i , in S and the target class, C.
To obtain a feature subset with minimum redundancy and maximum relevance, a mRMR function is obtained by combining (6) and (7):
max S F i S I ( C , f i ) 1 | S | i , j S I ( f i , f j )
The completed procedure of the mRMR can be refered to [4]. To obtain the desired feature subset, forward selection search [29] is employed.

4. PSO-LSWSVM

4.1. Least Squares Support Vector Machine (LSSVM)

Given a training set of N data points, ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x N , y N ) , where x i R d is the i t h input vector and y i ± 1 is the corresponding target, we employ the idea of the transformation of an input pattern into a reproducing kernel Hilbert space using a set of mapping functions, ϕ ( x ) . The reproducing kernel, K ( x , x ) , in the reproducing kernel Hilbert space is the dot product of the mapping functions at x and x , i.e., K ( x , x ) = ϕ ( x ) . ϕ ( x ) . In the new defined kernel space, a linear classifier usually has a form below:
y ( x ) = s i g n ( ω . ϕ ( x ) + b )
To facilitate the selection of the parameters ω and b, the LSSVM formulates the optimization problem as:
minimize ω , b , e F ( ω , b , e ) = 1 2 ω T ω + C 2 i = 1 N e i 2 subject   to y i [ ω T ϕ ( x i ) + b ] = 1 e i
The feature mapping, i.e., ϕ ( x ) , is usually unknown, and Mercer’s condition [30] can be appllied.
Ω i j = y i y j ϕ ( x i ) T ϕ ( x j )
The decision function of the LSSVM classifier becomes:
y i = s i g n j = 1 N α j y j K ( x i , x j ) + b
A kernel RBF can be chosen as:
K ( x i , x j ) = e x p ( | | x i x j | | 2 σ 2 )
where σ is a free parameter.

4.2. Least Squares Wavelet Support Vector Machine (LSWSVM)

Generally, the family of wavelet analysis has a form:
h a , c ( z ) = | a | 1 2 h ( z c a )
where z , a , c R , a is a dilation factor, c is a translation factor; and h ( z ) is the mother wavelet, which satisfies the following condition [31,32]:
W h = 0 | F ( ω ) | 2 | ω | <
where F ( ω ) is the output of h ( z ) using Fourier transform. Employing a wavelet transform for g ( z ) , one obtains:
W a , c ( g ) = g ( z ) , h a , c ( z )
where indicates the dot product. The function g ( z ) is provided by [31]:
g ( z ) = 1 W h 0 1 a 2 W a , c ( g ) h a , c ( z ) d a d c
Reformulate (17):
g ^ ( z ) = i = 1 N W i h a i , c i ( z )
where W i is the reconstruction coefficient, and g ( z ) is approximated by g ^ ( z ) .
A wavelet function can be selected as [31]:
h ( z ) = i = 1 N h ( z i )
where z = [ z 1 , z 2 , , z N ] T R . Then, if z , z R N , the dot-product wavelet kernels can be computed as:
K ( z , z ) = i = 1 N h z i c i a i = 1 N h z i c i a
and the following expression is used to describe the translation invariant wavelet kernels [31]:
K ( z , z ) = i = 1 N h z i z j a
Substituting (21) into (12), the decision function of the LSWSVM classifier has a form below:
y i = s i g n ( j 1 N α j y j j = 1 N h x t , j x i , j a i + b
where x t , j and x i , j denote the j t h element of x t and the i t h training sample, x i , respectively.
In order to approximate a general nonlinear model, in this paper, we propose to use the following wavelet kernel:
h ( x ) = λ c o s k x a . exp x 2 a 2
where a is a parameter of the RBF kernel; k and λ are new parameters that control the kernel shape. It is obvious to see from equation (23) that the performance of the defined wavelet kernel depends significantly on the selection of the parameters a, k, and λ . When the parameters a, k, and λ are changed, the shape of the kernel is changed. Therefore, it is needed to optimize these parameters to obtain a good performance of the system.

4.3. Particle Swarm Optimization (PSO) for Parameter Selection of LSWSVM—the PSO-LSWSVM Classifier

In order to get the optimal values of the parameters a, k, and λ , particle swarm optimization (PSO) [33] is employed in this paper. The detail description of the PSO can be referred to our previous work [4,19] to reduce the length of the paper. The velocity, position and the initial weight of the PSO are updated using the following three equations:
v i d t + 1 = ω . v i d t + c 1 r 1 . ( p b e s t , i d t x i d t + c 2 r 2 . ( g b e s t , d t x i d t ) )
x i d t + 1 = x i d t + v i d t + 1
ω k = ω m a x ω m a x ω m i n i t e r m a x × i t e r
The definitions of the parameters used in Equations (24)–(26) can be referred to [4,19].
The LSWSVM classification model constructed using the wavelet kernel function defined in (23) has four user-determined parameters, including a regularization parameter C and three kernel parameters, λ , k and a. In this paper, we use PSO to automatically select the parameters of the LSWSVM classifier; hence, a relatively new classifier, i.e., PSO-LSWSVM, is proposed. The step-by-step implementation details of parameters selection for the LSWSVM classifer based on PSO are described below.
Step 1:
Initializes the parameters of the PSO: the population N, the position and velocity of each particle (C, a, k and λ -parameters for LSWSVM).
Step 2:
Uses the following fitness function, which is obtained from the output of the LSWSVM classifier, to evaluate the initialized particles:
fitness function = N t N t + N f
where N t and N f denotes the number of true and false classification, respectively.
Step 3:
Creates a new swarm by updating the velocity and position of each particle using (24) and (25).
Step 4:
For the new obtained swarm, the fitness values are computed and compared to update the p b e s t , i and G b e s t of the swarm.
Step 5:
Checks the termination condition: If the maximum number is reached, goes to Step 6. Otherwise, return to Step 3 and continue the closed-loop process.
Step 6:
Encodes the optimal parameter of the wavelet kernel of the LSWSVM classifier from the global best position, G b e s t .

5. Fault Diagnosis Methodology

The proposed fault diagnosis methodology is briefly described as in Figure 1. The implementation is executed as follows:
Step 1:
A number of effective IMFs are obtained after filtering the vibration signals using the NLM and EMD.
Step 2:
Extracts the energy and time domain features to obtain a combined feature set.
Step 3:
Uses the mRMR feature selection technique to get an optimal feature subset.
Step 4:
Uses the wavelet kernel function defined in (23) for LSSVM classifier and optimizes the parameters using the PSO technique.
Step 5:
Classifies the bearing fault types using the PSO-LSWSVM classifer based on the ‘one to others’ multi-class classification strategy [34], which is illustrated in Figure 2, and the selected feature subset in Step 3.
Remark: Although the full fault diagnosis system, which includes feature extraction, feature selection, and feature classification, is presented in this paper, the major contribution of this paper is to introduce a novel PSO-LSWSVM classifier. The feature extraction tasks are mainly taken from the previous work [4], while the feature selection based on the mRMR is a standard and well-known technique in the literature.

6. Experimental Results

6.1. Training and Test Data Configuration

The data used in this experiment are taken from the Case Western Reserve University Bearing Data Center (2014) [35]. The bearing test-bed is shown in Figure 3. In this paper, four types of bearing conditions are considered, including one normal condition (no fault) which is labeled as NM and three fault conditions. The three fault conditions include fault at outer race, fault at inner race and fault at ball which are labeled as ORF, IRF and BF respectively. In each type of fault condition, fault size can have the value of 0.007, 0.014 or 0.021 mili-inches. Therefore, totally 10 conditions (10 classes) of bearing are taken into account.

6.2. Parameter Selection

In the first simulation set, we illustrate the performance of the NLM and EMD. Figure 4, Figure 5, Figure 6 and Figure 7 illustrate the denoising results using the NLM. The denoised signals are then passed through the EMD to obtain the effective IMF components. The 59 features are then extracted from the denoised signal and the IMF components as described in Section 2.
In the second and third simulation sets, the computed feature set is fed into the mMRM feature selection to get an optimal feature subset. The selected feature subset is then used as input to a classifier to identify the bearing conditions. The LSWSVM classifier was implemented based on a modification of the LS-SVMLabtoolbox [36]. In order to verify the effectiveness of the PSO and the proposed wavelet kernel function, we constructed four different classifiers: (1) an LSRBFSVM classifier using an RBF kernel for the LSSVM with parameters selected by the user; (2) a PSO-LSRBFSVM classifier (LSRBFSVM with parameters are selected by using PSO); (3) an LSWSVM classifier using the proposed wavelet kernel in (25) with parameters selected by the users; and (4) a PSO-LSWSVM classifier (using PSO to automatically select the parameters of the LSWSVM). In addition, to verify the effects of the parameters λ , k and a, the PSO-LSWSVM classifier is used in three different circumstances: (a) λ and k are firstly selected by user, and the PSO is used to tune the parameters a and C; (b) λ is firstly selected, and the PSO is used to tune the parameters k, a and C simultaneously; and (c) the PSO is used to tune the parameters λ , k, a and C simultaneously. These classifiers are also compared with the k-nearest neighbor (KNN) [37] and probability neural network (PNN) [38] classifiers, which are widely applied for bearing fault diagnosis, to further verify the effectiveness of the proposed classifier.

6.3. Performance Evaluation

According to the forward selection search algorithm [29], 59 feature subsets are created based on the mRMR feature selection. To compare the generalization performance of the classifiers, we consider each feature subset as an independent dataset. Thus, we have 59 different datasets corresponding to 59 feature subsets. To evaluate the performance of the methods, the extracted feature vectors are used as inputs for the classifiers to obtain the classification accuracies. In this paper, to estimate the generalized classification accuracy, l-fold cross-validation (CV) [39], where l is set to 3, is employed. To obtain a precisely classification result, l-fold CV is performed ten times in this study.

6.3.1. Training Process

First, the training process is performed to obtain an optimal feature subset of each classifier and the kernel parameters of the LSRBFSVM and LSWSVM classifiers. The PSO is performed at this training step. The validation accuracy in this study is computed as follows:
C a c c u r a c y = K N T P N S × 100 %
where K = 10 indicates number of classes, N T P indicates the number of true classifications, and N S is the number of samples used in this experiment.
The validation accuracy of 59 features dataset for the KNN, PNN, LSRBFSVM, PSO-LSRBFSVM, LSWSVM, and PSO-LSWSVM classifiers are shown in Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13, respectively. The mean and best results and the computational time (for one fold) of each method are also reported in Table 2 for the sake of comparison. The subspaces according to the best records are assigned as the optimal feature subset according to the forward selection search algorithm [29]. Observing from these figures, we can see that the combined 59 features yields a low classification accuracy due to the presence of the irrelevant and redundant features; for example, 43% for the KNN, 55.95% for the PNN, 45.71% for the LSRBFSVM, 68.57% for the PSO-LSRBFSVM, 62.86% for the LSWSVM, and around 90.95% for the PSO-LSWSVM. By using the mRMR criteria for feature selection, the classification accuracy is clearly increased. For example, for the KNN classifier, the peak value is obtained at 7 features with the accuracy increased up to 83.91%; for the PNN classifier, the peak value is obtained at 17 features with the accuracy increased up to 91.42%; for the LSRBFSVM, the peak value is obtained at 11 features with the accuracy increased up to 91.43%; for the PSO-LSRBFSVM, the peak value is obtained at 20 features with the accuracy increased up to 94.76%; for the LSWSVM, the peak value is obtained at 12 features with the accuracy increased up 99.05%; and for the PSO-LSWSVM, the peak value is obtained at 2 features with the accuracy increased up to 100%.
From these results, four observations can be obtained: (1) the feature subsets selected by the mMRM commonly yield higher accuracy than the use of all 59 features; (2) although the computational time of the PSO-LSWSVM (PSO: 30.52 s + LSWSVM: 0.422 s) classifier is higher than the KNN (0.125 s), PNN (0.109 s) and the PSO-LSRBFSVM classifier (PSO: 24.49 s + LSRBFSVM: 0.375 s), it gives much better performance. It should be notice that although the PSO requires a higher computational time, however the PSO training is done offline, and thus it will not affect to the real time fault diagnosis; (3) comparison results between Figure 12 with Figure 8, Figure 9 and Figure 10 shown that the LSWSVM classifier provides better accuracy compared to the KNN, PNN and LSRBFSVM classifiers; (4) by comparing Figure 11 with Figure 10 and Figure 13 with Figure 12, it is clear that using the PSO for parameters selection always provides better performance than using the random selection. In addition, comparisons between Figure 13a–c shown that all parameters, λ , k, a and C, have significant effects on the performance of the LSWSVM classifier, and that the selection of four parameters simultaneously will produce better generalization performance. Based on Table 2 and the forward selection search algorithm [29], 8 features, 17 features, 20 features and 2 features are selected as the optimal feature subset for the KNN, PNN, PSO-LSRBFSVM and PSO-LSWSVM classifiers, respectively.

6.3.2. Testing Process

After the optimal feature subset and optimal model are selected for each classifier, the testing data samples are used to verify the effectiveness of the classifiers. The confusion matrices that show the performances of the KNN, PNN, PSO-LSRBFSVM and PSO-LSWSVM (using PSO to automatically select all parameters, i.e., λ , k, a and C) classifiers are shown in Table 3, Table 4, Table 5 and Table 6, respectively. From the results, it is obvious to see that the proposed PSO-LSWSVM classifier (accuracy = 95.33%) gives superior classification accuracy compared to the KNN (accuracy = 83.05%), PNN (accuracy = 84.77%), and PSO-LSRBFSVM (accuracy = 86.84%).

7. Conclusions

Two major contributions have been presented in this paper:
  • A new pattern recognition approach for bearing fault diagnosis is developed by combining between feature extration based on a NLM-EMD method, a feature selection based on a mRMR and a new PSO-LSWSVM classifier.
  • A novel PSO-LSWSVM classifier, which combines between a least squares procedure, a new wavelet kernel function and the PSO, is proposed.
In the presented method, the combined NLM-EMD is first employed to acquire more effective IMF components of vibration signals. Then, for the de-noised signal and each IMF component, the energy and time-domain feature parameters are extracted to obtain characteristic parameters. Next, the mRMR feature selection technique is adopted to eliminate the irrelevant and redundant features and select the best combined feature subset. Finally, the selected feature subset is fed into the proposed PSO-LSWSVM classifier to identify the bearing conditions, wherein a novel combination of a PSO, a least squares procedure, and a new wavelet kernel is proposed to address the difficulties in the use of the traditional SVM classifier. By experimenting with a real bearing vibration signal, we verified that the proposed wavelet kernel function has a better generalization performance than the previous kernels, i.e., RBF kernel, and the proposed PSO-LSWSVM classifier can overcome all difficulties in the use of the traditional SVM classifer. In addition, the uses of the NLM-EMD for the feature extraction and mRMR for the feature selection are effective. Therefore the proposed fault diagnosis methodology based on the NLM-EMD, mMRM feature selection and PSO-LSWSVM classifier improves the bearing recognition accuracy significantly, up to 95.53%.

Author Contributions

Conceptualization, M.V.; methodology, M.V.; software, M.V.; validation, D.T.H.; formal analysis, D.T.H.; investigation, D.T.H.; resources, H.J.K.; data curation, M.V.; writing—original draft preparation, M.V.; writing—review and editing, D.T.H.; visualization, M.V; supervision, H.J.K.; project administration, H.J.K; funding acquisition, H.J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (2019R1D1A3A03103528), and partly supported by the start-up grant of Queen’s University Belfast (D8203EEC3054789).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

List of Abbreviation used in paper:
BFBall falut bearing
EMDEmpirical mode decomposition
IMFIntrinsic mode function
IRFInner race fault bearing
KNNK-nearest neighbor
PNNProbability neural network
PSOParticle swarm optimization
LSSVMLeast squares support vector machine
LSWSVMLeast squares wavelet support vector machine
LSRBFSVMLeast squares radial basis function support vector machine
mRMRMinimum-redundancy maximum-relevance
NLMNonlocal mean
NMNormal bearing
ORFOuter race fault bearing
RBFRadial basis function
SMOSequential minimal optimization
SVMSupport vector machine
WSVMWavelet support vector machine

References

  1. Kharche, P.P.; Kshirsagar, S.V. Review of fault detection in rolling element bearing. Int. J. Innov. Res. Adv. Eng. 2014, 1, 169–174. [Google Scholar]
  2. Huang, D.; Yang, J.; Zhou, D.; Litak, G. Novel Adaptive Search Method for Bearing Fault Frequency Using Stochastic Resonance Quantified by Amplitude-Domain Index. IEEE Trans. Instrum. Meas. 2020, 69, 109–121. [Google Scholar] [CrossRef]
  3. Hoang, D.T.; Kang, H.J. A Motor Current Signal Based Bearing Fault Diagnosis Using Deep Learning and Information Fusion. IEEE Trans. Instrum. Meas. 2019, 69, 3325–3333. [Google Scholar] [CrossRef]
  4. Van, M.; Kang, H.J. Bearing defect classification based on individual wavelet local fisher discriminant analysis with particle swarm optimization. IEEE Trans. Ind. Inform. 2015, 12, 124–135. [Google Scholar] [CrossRef]
  5. Van, M.; Kang, H.J.; Shin, K.S. Rolling element bearing fault diagnosis based on non-local means de-noising and empirical mode decomposition. IET Sci. Meas. Technol. 2014, 8, 571–578. [Google Scholar] [CrossRef]
  6. Radovic, M.; Ghalwash, M.; Filipovic, N.; Obradovic, Z. Minimum redundancy maximum relevance feature selection approach for temporal gene expression data. BMC Bioinform. 2017, 18, 9. [Google Scholar] [CrossRef] [Green Version]
  7. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  8. He, J.; Song, C.; Luo, Q.; Lan, L.; Yang, C.; Gui, W. Noise-robust self-adaptive support vector machine for residual oxygen concentration measurement. IEEE Trans. Instrum. Meas. 2020. [Google Scholar] [CrossRef]
  9. Wang, C.P.; Kim, H.J.; Yue, C.; Weygand, J.M.; Hsu, T.S.; Chu, X. Effects of solar wind ultralow-frequency fluctuations on plasma sheet electron temperature: Regression analysis with support vector machine. J. Geophys. Res. Space Phys. 2017, 122, 4210–4227. [Google Scholar] [CrossRef]
  10. Chen, D.; Tian, Y.; Liu, X. Structural nonparallel support vector machine for pattern recognition. Pattern Recognit. 2016, 60, 296–305. [Google Scholar] [CrossRef]
  11. Li, Y.; Wang, X.; Si, S.; Huang, S. Entropy based fault classification using the Case Western Reserve University data: A benchmark study. IEEE Trans. Reliab. 2019, 69, 754–767. [Google Scholar] [CrossRef]
  12. Huang, G.B.; Mao, K.; Siew, C.K.; Huang, D.S. Fast modular network implementation for support vector machines. IEEE Trans. Neural Netw. 2005, 16, 1651–1663. [Google Scholar] [CrossRef] [PubMed]
  13. Joachims, T. Making Large-Scale SVM Learning Practical; Advances in Kernel Methods—Support Vector Learning; Schölkopf, B., Burges, J.C.C., Smola, A.J., Eds.; MIT Press: Cambridge, MA, USA, 1998; pp. 169–184. [Google Scholar] [CrossRef]
  14. Platt, J. Sequential Minimal Optimization: A Fast Algorithm for Training Support Vector Machines. Available online: https://www.microsoft.com/en-us/research/publication/sequential-minimal-optimization-a-fast-algorithm-for-training-support-vector-machines/ (accessed on 11 June 2020).
  15. Keerthi, S.S.; Shevade, S.K.; Bhattacharyya, C.; Murthy, K.R.K. Improvements to Platt’s SMO algorithm for SVM classifier design. Neural Comput. 2001, 13, 637–649. [Google Scholar] [CrossRef]
  16. Suykens, J.A.; Vandewalle, J. Least squares support vector machine classifiers. Neural Process. Lett. 1999, 9, 293–300. [Google Scholar] [CrossRef]
  17. Xu, H.; Chen, G. An intelligent fault identification method of rolling bearings based on LSSVM optimized by improved PSO. Mech. Syst. Signal Process. 2013, 35, 167–175. [Google Scholar] [CrossRef]
  18. Smola, A.J.; Schölkopf, B.; Müller, K.R. The connection between regularization operators and support vector kernels. Neural Netw. 1998, 11, 637–649. [Google Scholar] [CrossRef] [Green Version]
  19. Van, M.; Kang, H.J. Wavelet kernel local fisher discriminant analysis with particle swarm optimization algorithm for bearing defect classification. IEEE Trans. Instrum. Meas. 2015, 64, 3588–3600. [Google Scholar] [CrossRef]
  20. Zhang, L.; Zhou, W.; Jiao, L. Wavelet support vector machine. IEEE Trans. Syst. Man Cybern. Part B (Cybern.) 2004, 34, 34–39. [Google Scholar] [CrossRef] [Green Version]
  21. Mazidi, M.H.; Eshghi, M. Detection of Heart Attack using Cross Wavelet Transformation and Support Vector Machine. Appl. Med. Inform. 2019, 41, 77–92. [Google Scholar]
  22. Xia, K.; He, S.; Tan, Y.; Jiang, Q.; Xu, J.; Yu, W. Wavelet packet and support vector machine analysis of series DC ARC fault detection in photovoltaic system. IEEJ Trans. Electr. Electron. Eng. 2019, 14, 192–200. [Google Scholar] [CrossRef]
  23. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar]
  24. Abdmouleh, Z.; Gastli, A.; Ben-Brahim, L.; Haouari, M.; Al-Emadi, N.A. Review of optimization techniques applied for the integration of distributed generation from renewable energy sources. Renew. Energy 2017, 113, 266–280. [Google Scholar] [CrossRef]
  25. Li, X.; Wu, S.; Li, X.; Yuan, H.; Zhao, D. Particle Swarm Optimization-Support Vector Machine Model for Machinery Fault Diagnoses in High-Voltage Circuit Breakers. Chin. J. Mech. Eng. 2020, 33, 1–10. [Google Scholar] [CrossRef] [Green Version]
  26. Tharwat, A.; Elhoseny, M.; Hassanien, A.E.; Gabel, T.; Kumar, A. Intelligent Bézier curve-based path planning model using Chaotic Particle Swarm Optimization algorithm. Clust. Comput. 2019, 22, 4745–4766. [Google Scholar] [CrossRef]
  27. Sengupta, S.; Basak, S.; Peters, R.A. Particle Swarm Optimization: A survey of historical and recent developments with hybridization perspectives. Mach. Learn. Knowl. Extr. 2019, 1, 157–191. [Google Scholar] [CrossRef] [Green Version]
  28. Huang, N.E.; Shen, Z.; Long, S.R.; Wu, M.C.; Shih, H.H.; Zheng, Q.; Yen, N.C.; Tung, C.C.; Liu, H.H. The empirical mode decomposition and the Hubert spectrum for nonlinear and non-stationary time series analysis. Proc. R. Soc. Math. Phys. Eng. Sci. 1998, 454, 903–995. [Google Scholar] [CrossRef]
  29. Fu, X.; Wang, L. Data dimensionality reduction with application to simplifying RBF network structure and improving classification performance. IEEE Trans. Syst. Man Cybern. Part B (Cybern.) 2003, 33, 399–409. [Google Scholar]
  30. Haykin, S. Neural Networks: A Comprehensive Foundation; Prentice Hall PTR: Upper Saddle River, NJ, USA, 1994. [Google Scholar]
  31. Zhang, Q.; Benveniste, A. Wavelet networks. IEEE Trans. Neural Netw. 1992, 3, 889–898. [Google Scholar] [CrossRef]
  32. Daubechies, I. The wavelet transform, time-frequency localization and signal analysis. IEEE Trans. Inf. Theory 1990, 36, 961–1005. [Google Scholar] [CrossRef] [Green Version]
  33. Szu, H.H.; Telfer, B.A.; Kadambe, S.L. Neural network adaptive wavelets for signal representation and classification. Opt. Eng. 1992, 31, 1907–1917. [Google Scholar] [CrossRef] [Green Version]
  34. Liu, Z.; Cao, H.; Chen, X.; He, Z.; Shen, Z. Multi-fault classification based on wavelet SVM with PSO algorithm to analyze vibration signals from rolling element bearings. Neurocomputing 2013, 99, 399–410. [Google Scholar] [CrossRef]
  35. Loparo, K.A. Bearing Data Center; Case Western Reserve University: Cleveland, OH, USA, 2013. [Google Scholar]
  36. De Brabanter, K.; Karsmakers, P.; Ojeda, F.; Alzate, C.; De Brabanter, J.; Pelckmans, K.; De Moor, B.; Vandewalle, J.; Suykens, J.A. LS-SVMlab Toolbox User’s Guide: Version 1.7; Katholieke Universiteit Leuven: Leuven, Belgium, 2010. [Google Scholar]
  37. Song, L.; Yan, R. Bearing fault diagnosis based on Cluster-contraction Stage-wise Orthogonal-Matching-Pursuit. Measurement 2019, 140, 240–253. [Google Scholar] [CrossRef]
  38. Liu, X.; Zhang, X.; Luan, Z.; Xu, X. Rolling bearing fault diagnosis based on EEMD sample entropy and PNN. J. Eng. 2019, 2019, 8696–8700. [Google Scholar] [CrossRef]
  39. Rodriguez, J.D.; Perez, A.; Lozano, J.A. Sensitivity analysis of k-fold cross validation in prediction error estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 569–575. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Flow chart of the developed strategy for bearing fault diagnosis.
Figure 1. Flow chart of the developed strategy for bearing fault diagnosis.
Sensors 20 03422 g001
Figure 2. ‘One to others’ multi-class fault classification system of PSO-LSWSVM.
Figure 2. ‘One to others’ multi-class fault classification system of PSO-LSWSVM.
Sensors 20 03422 g002
Figure 3. The bearing testbed
Figure 3. The bearing testbed
Sensors 20 03422 g003
Figure 4. The represented (top) vibration signal, and (bottom) de-noised signal using NLM when the bearing in normal operation.
Figure 4. The represented (top) vibration signal, and (bottom) de-noised signal using NLM when the bearing in normal operation.
Sensors 20 03422 g004
Figure 5. Denoising signal using NLM when the bearing in an inner race (IR) fault (0.021 in.).
Figure 5. Denoising signal using NLM when the bearing in an inner race (IR) fault (0.021 in.).
Sensors 20 03422 g005
Figure 6. Denoising signal using NLM when the bearing in an outer race (OR) fault (0.021 in.).
Figure 6. Denoising signal using NLM when the bearing in an outer race (OR) fault (0.021 in.).
Sensors 20 03422 g006
Figure 7. TDenoising signal using NLM when the bearing in a ball (B) fault (0.021 in.).
Figure 7. TDenoising signal using NLM when the bearing in a ball (B) fault (0.021 in.).
Sensors 20 03422 g007
Figure 8. Validation accuracy of the KNN classifier.
Figure 8. Validation accuracy of the KNN classifier.
Sensors 20 03422 g008
Figure 9. Validation accuracy of the PNN classifier.
Figure 9. Validation accuracy of the PNN classifier.
Sensors 20 03422 g009
Figure 10. Validation accuracy of the LSRBFSVM classifier; C = 4.5 , σ = 2.5 .
Figure 10. Validation accuracy of the LSRBFSVM classifier; C = 4.5 , σ = 2.5 .
Sensors 20 03422 g010
Figure 11. Validation accuracy of the PSO-LSRBFSVM classifier.
Figure 11. Validation accuracy of the PSO-LSRBFSVM classifier.
Sensors 20 03422 g011
Figure 12. Validation accuracy of the LSWSVM classifier.
Figure 12. Validation accuracy of the LSWSVM classifier.
Sensors 20 03422 g012
Figure 13. Validation accuracy of the PSO-LSWSVM classifier.
Figure 13. Validation accuracy of the PSO-LSWSVM classifier.
Sensors 20 03422 g013aSensors 20 03422 g013b
Table 1. Time-domain dimensionless parameters.
Table 1. Time-domain dimensionless parameters.
No.FeatureEquationNo.FeatureEquation
1Standard deviation x s t d = n = 1 N ( x ( n ) x m ) 2 N 6Root mean square x r m s = n = 1 N ( x ( n ) ) 2 N
2Peak x p = m a x | x ( n ) | 7Clearance factor C L F = x p 1 N n = 1 N | x ( n ) | 2
3Skewness x s k e w = n = 1 N ( x ( n ) x m ) 3 ( N 1 ) x s t d 3 8Shape factor S F = x r m s 1 N n = 1 N | x ( n ) |
4Kurtosis x k u r = n = 1 N ( x ( n ) x m ) 4 ( N 1 ) x s t d 4 9Impulse factor I F = x p 1 N n = 1 N | x ( n ) |
5Crest factor C F = x p x r m s where x ( n ) is a signal series for n = 1 , 2 , , N .
Table 2. Accuracy comparison (%) among classifiers.
Table 2. Accuracy comparison (%) among classifiers.
ClassifierMeanMaxPositionComputation Time (s)
kNN56.8383.9170.125
PNN70.9285.95170.109
LSRBFSVM71.3791.42110.375
PSO-LSRBFSVM82.8094.762024.49
LSWSVM81.4099.05120.422
PSO-LSWSVM (a)90.151001430.52
PSO-LSWSVM (b)95.97100530.52
PSO-LSWSVM (c)98.14100230.52
Table 3. Confusion matrix for showing classification results of the KNN classifier.
Table 3. Confusion matrix for showing classification results of the KNN classifier.
NMORF1IRF1BF1ORF2IRF2BF2ORF3IRF3BF3
NM145246010111231
ORF188891021545715500
IRF100112621034212877
BF1031011355307100
ORF22140124921373124
IRF2002990211443523
BF2381912011012700280
ORF3020218192111346220
IRF3074112633413770
BF303850331001304001465
Sensitivity(%)96.859.2775.0775.6783.2776.2784.6790.0791.897.67
Specificity(%)99.7298.8597.199.7399.1397.6498.0298.0799.4593.47
Accuracy (%)83.05
Table 4. Confusion matrix for showing classification results of the PNN classifier.
Table 4. Confusion matrix for showing classification results of the PNN classifier.
NMORF1IRF1BF1ORF2IRF2BF2ORF3IRF3BF3
NM1485218452510037
ORF101304435546273521
IRF11531053153051404153
BF1059181290102337169
ORF200510012501426500
IRF2002990511770031
BF214130010101337008
ORF3000363001325013
IRF30264660405010582
BF30670000375001436
Sensitivity(%)99.086.9370.286.083.3378.4789.1388.3370.5395.73
Specificity(%)99.3298.7994.598.8698.4297.7299.9999.4199.4598.86
Accuracy (%)84.77
Table 5. Confusion matrix for showing classification results of the PSO-LSRBFSVM classifier.
Table 5. Confusion matrix for showing classification results of the PSO-LSRBFSVM classifier.
NMORF1IRF1BF1ORF2IRF2BF2ORF3IRF3BF3
NM1483271125013
ORF10122760074614914
IRF1120100555252702011790
BF102011276300025
ORF20158013540416210
IRF2003420112155001
BF21578010013201230
ORF3050008000140859
IRF303177972890113011
BF315707010482301437
Sensitivity(%)98.8781.8067.0085.0790.2781.0088.0093.8786.7395.80
Specificity(%)99.8498.8295.9399.7799.0697.4299.0698.9399.4598.52
Accuracy (%)86.84
Table 6. Confusion matrix for showing classification results of the PSO-LSWSVM classifier.
Table 6. Confusion matrix for showing classification results of the PSO-LSWSVM classifier.
NMORF1IRF1BF1ORF2IRF2BF2ORF3IRF3BF3
NM1500020105002
ORF10148400101020
IRF100121001016400600
BF100251493102004
ORF2002001464033100
IRF2001780213175083
BF2012011181473100
ORF300012000144700
IRF3006500102114210
BF304050011091491
Sensitivity(%)10098.9399.4380.6797.6087.8098.2096.4794.7399.40
Specificity(%)99.9399.9799.9298.2799.6098.5599.7699.8499.4599.79
Accuracy (%)95.33

Share and Cite

MDPI and ACS Style

Van, M.; Hoang, D.T.; Kang, H.J. Bearing Fault Diagnosis Using a Particle Swarm Optimization-Least Squares Wavelet Support Vector Machine Classifier. Sensors 2020, 20, 3422. https://doi.org/10.3390/s20123422

AMA Style

Van M, Hoang DT, Kang HJ. Bearing Fault Diagnosis Using a Particle Swarm Optimization-Least Squares Wavelet Support Vector Machine Classifier. Sensors. 2020; 20(12):3422. https://doi.org/10.3390/s20123422

Chicago/Turabian Style

Van, Mien, Duy Tang Hoang, and Hee Jun Kang. 2020. "Bearing Fault Diagnosis Using a Particle Swarm Optimization-Least Squares Wavelet Support Vector Machine Classifier" Sensors 20, no. 12: 3422. https://doi.org/10.3390/s20123422

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop