Next Article in Journal
Exclusion and Underdetermined Qualia
Previous Article in Journal
Robust Variable Selection and Estimation Based on Kernel Modal Regression
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fault Diagnosis for Rolling Bearings Based on Fine-Sorted Dispersion Entropy and SVM Optimized with Mutation SCA-PSO

1
College of Electrical Engineering & New Energy, China Three Gorges University, Yichang 443002, China
2
Hubei Provincial Key Laboratory for Operation and Control of Cascaded Hydropower Station, China Three Gorges University, Yichang 443002, China
3
School of Hydropower and Information Engineering, Huazhong University of Science and Technology, Wuhan 430074, China
*
Author to whom correspondence should be addressed.
Entropy 2019, 21(4), 404; https://doi.org/10.3390/e21040404
Submission received: 22 March 2019 / Revised: 13 April 2019 / Accepted: 14 April 2019 / Published: 16 April 2019

Abstract

:
Rolling bearings are a vital and widely used component in modern industry, relating to the production efficiency and remaining life of a device. An effective and robust fault diagnosis method for rolling bearings can reduce the downtime caused by unexpected failures. Thus, a novel fault diagnosis method for rolling bearings by fine-sorted dispersion entropy and mutation sine cosine algorithm and particle swarm optimization (SCA-PSO) optimized support vector machine (SVM) is presented to diagnose a fault of various sizes, locations and motor loads. Vibration signals collected from different types of faults are firstly decomposed by variational mode decomposition (VMD) into sets of intrinsic mode functions (IMFs), where the decomposing mode number K is determined by the central frequency observation method, thus, to weaken the non-stationarity of original signals. Later, the improved fine-sorted dispersion entropy (FSDE) is proposed to enhance the perception for relationship information between neighboring elements and then employed to construct the feature vectors of different fault samples. Afterward, a hybrid optimization strategy combining advantages of mutation operator, sine cosine algorithm and particle swarm optimization (MSCAPSO) is proposed to optimize the SVM model. The optimal SVM model is subsequently applied to realize the pattern recognition for different fault samples. The superiority of the proposed method is assessed through multiple contrastive experiments. Result analysis indicates that the proposed method achieves better precision and stability over some relevant methods, whereupon it is promising in the field of fault diagnosis for rolling bearings.

1. Introduction

Rolling bearings are a crucial part in modern industrial manufacture, which can be found in linear guides, precision machine tools, and engine parts, etc., whose failure may result in serious safety accidents and economic loss. Therefore, one of the major topics to be investigated in preventing failures of mechanical systems is recognizing and diagnosing rolling bearing faults [1,2]. Unfortunately, rolling bearings are prone to failure to some extent because of complex operating conditions and structural characteristics [3]. Given this fact, effective and feasible fault diagnosis methods need to be developed for rolling bearings, thus, to improve the reliability of mechanical systems [4].
Due to the rich information carried by vibration signal, most of fault diagnosis methods for rolling bearings rely on analyzing vibration signal [5]. However, the vibration signal is commonly non-stationary in actual operation, which restricts the extraction efficiency for fault features and affects the accuracy of fault diagnosis. Addressing this issue, many time-frequency signal analysis methods have been proposed to analyze non-stationary signals, including wavelet transform (WT), empirical mode decomposition (EMD), ensemble empirical mode decomposition (EEMD), and variational mode decomposition (VMD). WT is a self-adaptive signal decomposition method, which has a good frequency resolution and time resolution for low-frequency components and provides a good frequency resolution and time resolution for high-frequency components [6], but the selection of an appropriate wavelet filter bank remains a challenge. EMD is adept in dealing with non-stationary signals and does not need to pre-set any basis function [7], but its performance is affected by end effects and mode mixing. To overcome these defects, as an improved version of EMD, EEMD was proposed by introducing noise-assisted analysis method [8], but it also increases the computational cost and cannot completely neutralize the added noise. Unlike the methods mentioned above, VMD is a quasi-orthogonal signal decomposition method, the main structure of which is to construct and solve a constrained variational problem [9]. Hence it realizes the separation of the frequency of each signal component, avoiding the problem of mode mixing in EMD and EEMD or filter selection in WT. Its advancement and effectiveness have been demonstrated in many previous studies. For instance, Fu et al. [10] applied VMD to decompose the wind speed time series into a set of sub-series with various frequency scales and then calculated the final predicted value of the raw wind speed by accumulating the predicted results of all components. Li et al. [11] established the approximate entropy feature vectors for each component of VMD and obtained a good fault diagnosis effect for rolling bearings. Fu et al. [12] achieved the feature extraction of signals by conducting multivariate autoregressive modelling for components of VMD.
As the non-stationarity of signals weakened by signal analysis methods, fault features can be smoothly extracted. Entropy, as a feature extraction method based on non-linear dynamic parameters, is widely applied in medical and mechanical fields, which can reflect the complexity of time series and is an effective tool for analysis of non-stationary time series. Approximate entropy (ApEn) [13] is the pioneer of introducing a feature extraction algorithm from the medical field into mechanical fault diagnosis, while it is susceptible to outliers. Permutation entropy (PE) [14] has strong anti-noise ability and is simple in calculation, but it does not take into account the difference between the average amplitude and actual amplitude. Fuzzy entropy (FE) [15] introduces the concept of uncertainty reasoning to solve the defect of sample entropy [16] which adopts a hard threshold as a discriminant criterion which may result in unstable results. This approach, however, also lowers the computational efficiency. Recently, a novel irregularity indicator termed dispersion entropy (DE), was proposed by Mostafa and Hamed [17]. DE has the advantages of fast calculation speed and less influence from outliers, which overcomes the defects of ApEn and PE to some extent, but the relationship information between neighboring amplitudes is not considered enough. In this paper, a concept of fine-sorted dispersion entropy (FSDE) is proposed to measure the features of vibration data and then construct fault feature vectors. By introducing an additional factor, FSDE is more perceptive to element values and may have better feature extraction ability than DE.
The essence of fault diagnosis for rolling bearings is actually a pattern recognition problem, that is, fault type is determined by the fault features of samples. In an engineering application, there are many alternative methods available for solving this problem. For instance, Bayesian decision has strong recognition ability by considering prior probability as well as class conditional probability, and its good accuracy requires the assumption of a suitable prior model [18]. K-nearest neighbor (KNN) based on Euclidean or Manhattan distances is easy to realize and susceptible to the distribution of samples [19]. Artificial neural network (ANN) has good ability in processing recognition problem with the support of a large number of samples [20]. Based on statistical learning and structural risk minimization theory, support vector machine (SVM) is an excellent machine learning method, which finds an optimal hyperplane that meets the classification requirements. The hyperplane maximizes the interval between two classes while promoting classification performance. Currently, SVM combined with the feature extraction method has been successfully employed in the field of pattern recognition [21,22].
Although SVM possesses superior ability in pattern recognition, its performance is affected by parameters. In view of this, various optimization algorithms have been developed and applied to search the best parameters, such as particle swarm optimization (PSO) [23], bacterial foraging algorithm (BFA) [24], artificial sheep algorithm (ASA) [25] and sine cosine algorithm (SCA) [26]. As a novel optimization approach, the effectiveness of SCA in parameter optimization has been proved in many previous studies [27,28]. To achieve better performance in convergence precision, a hybrid optimization strategy combining mutation operator, sine cosine algorithm, and particle swarm optimization (MSCAPSO) is developed in this paper. VMD was firstly applied to decompose the vibration signal into a set of intrinsic mode functions (IMFs), and the decomposing mode number K was pre-set with central frequency observation method. Then a modified dispersion entropy containing the additional factor (FSDE) is proposed to construct the feature vectors of IMFs for different types of faults. Afterward, a hybrid optimization strategy combining mutation operator, SCA and PSO called MSCAPSO was proposed for parameter selection of SVM. Finally, the optimal SVM model was employed to classify different fault samples. The engineering application and contrastive analysis reveal the availability and superiority of the proposed method.
The paper is organized as follows: Section 2 is dedicated to the basic knowledge of VMD and SVM. Section 3 introduces the proposed fault diagnosis method based on fine-sorted dispersion entropy and MSCA-PSO optimization approach. Section 4 presents the engineering application, the result analysis of which demonstrates the superiority of the proposed method. Some discussion about the method presented in this paper is in Section 5. The conclusion is summarized in Section 6.

2. Fundamental Theories

2.1. Variational Mode Decomposition

Variational mode decomposition (VMD) is a new adaptive signal preprocessing method [29]. By setting a mode number K in advance, the given signal can be decomposed into K band-limited intrinsic mode functions (IMFs). The cyclic iterative method is used to obtain the optimal solution of the constrained variational problem, through which the frequency center and bandwidth of each mode are determined. The constrained variation problem can be described as follows:
min m k , ω k { k t [ ( δ ( t ) + j π t ) m k ( t ) ] e j ω k t 2 2 } s . t .   k = 1 K m k ( t ) = f ( t ) ,   k = 1 , 2 , , K
where mk = [m1, m2, …, mk] represents the set of K mode functions and ω k = [ ω 1 , ω 2 , , ω k ] represents the set of central frequencies, while t and δt are the partial derivative of time t for the function and unit pulse function, respectively. f(t) is the given real valued input signal.
For the above variational problem, quadratic penalty function term and Lagrange multiplier are employed to transform it into an unconstrained problem. Then Problem (1) can be specified as follows:
L ( m k , ω k , β ) = α k t [ ( δ ( t ) + j π t ) m k ( t ) ] e j ω k t 2 2 + f ( t ) k m k ( t ) 2 2 + β ( t ) , f ( t ) k m k ( t )
where α and β(t) represent the penalty factor and Lagrange multiplier, respectively.
Then alternate direction method is utilized to obtain the saddle point of Lagrange multiplier, that is, solving (2) by optimizing mk, ωk, and β alternately. The optimization problems of mk and ωk are formulated as (3) and (4), respectively.
m k n + 1 = min { α t [ ( δ ( t ) + j π t ) m k ( t ) ] e j ω k t 2 2 + f ( t ) i m i ( t ) + β ( t ) 2 2 2 }
ω k n + 1 = min { t [ ( δ ( t ) + j π t ) m k ( t ) ] e j ω k t 2 2 }
Solving problems (3) and (4), the iterative equations are deduced as follows:
m ^ k n + 1 ( ω ) = f ^ ( ω ) i k m ^ i ( ω ) + β ^ ( ω ) 2 1 + 2 α ( ω ω k ) 2
ω k n + 1 = 0 ω | m ^ k ( ω ) | 2 d ω 0 | m ^ k ( ω ) | 2 d ω
The Lagrange multipliers can be iterated with Equation (7).
β ^ n + 1 ( ω ) = β ^ n ( ω ) + γ ( f ( ω ) k m ^ k n + 1 ( ω ) )
where γ is an updating parameter.
The main steps of VMD can be summarized as follows:
Step 1: Initialize mk1, ωk1, β1, n = 1;
Step 2: Execute loop, n = n + 1;
Step 3: Update mk and ωk based on Equations (5) and (6);
Step 4: Update β based on Equation (7);
Step 5: If k m ^ k n + 1 m ^ k n 2 2 / m ^ k n 2 2 < ε loop end, else turn to step 2 for next iteration.

2.2. Support Vector Machine

Support vector machine (SVM) is a machine learning model developed by Vapnik [30], which can deduce the optimal solution between model complexity and learning ability based on limited information, thus, to obtain the best classification accuracy, showing unique advantages in solving learning problems with limited samples, non-linear, and high dimensional data. For a given sample set { ( x i , y i ) | i = 1 ,   2 , , n } , the main idea of SVM is to map samples into a high-dimensional feature space and find a hyper-plane to solve the corresponding non-linear problem in low-dimensional space. The hyper-plane function can be constructed as:
w x + b = 0
where w is a weight vector, and b is a bias parameter, while w·x represents the inner product of w and x.
Taking binary classification issue as an example, to classify samples correctly with enough into classification intervals, all samples are required to meet the following constraint:
w x i + b { > 1   f o r   y i = 1 < 1   f o r   y i = 1
In the sequel, the classification interval can be calculated as 2/‖w2, and maximizing the interval is equivalent to minimizing ‖w2. To solve the linear indivisibility problem, the slack term ξ and penalty factor C are brought into Equation (9), hence the construction of the optimal hyper-plane is transformed into a constrained optimization problem:
{ min f = 1 2 w 2 + C i = 1 n ξ i s . t .   y i ( w T x i + b ) 1 ξ i ,   i = 1 , 2 , , n
When mapping samples to high-dimensional space, different inner product kernel functions will form different algorithms. Among them, radial basis function is widely employed in the application of pattern recognition, which can be described as:
K ( x i , x j ) = ϕ ( x i ) ϕ ( x j ) = exp ( g x i x j 2 )
where g is a kernel parameter.
The solution of the constrained optimization problem (10) is determined by a saddle point of Lagrange function, and this quadratic programming problem is transformed into the corresponding dual problem, that is:
max L = i = 1 n μ i 1 2 i , j = 1 n μ i μ j y i y j K ( x i , x j ) s . t .   i = 1 n μ i y i = 0 ,   μ i 0 ,   i = 1 , 2 , , n
where µi is Lagrange multiplier.
With the optimal value of µi obtained from the above dual problem, the optimal classification discriminant function can be ascertained as:
f ( x ) = s g n ( i = 1 n μ i K ( x i , x ) + b )

3. Fault Diagnosis for Rolling Bearings by Fine-sorted Dispersion Entropy and Mutation SCA-PSO Optimized SVM

3.1. Fine-Sorted Dispersion Entropy

3.1.1. Dispersion Entropy

For a time series x = [x1, x2, …, xn] of length n, xi is firstly normalized by a mapping function, where the standard normal cumulative distribution function is generally used [17].
y i = 1 σ 2 π x i e ( s μ ) 2 2 σ 2 d s
where i = 1, 2, …, n, σ, and μ are the variance and expectation of the normal distribution, respectively. The time series x is then normalized to y = [y1, y2, …, yn], yi ∈ (0, 1). Further, the phase space is reconstructed for y:
y j m = [ y j , y j + τ , , y j + ( m 1 ) τ ]
where j = 1, 2, …, n − (m − 1)τ, m and τ are the embedding dimension and time delay, respectively. Then y j m is mapped to c class (1 to c):
z i c = r o u n d ( c y i + 0.5 )
z j m , c = [ z j c , z j + τ c , , z j + ( m 1 ) τ c ]
where z i c is the i-th element of class sequence zc, while round means rounding. Each z j m , c corresponds to a dispersion pattern π v 0 v 1 v m 1 with elements z j c = v 0 ,   z j + d c = v 1 , ,   z j + ( m 1 ) τ c = v m 1 . Since π v 0 v 1 v m 1 contains m elements, each element contains c values, so the number of all possible dispersion patterns are cm.
The relative frequency of π v 0 v 1 v m 1 can be deduced as:
p = N u m b e r { j | j n ( m 1 ) τ , π v 0 v 1 v m 1 } n ( m 1 ) τ
where N u m b e r { j | j n ( m 1 ) τ , π v 0 v 1 v m 1 } is the number of occurrences of each π v 0 v 1 v m 1 corresponding to z j m , c :
Dispersion entropy can be defined according to information entropy theory:
D E ( x , m , c , τ ) = π = 1 c m p ln ( p )
A large DE value means that the time series is more irregular; on the contrary, the smaller the DE value, the lower the irregularity.

3.1.2. Fine-Sorted Dispersion Entropy

Dispersion entropy, as an information entropy feature, can effectively monitor the dynamic changes in vibration signal and measure its irregularity, so it can be effectively applied to the state monitoring of rolling bearings. However, its ability to feature extraction for the complex non-stationary signal can be further improved in some respects. One possible improvement would be to enhance the perception of neighboring element values. In the theory of dispersion entropy, dispersion entropy focuses on the class sequence that mapping the elements of time series into positive integers. The relationship between elements, as comparatively critical information, has not been considered enough, which suggests that the degree of difference between neighboring elements is not concerned. According to the mapping rule of dispersion entropy, the same dispersion pattern may come from multiple forms of sample vectors. Figure 1 is an example of possible vectors corresponding to the same dispersion pattern (dispersion pattern [2 2 3 3]).
To enhance the feature extraction ability of dispersion entropy, an improved dispersion entropy called fine-sorted dispersion entropy (FSDE) is proposed. FSDE introduces a factor f to measure the difference between neighboring elements in the vector y j m . The factor f is then added in class sequence z j m , c as an additional element. Therefore, the relationship information between elements is brought into dispersion entropy. This operation not only inherits the features within the result of original DE but also improves the case that vectors with a distinct trend are assigned to the same class sequence, which may provide a powerful aid to feature extraction process of FSDE [31,32]. The calculation equation for factor f is as follows:
f = max ( | d y | ) ρ s t d ( | d x | )
where returns the largest integer that is less than or equal to the value in brackets, d x = { x i + 1 x i | i = 1 , 2 , , n 1 } , d y = { y i + 1 y i | i = 1 , 2 , , m 1 } , std represent the standard deviation and ρ is a precision parameter. Obviously, if ρ > max ( | d y | ) s t d ( | d x | ) , the value of f can only be 0, and this additional element will not affect the results of dispersion pattern, which means that the value of FSDE will be exactly the same as original DE. If ρ ( 0 , max ( | d y | ) s t d ( | d x | ) ] , f will have more possible values, the closer ρ gets to 0, the original dispersion pattern will be subdivided into more new dispersion patterns. As an example, time series [0.5114, 0.5131, 05134, 0.5130], [0.5012, 0.5085, 0.5152, 0.5212] and [0.5326, 0.5249, 0.5158, 0.5072] are constructed with τ = 1, m = 4, ρ = 1, c = 6. These time series first map to class sequences according to the mapping strategy of DE, i.e., [4, 4, 4, 4], [4, 4, 4, 4], and [4, 4, 4, 4]. It can be seen that all the time series are mapped to the same dispersion pattern, although their trends are different. Meanwhile, the time series are mapped by the proposed FSDE, factor f is added as an additional element to the end of the class sequence, the new class sequences are [4, 4, 4, 4, 0], [4, 4, 4, 4, 1], and [4, 4, 4, 4, 2]. In the example, the time series originally mapped to the same dispersion pattern are finely mapped to more dispersion patterns. Figure 2 illustrates the way that f is added to the sequence z j m , c as an additional element.
After obtaining the new class sequence z j m , c , π v 0 v 1 v m 1 denotes a dispersion pattern corresponding to a z j m , c . The relative frequency of new dispersion pattern is calculated as:
p = N u m b e r { j | j n ( m 1 ) τ , π v 0 v 1 v m 1 } n ( m 1 ) τ
Then, similar to the definition of dispersion entropy, the proposed fine-sorted dispersion entropy can be defined as:
F S D E ( x , m , c , τ , f ) = π = 1 c m p ln ( p )

3.2. Mutation SCA-PSO Optimization

3.2.1. Sine Cosine Algorithm

Sine cosine algorithm (SCA) is a new swarm optimization algorithm, which is simple in structure and easy to realize. Using only sine and cosine functions, SCA achieves a good optimization effect. Exploration and exploitation are the two key phases when SCA processes optimization a problem (see [26] for more details). With an elaborate mechanism of exploration and exploitation, SCA is able to search for feasible solutions quickly in the searching space. Let Zi = (Zil, Zi2, …, Zid)T be the position of i-th individual, whereupon each solution of the optimization problem corresponds to a position of a corresponding individual in the searching space, where d is the dimension of individuals. The core updating equation of SCA is as follows:
Z i ( l + 1 ) = { Z i ( l ) + r 1 × sin ( r 2 ) × | r 3 P i ( l ) Z i ( l ) | ,   r 4 < 0.5 Z i ( l ) + r 1 × cos ( r 2 ) × | r 3 P i ( l ) Z i ( l ) | ,   r 4 0.5
where Pi = (Pil, Pi2, …, Pid)T is the best position of individual i. The parameter r1 = al(a/lmax) determine the region of individual i at next iteration, where lmax, l, and a are the maximum number of iterations, the current number of iterations, and a constant, respectively. The parameter r2 is a random number in the scope of [0, 2π], which defines the distance that the next movement of individuals should be towards. To stochastically emphasize (r3 > 1) or deemphasize (r3 < 1) the effect of the best position searched so far, a random weight r3 within [0, 2] is brought into the equation. Finally, the parameter r4 is a random number in the range of [0, 1] to switch fairly between sine and cosine components.

3.2.2. Particle Swarm Optimization

Particle swarm optimization (PSO) is a classical stochastic optimization algorithm based on swarm intelligence. In the PSO algorithm, each particle corresponds to a potential solution of the problem to be optimized [33]. Particle Pi has its own velocity and position, which are expressed as d-dimensional vectors vi = (vil, vi2, …, vid)T and si = (sil, si2, …, sid)T, respectively. Particle Pi keeps track of the optimal solution it has found so far, namely, individual extreme value Pibest. The optimal solution found by the whole population is called the global extreme value Pigbest. Particles of the swarm update themselves by tracking the above two extremes values. PSO generates a group of particles randomly to initiate the optimization and then solves the problem by iterations. In the l-th iteration, particle Pi updates its velocity and position according to the following two equations:
v i ( l + 1 ) = w v i ( l ) + c 1 r a n d ( P i b e s t P i ( l ) ) + c 2 r a n d ( P i g b e s t P i ( l ) )
s i ( l + 1 ) = s i ( l ) + v i ( l )
where w is the inertia weight which is employed to balance the global and local search ability of the algorithm. c1 and c2 are learning factors, rand is a random number between (0, 1), l is the current number of iterations.

3.2.3. Mutation SCA-PSO Optimization

In the updating strategy of SCA, sine and cosine functions in Equation (23) possess the outstanding ability of exploration but also have some drawbacks, such as slow convergence and the likelihood of being trapped in a local optimum. The reason for the above problems is that parameters of SCA are difficult to set properly, so a poor combination of parameters will result in weak exploitation, but its exploration phase will not be affected [34]. On the other hand, the PSO algorithm has the advantages of information exchange among particles and good robustness, which lead to the high probability of exploiting local optimal solution. However, similar to SCA, PSO also has the problem of local optimum in the later stage of iterations.
Since SCA and PSO have their own strengths, the combination of capabilities between SCA and PSO is expected to improve the convergence performance, which may make the hybrid algorithm powerful in both exploring different regions and exploiting promising regions. As shown in Figure 3, the SCA-PSO approach is constructed in a hierarchical form. The top layer contains M particles which are searching solutions according to the strategy of PSO. The bottom layer contains M·N individuals which are evenly distributed to each top layer particle (total M groups) and their positions update by SCA strategy.
In iterations, the optimal value found by each group in the bottom layer is saved by a top layer particle. Through this operation, the top layer particles focus on exploitation by using advantages of PSO, while the bottom layer individuals focus on exploration, giving play to advantages of SCA [34]. The updating strategy of bottom layer individuals is changed to the following equation:
Z i j l + 1 = { Z i j l + r 1 × sin ( r 2 ) × | r 3 s i l Z i j l | ,   r 4 < 0.5 Z i j l + r 1 × cos ( r 2 ) × | r 3 s i l Z i j l | ,   r 4 0.5
where Zijl is the position of j-th (j = 1, 2, …, N) bottom layer individual that belongs to the i-th (i = 1, 2, …, M) top layer particle. The position of i-th top layer particle is described as sil. l is the current number of iterations.
Moreover, analysis of PSO shows that the population diversity of PSO gradually decreasing with each iteration and it is more likely to result in a local optimum in the later stage of iterations. To solve this problem, a mutation operator is introduced to enrich the diversity of PSO particles [35]. The updating rules of top layer particles containing the mutation operators are as follows:
v i l = w v i l + c 1 r a n d ( s i b e s t s i l ) + c 2 r a n d ( s g b e s t s i l )
s i l + 1 = { s i l + v i l , l n T s i l [ 1 + G ( 0.5 u n i f ) ] , l = n T n = 1 , 2 ,
where G and T (T < lmax) are the given mutation amplitude and mutation period, respectively, while unif represents a random number that conforms to the uniform distribution U (0, 1). It can be observed that during the updating process, the position si will mutate periodically, which gives PSO particles a certain ability to jump out of the local optimum even in the later stage of iterations.
The computational complexity of the proposed mutation sine cosine algorithm and particle swarm optimization (SCA-PSO) algorithm is O(lM(Ntsca+tpso)), where tsca and tpso are the computational costs of updating all the search agents in each iteration for SCA and PSO, respectively. Accordingly, the computational complexity of SCA and PSO are O(lntsca) and O(lntpso), respectively.
The main steps of proposed mutation SCA-PSO (MSCAPSO) approach are as below, and its block diagram is shown in Figure 4:
Step 1: Initialize M particles in the top layer as well as M·N individuals in the bottom layer randomly and set mutation parameters;
Step 2: Calculate the fitness values of all searching agents;
Step 3: Update si on the basis of the best solution found by each group in the bottom layer according to Equation (26);
Step 4: Update sibest in top layer based on Equations (27) and (28);
Step 5: If the maximum number of iterations is not meet, go to step 3;
Step 6: Return the value of sigbest as the optimal solution.

3.3. Fault Diagnosis by FSDE and MSCAPSO Optimized SVM

In this section, a novel fault diagnosis method based on FSDE and SVM optimized by MSCAPSO is proposed to improve the fault diagnosis accuracy. The specific steps are detailed as follows:
Step 1: Determine the mode number K of VMD by central frequency observation method;
Step 2: Decompose samples into sets of IMFs with VMD;
Step 3: Calculate the FSDE value of each IMF;
Step 4: Construct feature vectors of different fault samples with FSDE values;
Step 5: Optimize parameters C and g for SVM with the proposed MSCAPSO optimization method;
Step 6: Train the SVM model with optimal parameters C and g;
Step 7: Apply the optimal SVM model to classify different fault samples and evaluate the performance of the model.
The flowchart of the proposed fault diagnosis model is shown in Figure 5.

4. Engineering Application

4.1. Data Collection

Vibration signals with different fault locations and sizes gathered from Bearings Data Center of Case Western Reserve University (CWRU) [36] were employed as the experiment data in this research. The experiment stand was mainly composed of a motor, an accelerometer, a torque sensor/encoder and a dynamometer. The bearing was a deep groove ball bearing with model SKF6205-2RS. Accelerometers were applied to collect vibration signal. Meanwhile, the torque sensor/encoder was for the measurement of power and speed. By using electro-discharge machining, the experiment device simulated three fault states of the rolling bearing: inner race fault, ball element fault, and outer race fault, and the depth of faults was 0.011 inches. In the experiment, vibration signals collected from the drive end were taken as the research objects, where the sample frequency was 12 kHz. To fully verify the effectiveness of the proposed fault diagnosis method, nine types samples were applied in this paper, namely inner race fault, ball fault, and outer race fault with diameters of 0.007, 0.014, and 0.021 inches. In addition, the experiments are conducted under the load of 2, 1, and 0 hp with motor speed 1750, 1772, and 1797 rpm. Further, the vibration signal of each type of faults contained 59 samples, and each sample contained 2048 sample points. The experimental data are listed in Table 1.

4.2. Application to Fault Diagnosis of Rolling Bearings

The proposed FSDE-MSCAPSO method has been applied to diagnose nine types of faults and compared with other six relevant fault diagnosis methods at feature extraction as well as parameter optimization stage. More specifically, fuzzy entropy (FE) [37], permutation entropy (PE) [4], and dispersion entropy (DE) were applied at the feature extraction stage for comparison; SCA and PSO were applied at the parameter optimization stage. The parameters of the contrastive experiments were set in the same way as the proposed method.
In the proposed method, the first step is to decompose the signal of each type of fault into a set of IMFs. The mode number K of VMD needs to be determined in advance with the central frequency observation method. If K is too large, central frequencies of neighboring IMFs may be too close, then mode mixing will emerge. Normalized central frequencies under different K values are listed in Table 2, where K was ascertained by the sample of ball fault with a diameter of 0.014 inches (L5). As listed in Table 2, when K was set 5, the first two central frequencies were relatively close. Similarly, as illustrated in Figure 6, during the iterative calculation of VMD, if K was 5 or greater, there were some central frequencies of IMFs close to each other, which meant excessive decomposition had occurred. Hence, parameter K was set to 4.
The waveforms of the original signals are shown in Figure 7. It can be observed that waveforms of different fault positions or different defect sizes are different to a certain extent. The examples of decomposition results with different fault positions (L2, L5, L8) and different defect sizes (L4, L5, L6) are illustrated in Figure 8, which shows that IMFs decomposed from the same types of faults are quite different.
When the IMFs of all samples were obtained, FSDE values were calculated to construct fault feature vectors, during which four parameters needed to be chosen properly, namely, embedding dimension m, number of class c, time delay τ, and precision parameter ρ. Theoretically, m is usually set from two to seven, if m is large, there will be a high computational cost. When dealing with c, it is usually suggested to choose from four to eight, if c is small, it may not be able to recognize two values with very different amplitudes [38]. ρ is the key parameter that distinguishes FSDE from DE. The closer ρ to 0, an original dispersion pattern will be subdivided into more new dispersion patterns, however, FSDE may also be oversensitive to small changes and increase the computation. Considering all these factors, embedding dimension m was set at three and the number of class c was set at six in our research. Moreover, for the practical purpose, it was recommended to assume the time delay τ = 1 [38], while precision parameter ρ was set as 2.9 to balance the precision and computational complexity.
The first three FSDE and DE vectors of different fault samples (L1–L9) are compared in Table 3, which shows that FSDE value of every IMF is different from the corresponding DE value, but the overall trend remains similar. 3D projection of entropy vectors for different types of faults is shown in Figure 9. It can be seen that feature vectors extracted by PE are not clearly distinguished, quite a lot of feature vectors from different types of faults are still mixed together. This situation is improved in the figures of FE and DE to a certain extent, while feature vectors extracted by FSDE are basically well distinguished.
In the optimization stage, the feature vectors from different types of faults were divided randomly into two parts with 30 and 29 vectors, where 30 were used for training while the rest, 29, for testing. The proposed MSCAPSO approach was applied to optimize the penalty factor C and the kernel parameter g of SVM. The total number of searching agents and iteration times were set 40 and 100, respectively, while searching ranges of C and g were both [2−10, 210]. Considering the maximum number of iterations and the number of optimization parameters, the mutation period was set at five and mutation amplitude was set at one. Fitness values were calculated by five-fold cross-validation for the training samples. Hence, the optimization effect of the given C and g could be measured by the average accuracy of cross-validation, then the optimal C and g were obtained.
The convergence procedure of MSCAPSO is depicted in Figure 10a, from which it can be observed that the average fitness value of particles rises rapidly at an early stage of the iterations and then tends to be stable. That is, the particles are close to the global optimal solution. The shaded part shows the distribution of convergence curves in 10 runs. The comparison of different optimization methods is shown in Figure 10b, where all convergence experiments are based on the same FSDE feature vectors, and each convergence curve is the average of 10 runs. The figure shows that the convergence curve of PSO fluctuates greatly, and the average fitness value is difficult to be further improved after 50 iterations. SCA has a good overall convergence trend, but the convergence rate is significantly lower than the other two methods. MSCAPSO has already converged to a high level in the early stage and keeps the best convergence effect among all methods in the whole iterations.
With the optimal parameters C and g, the SVM model optimized by MSCAPSO was trained and applied to recognize testing samples. For an in-depth verification of the proposed method, diagnosis results were averaged over 10 runs, and training samples were chosen at random in every run. Furthermore, the deviation of each result was calculated as a reference for result evaluation. In this application, adjusted Rand index (ARI), normalized mutual information (NMI), F-measure (F), and accuracy (ACC) as four widely used evaluation metrics are employed to evaluate the diagnosis results [39,40], which reflects the matching degree between classification result and real distribution information of samples. The scope of ARI is [−1, 1] and the rest scopes are all [0, 1]. A higher value indicates a better classification performance. The optimal C and g are presented corresponding to the best accuracy among 10 runs.
Fault diagnosis results with different methods under variable load conditions are shown in Table 4 and Figure 11. It can be observed from Table 4 that four evaluation metrics of the proposed FSDE-MSCAPSO method are all the highest when the motor loads are 2 hp and 1 hp, i.e., 0.9637, 0.9646, 0.9839, 0.9839 and 0.9615, 0.9598, 0.9827, 0.9828. As well, the deviations of the evaluation values are also at a very small level among all methods. When the motor load is 0 hp, although the value of proposed method is slightly lower than FE-PSO in each metric, its performance is still better than DE-PSO/SCA methods than before improvements. To be specific, the comparison of FE-PSO, DE-PSO and FSDE-PSO indicates that the proposed FSDE has a better feature extraction ability than FE, PE and DE in the field of fault diagnosis. Similarly, this fact is also verified in the comparison among PE-SCA, DE-SCA, and FSDE-SCA. Moreover, it can be concluded from the contrastive experiments between FSDE-SCA/PSO and FSDE-MSCAPSO that the proposed MSCAPSO optimization strategy improves the diagnosis performance of SVM.
The boxplots of evaluation values are shown in Figure 12, illustrating the performances of different diagnostic methods. As shown in the figure, the proposed method achieves better overall performance and stability for fault diagnosis than other contrastive methods. Furthermore, several statistical metrics that have been widely used for robustness evaluation [10,41,42] are applied to demonstrate the robustness of the diagnosis models, i.e., mean absolute error (MAE), root mean square error (RMSE), and mean absolute percentage error (MAPE). The error metrics of the outputs of different models under various loads are listed in Table 5, as well as the corresponding histograms shown in Figure 13. It can be observed that error metrics of FSDE-MSCAPSO are the lowest when the motor loads are 2 hp and 1 hp. When the motor load is 0 hp, although the metric values of the proposed method are not the lowest among all models, it is still at a low level and in the second position in each metric, indicating the proposed model has good robustness on the whole.

5. Discussion

After the detailed comparative analysis above, the superiority of the proposed FSDE and MSCAPSO methods have been effectively demonstrated, as well, the accuracy of the diagnosis model has also been verified by experiments on various loads, fault sizes, and locations. It is worth noting that the proposed FSDE method only considers the relationship between neighboring elements in a given series, and the overall information of the series is not taken into account. According to the previous references [31,43], the improved method using the overall amplitude information of the series has been realized on other entropy methods, which is expected to further improve the performance of the proposed method. Furthermore, some of the parameters in the FSDE are still empirical values in the present paper. In future studies, reasonable metrics can be introduced to evaluate the influence of parameters on the approximation between FSDE and DE, so as to better determine the value of parameters. Furthermore, the proposed MSCAPSO algorithm can be optimized through better strategies to further improve the convergence speed and global search ability. In addition, the multi-objective optimization that has been widely utilized in the field of controlling [44,45,46,47] could be implemented in fault diagnosis, which is expected to improve the accuracy and reduce the variance of the outputs of the model [48]. Moreover, although the diagnosis experiments considered various load conditions, fault sizes, and locations, the signal to noise ratio of the CWRU bearing data is relatively high, and the conclusions obtained in the experiment may not be very comprehensive. In the future work, we plan to build our own bearing test device to collect many more vibration signals, and the results obtained under such conditions may be more practical.
Considering that the CWRU bearing data set has been selected by many researchers as experimental data to validate their methods and techniques, the diagnosis accuracies of some other methods based on this data set are listed in Table 6 to compare with the methods presented in this paper. Note that the segmentation or fault types of experimental samples may not be exactly the same, the accuracies listed is for general description.

6. Conclusions

A novel fault diagnosis method for rolling bearings by fine-sorted dispersion entropy and mutation SCA-PSO optimized SVM is presented to diagnose faults of various sizes, locations, and motor loads in this study. Due to the non-stationarity of original vibration signals, signals collected from different types of faults were firstly split by VMD into sets of IMFs, before which the decomposing mode number K was determined by a central frequency observation method. Then an improved dispersion entropy containing an additional factor, termed FSDE, was proposed to enhance the perception for relationship information between neighboring elements. It was employed to construct the feature vectors of different fault samples later. Afterward, a hybrid optimization strategy combining advantages of mutation operator, SCA and PSO called MSCAPSO, was proposed to optimize the SVM model. The optimal SVM model was subsequently applied to realize the classification for different fault samples. The contrastive experiments in Section 4 were implemented among the proposed FSDE-MSCAPSO method and other relevant methods, where nine types of faults with different locations and sizes were successfully diagnosed. The analysis of results indicates that the proposed method achieves the highest 0.9637, 0.9646, 0.9839, and 0.9839 in four widely used evaluation metrics, which exhibits its superior precision and stability over other methods.

Author Contributions

W.F. conceived the research, J.T. performed the experiments and was responsible for paper writing, Y.X. provided guidance for experiments and contributed to revise the paper, K.W. helped analyze results, T.C. provided recommendations for this research.

Funding

This work was funded by the National Natural Science Foundation of China (NSFC) [51741907, 51809099], Hubei Provincial Major Project for Technical Innovation [2017AAA132], the Open Fund of Hubei Provincial Key Laboratory for Operation and Control of Cascaded Hydropower Station [2017KJX06], Research Fund for Excellent Dissertation of China Three Gorges University [2019SSPY070].

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

SymbolsAbbreviations
Kmode numberApEnapproximate entropy
mkset of K mode functionsANNartificial neural network
ωkset of central frequenciesASAartificial sheep algorithm
tpartial derivative of time tBFAbacterial foraging algorithm
δtunit pulse function of time tBPNNback-propagation neural network
f(t)real valued input signalCWRUCase Western Reserve University
αpenalty factor for VMDCQSCAchaos quantum sine cosine algorithm
β(t)Lagrange multiplier for VMDDEdispersion entropy
wweight vectorEMDempirical mode decomposition
bbias parameterEEMDensemble empirical mode decomposition
ξslack termEWTempirical wavelet transforms
Cpenalty factor for SVMELMextreme learning machine
gkernel parameterFEfuzzy entropy
μiLagrange multiplier for SVMFSDEfine-sorted dispersion entropy
xtime seriesGCMPEgeneralized composite multiscale permutation entropy
nlength of time seriesHGSAhybrid gravitational search algorithm
σvariance of the normal distributionIMPEimproved multiscale permutation entropy
μexpectation of the normal distributionIMDEimproved multiscale dispersion entropy
y j m reconstruction matrix of xIMFintrinsic mode function
membedding dimensionKNNk-nearest neighbour
τtime delayLapSVMLaplacian support vector machine
cnumber of classLSSVMleast squares support vector machine
z j m , c class sequence of DEMSCAPSOmutation sine cosine algorithm and particle swarm optimization
π v 0 v 1 v m 1 dispersion pattern of DEMHGWOSCAmutation hybrid grey wolf optimizer and sine cosine algorithm
ffactor of FSDEmRMRmax-relevance min-redundancy
ρprecision parameterMARmultivariate autoregressive
z j m , c class sequence of DEMAEmean absolute error
π v 0 v 1 v m 1 dispersion pattern of FSDEMAPEroot mean square error
Ziposition of individualPSOparticle swarm optimization
Pibest position of individualPEpermutation entropy
vivelocity of particleQPSOquantum behaved particle swarm optimization
siposition of particleRMSEmean absolute percentage error
Pibestindividual extreme valueSCAsine cosine algorithm
Pigbestglobal extreme valueSVMsupport vector machine
c1, c2learning factorSSKMFAsemi-supervised kernel Marginal Fisher analysis
Zijlposition of bottom layer individualTSMFEtime shift multiscale fuzzy entropy
silposition of top layer particleVMDvariational mode decomposition
Mnumber of top layer particle
Nnumber of bottom layer individual
Gmutation amplitude
lcurrent number of iterations

References

  1. Brkovic, A.; Gajic, D.; Gligorijevic, J.; Savicgajic, I.; Georgieva, O.; Gennaro, S.D. Early fault detection and diagnosis in bearings for more efficient operation of rotating machinery. Energy 2017, 136, 63–71. [Google Scholar] [CrossRef]
  2. Zhao, X.; Qin, Y.; He, C.; Jia, L.; Kou, L. Rolling element bearing fault diagnosis under impulsive noise environment based on cyclic correntropy spectrum. Entropy 2019, 21, 50. [Google Scholar] [CrossRef]
  3. Mbo’O, C.P.; Hameyer, K. Fault diagnosis of bearing damage by means of the linear discriminant analysis of stator current features from the frequency selection. IEEE Trans. Ind. Appl. 2016, 52, 3861–3868. [Google Scholar] [CrossRef]
  4. Fu, W.; Tan, J.; Li, C.; Zou, Z.; Li, Q.; Chen, T. A hybrid fault diagnosis approach for rotating machinery with the fusion of entropy-based feature extraction and SVM optimized by a chaos quantum sine cosine algorithm. Entropy 2018, 20, 626. [Google Scholar] [CrossRef]
  5. Yuan, R.; Lv, Y.; Song, G. Fault diagnosis of rolling bearing based on a novel adaptive high-order local projection denoising method. Complexity 2018, 2018, 3049318. [Google Scholar] [CrossRef]
  6. Gligorijevic, J.; Gajic, D.; Brkovic, A.; Savicgajic, I.; Georgieva, O.; Gennaro, S.D. Online condition monitoring of bearings to support total productive maintenance in the packaging materials industry. Sensors 2016, 16, 316. [Google Scholar] [CrossRef]
  7. Ali, J.B.; Fnaiech, N.; Saidi, L.; Chebel-Morello, B.; Fnaiech, F. Application of empirical mode decomposition and artificial neural network for automatic bearing fault diagnosis based on vibration signals. Appl. Acoust. 2015, 89, 16–27. [Google Scholar]
  8. Meng, L.; Li, C.; Zhang, X.; Li, R.; An, X. Compound feature selection and parameter optimization of ELM for fault diagnosis of rolling element bearings. ISA Trans. 2016, 65, 556–566. [Google Scholar]
  9. Liu, H.; Jing, J.; Jian, M. Fault diagnosis of electromechanical actuator based on VMD multifractal detrended fluctuation analysis and PNN. Complexity 2018, 2018, 9154682. [Google Scholar] [CrossRef]
  10. Fu, W.; Wang, K.; Zhou, J.; Xu, Y.; Tan, J.; Chen, T. A hybrid approach for multi-step wind speed forecasting based on multi-scale dominant ingredient chaotic analysis, KELM and synchronous optimization strategy. Sustainability 2019, 11, 1804. [Google Scholar] [CrossRef]
  11. Li, K.; Su, L.; Wu, J.; Wang, H.; Chen, P. A rolling bearing fault diagnosis method based on variational mode decomposition and an improved kernel extreme learning machine. Appl. Sci. 2017, 7, 1004. [Google Scholar] [CrossRef]
  12. Fu, W.; Tan, J.; Zhang, X.; Chen, T.; Wang, K. Blind parameter identification of MAR model and mutation hybrid GWO-SCA optimized SVM for fault diagnosis of rotating machinery. Complexity 2019, 2019, 3264969. [Google Scholar]
  13. Pincus, S. Approximate entropy (ApEn) as a complexity measure. Chaos 1995, 5, 110–117. [Google Scholar] [CrossRef] [PubMed]
  14. Bandt, C.; Pompe, B. Permutation entropy: A natural complexity measure for time series. Phys. Rev. Lett. 2002, 88, 174102. [Google Scholar] [CrossRef] [PubMed]
  15. Chen, W.; Wang, Z.; Xie, H.; Yu, W. Characterization of surface EMG signal nased on fuzzy entropy. IEEE Trans. Neural Syst. Rehabil. Eng. 2007, 15, 266–272. [Google Scholar] [CrossRef] [PubMed]
  16. Richman, J.S.; Moorman, J.R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiol.-Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef] [PubMed]
  17. Rostaghi, M.; Azami, H. Dispersion entropy: A measure for time-series analysis. IEEE Signal Process. Lett. 2016, 23, 610–614. [Google Scholar] [CrossRef]
  18. Cai, B.; Huang, L.; Xie, M. Bayesian networks in fault diagnosis. IEEE Trans. Ind. Inform. 2017, 13, 2227–2240. [Google Scholar] [CrossRef]
  19. Wang, D. K-nearest neighbors based methods for identification of different gear crack levels under different motor speeds and loads: Revisited. Mech. Syst. Signal Process. 2016, 70–71, 201–208. [Google Scholar] [CrossRef]
  20. Raj, N.; Jagadanand, G.; George, S. Fault detection and diagnosis in asymmetric multilevel inverter using artificial neural network. Int. J. Electron. 2017, 105, 559–571. [Google Scholar] [CrossRef]
  21. Shi, P.; Liang, K.; Han, D.; Zhang, Y. A novel intelligent fault diagnosis method of rotating machinery based on deep learning and PSO-SVM. J. Vibroeng. 2017, 19, 5932–5946. [Google Scholar] [CrossRef]
  22. Zhu, X.; Zheng, J.; Pan, H.; Bao, J.; Zhang, Y. Time-shift multiscale fuzzy entropy and laplacian support vector machine based rolling bearing fault diagnosis. Entropy 2018, 20, 602. [Google Scholar] [CrossRef]
  23. Jiang, M.; Jiang, L.; Jiang, D.; Li, F.; Song, H. A sensor dynamic measurement error prediction model based on NAPSO-SVM. Sensors 2018, 18, 233. [Google Scholar] [CrossRef]
  24. Xu, Y.; Zheng, Y.; Du, Y.; Yang, W.; Peng, X.; Li, C. Adaptive condition predictive-fuzzy PID optimal control of start-up process for pumped storage unit at low head area. Energy Convers. Manag. 2018, 177, 592–604. [Google Scholar] [CrossRef]
  25. Xu, Y.; Li, C.; Wang, Z.; Zhang, N.; Peng, B. Load frequency control of a novel renewable energy integrated micro-grid containing pumped hydropower energy storage. IEEE Access 2018, 6, 29067–29077. [Google Scholar] [CrossRef]
  26. Mirjalili, S. SCA: A sine cosine algorithm for solving optimization problems. Knowl-Based Syst. 2016, 96, 120–133. [Google Scholar] [CrossRef]
  27. Fu, W.; Wang, K.; Li, C.; Li, X.; Li, Y.; Zhong, H. Vibration trend measurement for a hydropower generator based on optimal variational mode decomposition and an LSSVM improved with chaotic sine cosine algorithm optimization. Meas. Sci. Technol. 2019, 30, 015012. [Google Scholar] [CrossRef]
  28. Fu, W.; Wang, K.; Li, C.; Tan, J. Multi-step short-term wind speed forecasting approach based on multi-scale dominant ingredient chaotic analysis, improved hybrid GWO-SCA optimization and ELM. Energy Convers. Manag. 2019, 187, 356–377. [Google Scholar] [CrossRef]
  29. Dragomiretskiy, K.; Zosso, D. Variational mode decomposition. IEEE Trans. Signal Process. 2014, 62, 531–544. [Google Scholar] [CrossRef]
  30. Vapnik, V.N. The Nature of Statistical Learning Theory; Springer: New York, NY, USA, 1995. [Google Scholar]
  31. Fadlallah, B.H.; Chen, B.; Keil, A.; Principe, J.C. Weighted-permutation entropy: A complexity measure for time series incorporating amplitude information. Phys. Rev. E 2013, 87, 022911. [Google Scholar] [CrossRef]
  32. Xiaofeng, L.; Yue, W. Fine-grained permutation entropy as a measure of natural complexity for time series. Chin. Phys. B 2009, 18, 2690–2695. [Google Scholar] [CrossRef] [Green Version]
  33. Kennedy, J.; Eberhart, R.C. Particle swarm optimization. In Proceedings of the International Symposium on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; pp. 1942–1948. [Google Scholar]
  34. Issa, M.; Hassanien, A.E.; Oliva, D.; Helmi, A.; Ziedan, I.; Alzohairy, A.M. ASCA-PSO: Adaptive sine cosine optimization algorithm integrated with particle swarm for pairwise local sequence alignment. Expert Syst. Appl. 2018, 99, 56–70. [Google Scholar] [CrossRef]
  35. Pehlivanoglu, Y.V. A new particle swarm optimization method enhanced with a periodic mutation strategy and neural networks. IEEE Trans. Evol. Comput. 2013, 17, 436–452. [Google Scholar] [CrossRef]
  36. Bearing Data Center of the Case Western Reserve University. Available online: http://csegroups.case.edu/bearingdatacenter/pages/download-data-file (accessed on 28 December 2018).
  37. Zhao, H.; Sun, M.; Deng, W.; Yang, X. A new feature extraction method based on EEMD and multi-scale fuzzy entropy for motor bearing. Entropy 2016, 19, 14. [Google Scholar] [CrossRef]
  38. Rostaghi, M.; Ashory, M.R.; Azami, H. Application of dispersion entropy to status characterization of rotary machines. J. Sound Vib. 2019, 438, 291–308. [Google Scholar] [CrossRef]
  39. Nguyen, T.P.Q.; Kuo, R.J. Partition-and-merge based fuzzy genetic clustering algorithm for categorical data. Appl. Soft Comput. 2019, 75, 254–264. [Google Scholar] [CrossRef]
  40. Fahad, A.; Alshatri, N.; Tari, Z.; Alamri, A.; Khalil, I.; Zomaya, A.Y.; Foufou, S.; Bouras, A. A survey of clustering algorithms for big data: Taxonomy and empirical analysis. IEEE Trans. Emerg. Top. Comput. 2014, 2, 267–279. [Google Scholar] [CrossRef]
  41. Zhang, Q.; Gao, J.; Dong, H.; Mao, Y. WPD and DE/BBO-RBFNN for solution of rolling bearing fault diagnosis. Neurocomputing 2018, 312, 27–33. [Google Scholar] [CrossRef]
  42. Yang, Z.; Wang, J. A combination forecasting approach applied in multistep wind speed forecasting based on a data processing strategy and an optimized artificial intelligence algorithm. Appl. Energy 2018, 230, 1108–1125. [Google Scholar] [CrossRef]
  43. Azami, H.; Escudero, J. Amplitude-aware permutation entropy: Illustration in spike detection and signal segmentation. Comput. Methods Programs Biomed. 2016, 128, 40–51. [Google Scholar] [CrossRef] [Green Version]
  44. Lai, X.; Li, C.; Zhou, J.; Zhang, N. Multi-objective optimization for guide vane shutting based on MOASA. Renewable Energy 2019, 139, 302–312. [Google Scholar] [CrossRef]
  45. Li, C.; Wang, W.; Chen, D. Multi-objective complementary scheduling of hydro-thermal-RE power system via a multi-objective hybrid grey wolf optimizer. Energy 2019, 171, 241–255. [Google Scholar] [CrossRef]
  46. Liu, Y.; Qin, H.; Mo, L.; Wang, Y.; Chen, D.; Pang, S.; Yin, X. Hierarchical flood operation rules optimization using multi-objective cultured evolutionary algorithm based on decomposition. Water Resour. Manag. 2019, 33, 337–354. [Google Scholar] [CrossRef]
  47. Zhang, C.; Peng, T.; Li, C.; Fu, W.; Xia, X.; Xue, X. Multiobjective optimization of a fractional-order PID controller for pumped turbine governing system using an improved NSGA-III algorithm under multiworking conditions. Complexity 2019, 2019, 5826873. [Google Scholar] [CrossRef]
  48. Wang, W.; Tahir, H.; Ren, Z.; Zhou, K. Robust fault detection and diagnosis for uncertain nonlinear systems. J. Syst. Eng. Electron. 2009, 20, 1031–1036. [Google Scholar]
  49. Yan, X.; Jia, M. Intelligent fault diagnosis of rotating machinery using improved multiscale dispersion entropy and mRMR feature selection. Knowl. Based Syst. 2019, 163, 450–471. [Google Scholar] [CrossRef]
  50. Zheng, J.; Pan, H.; Yang, S.; Cheng, J. Generalized composite multiscale permutation entropy and laplacian score based rolling bearing fault diagnosis. Mech. Syst. Signal Process. 2018, 99, 229–243. [Google Scholar] [CrossRef]
  51. Jiang, L.; Xuan, J.; Shi, T. Feature extraction based on semi-supervised kernel marginal fisher analysis and its application in bearing fault diagnosis. Mech. Syst. Signal Process. 2013, 41, 113–126. [Google Scholar] [CrossRef]
Figure 1. Example of possible vectors corresponding to the same dispersion pattern.
Figure 1. Example of possible vectors corresponding to the same dispersion pattern.
Entropy 21 00404 g001
Figure 2. f is added in class vector of dispersion entropy as an additional element.
Figure 2. f is added in class vector of dispersion entropy as an additional element.
Entropy 21 00404 g002
Figure 3. The structure of the mutation sine cosine algorithm and particle swarm optimization (SCA-PSO).
Figure 3. The structure of the mutation sine cosine algorithm and particle swarm optimization (SCA-PSO).
Entropy 21 00404 g003
Figure 4. Block diagram of the proposed mutation operator, sine cosine algorithm and particle swarm optimization (MSCAPSO).
Figure 4. Block diagram of the proposed mutation operator, sine cosine algorithm and particle swarm optimization (MSCAPSO).
Entropy 21 00404 g004
Figure 5. Flowchart of the proposed fault diagnosis method.
Figure 5. Flowchart of the proposed fault diagnosis method.
Entropy 21 00404 g005
Figure 6. The variation of central frequency with iteration under different K values.
Figure 6. The variation of central frequency with iteration under different K values.
Entropy 21 00404 g006
Figure 7. Time domain waveforms of fault signals: (a) signals with different fault positions (L2, L5, L8), (b) signals with different defect sizes (L4, L5, L6).
Figure 7. Time domain waveforms of fault signals: (a) signals with different fault positions (L2, L5, L8), (b) signals with different defect sizes (L4, L5, L6).
Entropy 21 00404 g007
Figure 8. The variational mode decomposition (VMD) results of fault signals: (a) decomposition results with different fault positions (L2, L5, L8), (b) decomposition results with different defect sizes (L4, L5, L6).
Figure 8. The variational mode decomposition (VMD) results of fault signals: (a) decomposition results with different fault positions (L2, L5, L8), (b) decomposition results with different defect sizes (L4, L5, L6).
Entropy 21 00404 g008
Figure 9. 3D projection of entropy vectors for different types of faults: (a) vectors of permutation entropy, (b) fuzzy entropy, (c) dispersion entropy(d), fine-sorted dispersion entropy.
Figure 9. 3D projection of entropy vectors for different types of faults: (a) vectors of permutation entropy, (b) fuzzy entropy, (c) dispersion entropy(d), fine-sorted dispersion entropy.
Entropy 21 00404 g009aEntropy 21 00404 g009b
Figure 10. Convergence procedure of proposed optimization approach: (a) distribution of convergence curves in 10 runs (b) comparison of different optimization methods.
Figure 10. Convergence procedure of proposed optimization approach: (a) distribution of convergence curves in 10 runs (b) comparison of different optimization methods.
Entropy 21 00404 g010
Figure 11. Comparison of evaluation values for different methods (2 hp).
Figure 11. Comparison of evaluation values for different methods (2 hp).
Entropy 21 00404 g011
Figure 12. Boxplots of evaluation values for different methods under variable load conditions: (a) 2hp, (b) 2hp, (c) 0hp; the x-axis tick labels correspond to: 1: Fuzzy entropy-particle swarm optimization (FE-PSO), 2: Permutation entropy-sine cosine algorithm (PE-SCA), 3: Dispersion entropy-particle swarm optimization (DE-PSO), 4 Dispersion entropy-sine cosine algorithm (DE-SCA), 5: Fine-sorted dispersion entropy-particle swarm optimization (FSDE-PSO), 6: Fine-sorted dispersion entropy-sine cosine algorithm (FSDE-SCA), 7: Fine-sorted dispersion entropy-mutation operator, sine cosine algorithm and particle swarm optimization (FSDE-MSCAPSO).
Figure 12. Boxplots of evaluation values for different methods under variable load conditions: (a) 2hp, (b) 2hp, (c) 0hp; the x-axis tick labels correspond to: 1: Fuzzy entropy-particle swarm optimization (FE-PSO), 2: Permutation entropy-sine cosine algorithm (PE-SCA), 3: Dispersion entropy-particle swarm optimization (DE-PSO), 4 Dispersion entropy-sine cosine algorithm (DE-SCA), 5: Fine-sorted dispersion entropy-particle swarm optimization (FSDE-PSO), 6: Fine-sorted dispersion entropy-sine cosine algorithm (FSDE-SCA), 7: Fine-sorted dispersion entropy-mutation operator, sine cosine algorithm and particle swarm optimization (FSDE-MSCAPSO).
Entropy 21 00404 g012
Figure 13. Comparison of mean absolute error (MAE), root mean square error (RMSE), and mean absolute percentage error (MAPE) for diagnosis models under different load conditions: (a) MAE, (b) RMSE, (c) MAPE.
Figure 13. Comparison of mean absolute error (MAE), root mean square error (RMSE), and mean absolute percentage error (MAPE) for diagnosis models under different load conditions: (a) MAE, (b) RMSE, (c) MAPE.
Entropy 21 00404 g013
Table 1. Description of the experimental data.
Table 1. Description of the experimental data.
Motor Load (hp)Position of FaultDefect Size (Inches)Label of ClassesNumber of Samples
2/1/0Inner race0.007L159
Inner race0.014L259
Inner race0.021L359
Ball0.007L459
Ball0.014L559
Ball0.021L659
Outer race0.007L759
Outer race0.014L859
Outer race0.021L959
Table 2. Normalized central frequencies with different K values.
Table 2. Normalized central frequencies with different K values.
Number of ModesNormalized Central Frequencies
20.26230.0438
30.26980.22270.0418
40.27000.22460.11120.0367
50.27600.26140.22090.11070.0366
60.29920.27450.26060.22000.11060.0366
70.33010.27540.26100.22060.11460.05480.0315
80.36630.28470.27110.25480.21800.11440.05460.0314
Table 3. Fine-sorted dispersion entropy and dispersion entropy of intrinsic mode functions (IMFs).
Table 3. Fine-sorted dispersion entropy and dispersion entropy of intrinsic mode functions (IMFs).
Fault LabelSample NumberFine-Sorted Dispersion EntropyDispersion Entropy
IMF1IMF2IMF3IMF4IMF1IMF2IMF3IMF4
L110.44280.41260.38600.31540.41100.37400.32020.2509
20.44210.41430.38540.31540.40740.37340.32040.2520
30.44020.41160.38620.31510.40300.37000.32010.2495
L210.37990.36710.38280.30030.34860.33440.31570.2370
20.38060.37130.38490.29590.36060.35200.31700.2380
30.38220.36750.38640.30690.36110.33470.31630.2452
L310.40740.54480.42770.30610.37300.52250.38810.2514
20.42310.51330.42220.30500.39070.49240.38290.2519
30.39860.51130.42460.30640.35760.48590.37730.2509
L410.39930.38380.36110.26230.31460.30560.29700.2031
20.40340.38760.35550.24520.31200.30410.29140.1789
30.40280.38740.36220.28910.31350.31080.29680.2288
L510.37080.37070.38840.27690.30630.31120.31760.2218
20.36380.37250.38610.27990.30430.31290.31430.2198
30.37310.38560.38730.27830.30210.32030.31620.2206
L610.35130.39160.39380.30260.28970.31570.31900.2397
20.38470.39430.39440.30310.31090.31530.31780.2415
30.38510.38820.40210.31270.31170.31350.31870.2499
L710.45610.58150.49640.35460.43720.55820.47180.2926
20.46130.58200.49100.34440.44140.55690.46850.2859
30.46060.58640.49600.37540.43950.56320.47360.3167
L810.40520.38920.34770.27390.32820.30830.28250.2159
20.41620.39530.33240.27700.33290.31590.26690.2165
30.40240.38840.35390.29790.32960.31400.28920.2320
L910.47930.41150.36430.30880.47460.40410.32440.2656
20.48990.38630.39960.31690.48550.37580.37430.2778
30.49060.50400.42910.31390.48250.49660.42340.2751
Table 4. Fault diagnosis results with different methods under variable load conditions.
Table 4. Fault diagnosis results with different methods under variable load conditions.
Motor Load (hp)MethodsCgEvaluation Metrics
Adjusted Rand Index (ARI)Normalized Mutual Information (NMI)F-Measure (F)Accuracy (ACC)
2FE-PSO204.7667.9780.9309
[−0.025, 0.043]
0.9288
[−0.021, 0.038]
0.9687
[−0.010, 0.020]
0.9686
[−0.011, 0.020]
PE-SCA616.339411.1110.9185
[−0.043, 0.046]
0.9207
[−0.032, 0.048]
0.9628
[−0.020, 0.022]
0.9628
[−0.020, 0.022]
DE-PSO13.12410240.9461
[−0.023, 0.028]
0.9463
[−0.021, 0.033]
0.9759
[−0.010, 0.013]
0.9759
[−0.010, 0.013]
DE-SCA12.762420.7620.9456
[−0.022, 0.028]
0.9513
[−0.020, 0.019]
0.9755
[−0.010, 0.013]
0.9755
[−0.010, 0.013]
FSDE-PSO54.370151.0240.9572
[−0.016, 0.009]
0.9603
[−0.019, 0.011]
0.9808
[−0.008, 0.004]
0.9808
[−0.008, 0.004]
FSDE-SCA25.206200.2670.9549
[−0.014, 0.019]
0.9589
[−0.021, 0.019]
0.9797
[−0.007, 0.009]
0.9797
[−0.007, 0.009]
FSDE-MSCAPSO6.085390.6270.9637
[−0.019, 0.019]
0.9646
[−0.016, 0.014]
0.9839
[−0.007, 0.008]
0.9839
[−0.007, 0.008]
1FE-PSO6.9082.3580.9304
[−0.053, 0.035]
0.9243
[−0.050, 0.035]
0.9681
[−0.027, 0.017]
0.9682
[−0.026, 0.016]
PE-SCA69.617294.3580.9019
[−0.023, 0.032]
0.8942
[−0.0433, 0.044]
0.9537
[−0.012, 0.015]
0.9544
[−0.012, 0.015]
DE-PSO2.217586.4660.8814
[−0.040, 0.051]
0.8827
[−0.046, 0.041]
0.9429
[−0.021, 0.027]
0.9433
[−0.020, 0.026]
DE-SCA331.824250.2050.8864
[−0.042, 0.047]
0.8869
[−0.035, 0.034]
0.9458
[−0.023, 0.024]
0.9460
[−0.023, 0.023]
FSDE-PSO2.7541024.0000.9483
[−0.023, 0.018]
0.9468
[−0.024, 0.024]
0.9765
[−0.011, 0.008]
0.9766
[−0.011, 0.008]
FSDE-SCA4.532690.2480.9386
[−0.029, 0.044]
0.9387
[−0.025, 0.040]
0.9718
[−0.015, 0.021]
0.9720
[−0.014, 0.020]
FSDE-MSCAPSO5.0041024.0000.9615
[−0.038, 0.021]
0.9598
[−0.025, 0.019]
0.9827
[−0.018, 0.010]
0.9828
[−0.017, 0.010]
0FE-PSO9.4539.7810.9141
[−0.034, 0.052]
0.9144
[−0.035, 0.051]
0.9596
[−0.018, 0.025]
0.9598
[−0.017, 0.025]
PE-SCA853.00232.9320.8850
[−0.045, 0.031]
0.8862
[−0.048, 0.038]
0.9447
[−0.021, 0.018]
0.9448
[−0.021, 0.017]
DE-PSO16.319456.1700.8942
[−0.055, 0.038]
0.8997
[−0.054, 0.032]
0.9512
[−0.027, 0.018]
0.9510
[−0.028, 0.018]
DE-SCA155.860163.4980.8753
[−0.037, 0.035]
0.8855
[−0.025, 0.028]
0.9396
[−0.020, 0.018]
0.9398
[−0.020, 0.018]
FSDE-PSO16.138549.6630.9036
[−0.020, 0.045]
0.9111
[−0.034, 0.046]
0.9554
[−0.010, 0.022]
0.9556
[−0.009, 0.021]
FSDE-SCA1024.0002.4000.8873
[−0.034, 0.022]
0.8948
[−0.042, 0.024]
0.9473
[−0.016, 0.011]
0.9475
[−0.016, 0.010]
FSDE-MSCAPSO216.051448.1830.9079
[−0.031, 0.033]
0.9086
[−0.035, 0.033]
0.9567
[−0.015, 0.017]
0.9571
[−0.015, 0.016]
Table 5. Error metrics of results of different methods.
Table 5. Error metrics of results of different methods.
MethodMAERMSE MAPE
2 hp1 hp0 hp2 hp1 hp0 hp2 hp1 hp0 hp
FE-PSO0.03140.03180.04020.03240.03540.04200.03250.03310.0421
PE-SCA0.03720.04560.05520.03900.04620.05660.03880.04780.0586
DE-PSO0.02410.05670.04900.02530.05920.05160.02480.06050.0519
DE-SCA0.02450.05400.06020.02560.05580.06170.02520.05730.0642
FSDE-PSO0.01920.02340.04440.01950.02400.04530.01950.02400.0466
FSDE-SCA0.02030.02800.05250.02100.03000.05310.02080.02890.0555
FSDE-MSCAPSO0.01610.01720.04290.01690.01900.04430.01640.01760.0450
Table 6. Comparison among several other fault diagnosis model. The abbreviations are as follows: Chaos quantum sine cosine algorithm (CQSCA), hybrid gravitational search algorithm (HGSA), extreme learning machine (ELM), multivariate autoregressive (MAR), time shift multiscale fuzzy entropy (TSMFE), Laplacian support vector machine (LapSVM), improved multiscale dispersion entropy (IMDE), max-relevance min-redundancy (mRMR), back-propagation neural network (BPNN), improved multiscale permutation entropy (IMPE), quantum behaved particle swarm optimization (QPSO), least squares support vector machine (LSSVM), generalized composite multiscale permutation entropy (GCMPE), semi-supervised kernel Marginal Fisher analysis (SSKMFA).
Table 6. Comparison among several other fault diagnosis model. The abbreviations are as follows: Chaos quantum sine cosine algorithm (CQSCA), hybrid gravitational search algorithm (HGSA), extreme learning machine (ELM), multivariate autoregressive (MAR), time shift multiscale fuzzy entropy (TSMFE), Laplacian support vector machine (LapSVM), improved multiscale dispersion entropy (IMDE), max-relevance min-redundancy (mRMR), back-propagation neural network (BPNN), improved multiscale permutation entropy (IMPE), quantum behaved particle swarm optimization (QPSO), least squares support vector machine (LSSVM), generalized composite multiscale permutation entropy (GCMPE), semi-supervised kernel Marginal Fisher analysis (SSKMFA).
ReferenceFeature ExtractionClassificationAccuracy
[4]IMFs decomposed by VMD+PECQSCA-SVM0.9789
[8]IMFs decomposed by EEMDHGSA-ELM0.9938
[12]IMFs decomposed by VMD+MARHMGWOSCA-SVM0.9808
[20]IMPEQPSO-LSSVM0.9800
[22]TSMFEC1: LapSVM
C2: SVM
C1: 100
C2: 98.44
[49]IMDEC1: mRMR+ELM
C2: mRMR+BPNN
C3: mRMR+SVM
C1: 0.9056
C2:0.8760
C3: 0.8840
[50]GCMPEPSO-SVM0.9889
[51]SSKMFAKNN0.9850

Share and Cite

MDPI and ACS Style

Fu, W.; Tan, J.; Xu, Y.; Wang, K.; Chen, T. Fault Diagnosis for Rolling Bearings Based on Fine-Sorted Dispersion Entropy and SVM Optimized with Mutation SCA-PSO. Entropy 2019, 21, 404. https://doi.org/10.3390/e21040404

AMA Style

Fu W, Tan J, Xu Y, Wang K, Chen T. Fault Diagnosis for Rolling Bearings Based on Fine-Sorted Dispersion Entropy and SVM Optimized with Mutation SCA-PSO. Entropy. 2019; 21(4):404. https://doi.org/10.3390/e21040404

Chicago/Turabian Style

Fu, Wenlong, Jiawen Tan, Yanhe Xu, Kai Wang, and Tie Chen. 2019. "Fault Diagnosis for Rolling Bearings Based on Fine-Sorted Dispersion Entropy and SVM Optimized with Mutation SCA-PSO" Entropy 21, no. 4: 404. https://doi.org/10.3390/e21040404

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop