Next Article in Journal
Dynamic Corrosion Test Using LiNO3 Containing Molten Salt for CSP Applications
Previous Article in Journal
Semi-Automatic Cloud-Native Video Annotation for Autonomous Driving
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Fault Diagnosis Algorithm for Rolling Bearings Based on One-Dimensional Convolutional Neural Network and INPSO-SVM

School of Mechanical, Electrical & Information Engineering, Shandong University, Weihai 264209, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(12), 4303; https://doi.org/10.3390/app10124303
Submission received: 25 May 2020 / Revised: 13 June 2020 / Accepted: 19 June 2020 / Published: 23 June 2020
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Deep learning based intelligent fault diagnosis methods have become a research hotspot in the fields of fault diagnosis and the health management of rolling bearings in recent years. To effectively identify incipient faults in rotating machinery, this paper proposes a novel hybrid intelligent fault diagnosis framework based on a convolutional neural network and support vector machine (SVM). First, an improved one-dimensional convolutional neural network (1DCNN) was adopted to extract fault features, and the state information and intrinsic properties of the raw vibration signals were mined. Second, the extracted features were used to train the SVM, which was applied to classify the fault category. The proposed hybrid framework combined the excellent classification performance of the SVM for small samples and the strong feature-learning ability of CNN network. In order to tune the parameters of the SVM, an improved novel particle swarm optimization algorithm (INPSO) which combined the Tent map and Lévy flight strategy was proposed. Numerical experimental results indicated that the proposed PSO variant had a better performance in searching accuracy and convergence speed. At last, multiple groups of rolling bearing fault diagnosis experiments were carried out and experimental results showed that, with the proposed 1DCNN-INPSO-SVM model, the hybrid framework was capable of diagnosing with high precision for rolling bearings and superior to some traditional fault diagnosis methods.

1. Introduction

Rotating machinery plays an important part in modern industrial manufacture, and its failure may cause unpredictable security accidents and economic losses [1]. Therefore, the identification and diagnosis of rotating machine faults is one of the major topics in the study of mechanical systems fault prevention [1,2]. Rolling element bearings are considered to be the key components in a rotating mechanism system, and faults occurring in the rolling element bearings can not only have a direct and massive influence on the stability of the system, but also shorten the service life of the machine [3]. Therefore, after the vibration signals are collected from the sensors, the most effective way for the rolling bearings to avoid possible damage is to establish a highly efficient and intelligent fault diagnosis and health monitoring system [4]. Generally, intelligent fault diagnosis methods can be divided into two steps, namely feature extraction based on signal processing and health status recognition based on the classification model [5]. The raw signals collected from multiple sensors contain useful information about the running status of the machine, as well as useless noises in the time domain. Therefore, it is necessary to find an effective feature extraction method to obtain intrinsic features of the signals. Wavelet packet transformation (WPT) [6], empirical mode decomposition (EMD) [7] and fast Fourier transformation (FFT) [8] are three kinds of widely applied feature extraction methods in the vibration signal processing field. After feature extraction, principal component analysis (PCA) [9] and independent component analysis (ICA) [10] methods usually are selected to remove the low sensitivity and useless data from the extracted features. With the feature vectors selected and extracted from raw signals, state classification is accomplished by K nearest neighbor (KNN) [11], support vector machine (SVM) [12] and other machine learning algorithms and statistical frameworks [13,14]. The application of traditional fault diagnosis algorithms has achieved certain results, but there are still some shortcomings, such as the feature selection excessively relies on the experience and prior knowledge of the engineers [15], furthermore, it is very subjective and time-consuming [16]. At the same time, traditional machine learning models are faced with dimensional disaster, overfitting and other problems, and the shallow models are difficult to represent the complex mapping relationships between signal samples and health status [17].
As Hinton et al. [18] proposed the method of layer-by-layer unsupervised training combined with supervised fine-tuning, deep learning (DL) theory has become a great breakthrough in the field of machine learning and artificial intelligence [16]. Compared with traditional intelligent fault diagnosis methods, the deep model can learn representative invariable features and complex nonlinear relationships from the raw data automatically. As a classical deep learning model, the convolutional neural network (CNN) has a strong feature learning ability with its unique convolution structure, which has achieved great success in the field of image recognition and has been widely applied in fault diagnosis. Chen et al. [19] proposed a CNN model used for intelligent fault diagnosis of gearboxes with manually extracted features. Wang et al. [20] proposed a CNN model for rotating machinery conditions recognition whose inputs are two-dimensional matrix transformed by time domain signals. Janssens et al. [21] adopted the data processed by the discrete Fourier transform method as the inputs of a three-layer CNN model for fault diagnosis. Many research results show that the CNN network has a strong ability for learning sensitivity and robustness. However, how to preprocess the raw data efficiently, e.g., the transformation process of one-dimensional vibration signals into two-dimensional matrices for CNN inputs, is a thought-provoking question. To the best of our knowledge there is no complete theory to guarantee the transformation is reasonable. In this paper, the proposed 1DCNN network avoids time-consuming data preprocessing and extracts features more accurately by carrying out automatic adaptive learning on one-dimensional raw vibration signals. In the field of image processing, the fully connected layer of a CNN network generally adopts softmax function to classify, which can achieve a good classification accuracy. However, for fault diagnosis problems, the training data volume is much smaller than that of the image processing, which means that the depth model is difficult to obtain ideal results in dealing with fault diagnosis problems independently. For SVM, the idea of kernel function is adopted to map nonlinear samples to a high-dimensional space and make it linearly separable, and then SVM finds an optimal hyperplane that maximizes the classification interval between data sets, thus promoting classification performance [22]. Many research works have proved that the SVM has a strong learning ability for small sample classification problems. Therefore, by combining 1DCNN with the SVM, the extracted sparse representative feature vectors from the second fully connected layer of the network are taken as the inputs of the SVM for achieving optimal classification accuracy. However, the standard SVM needs to preset two crucial parameters (i.e., penalty factor c and kernel parameter g), which have a great influence on its classification performance [23]. To solve this problem, various optimization methods are applied to tune SVM parameters, such as genetic algorithm (GA) [24] and particle swarm optimization (PSO) [23]. Recently, the particle swarm optimization algorithm has achieved remarkable results in many works. However, the original PSO algorithm has some drawbacks, e.g., premature and slow convergence speed. To avoid those shortcomings, this paper proposed an improved novel particle swarm optimization algorithm (INPSO) for SVM parameter optimization.
In this paper, we propose a novel fault diagnosis method based on the one-dimensional convolution neural network (1DCNN) and SVM optimized by INPSO (INPSO-SVM) to improve diagnosis accuracy. To sum up, there are three main contributions existing in the proposed method. Firstly, the proposed method directly extracts the features of the original time-domain vibration signals by applying a 1DCNN network, without any time-frequency conversion. Secondly, the excellent small sample learning abilities of the SVM is combined with the powerful deep feature adaptive learning ability of 1DCNN to improve the diagnosis accuracy. Finally, an improved particle swarm optimizer variant is presented to tune the parameters of SVM, which can improve the classification accuracy and generalization performance of SVM. The engineering application and contrastive tests reveal the effectiveness and superiority of the proposed method.
The remainder of this paper is organized as follows: Section 2 is dedicated to the basic knowledge of CNN, SVM, and PSO. Section 3 introduces the proposed fault diagnosis method based on 1DCNN model and INPSO optimized SVM. Section 4 discusses the results of the experiments. The conclusion is summarized in Section 5.

2. Fundamental Theories

2.1. Convolutional Neural Network

CNN is a typical feed-forward neural network model which is composed by a trainable feature extraction part and a classification part [16,25]. The feature extraction part mainly includes convolutional layers and pooling layers [21], and the classification part is a multilayer perceptron consisting of several fully connected layers [26].
In the convolutional layers, the input samples are convolved with a bunch of learnable convolution kernels to extract feature vectors [16]. Usually a convolution layer has multiple convolution kernels and a class of features learned from a convolution kernel are called mapping graphs. The operation can be expressed as follows.
x i k ( j ) = w i k x i 1 ( j ) + b i k
where donates the convolution operator, i donates the layer number of the network, k is the index of the output feature maps, x i 1 ( j ) is the input feature map of j-th local region, and x i k ( j ) is the k-th output feature map after the convolution calculation. w i k and b i k are the weight and the bias of the k-th group filter in the i-th layer.
In recent years, the rectified linear unit (ReLU) has been widely used as an activation element to accelerate the convergence of neural networks [26]. ReLU model can be described by Equation (2).
A i k ( j ) = f ( x i k ( j ) ) = max { 0 , x i k ( j ) }
where x i k ( j ) is the output value of the convolution operation, A i k ( j ) is the activation of x i k ( j ) .
A pooling operation is adopted to reduce the computational complexity, and effectively control the risk of overfitting [27]. Max-pooling is defined as follows.
P i k ( j ) = max ( j 1 ) S + 1 < t j S { A i k ( t ) }
where A i k ( t ) denotes the value of t-th neuron in the k-th output feature map of the i-th layer. S is the pooling length, and P i k is the output feature map after the pooling operation.
In a traditional convolutional neural network, after several convolutional layers and pooling layers, the feature maps learned by each kernel function are flattened into a one-dimensional array as the input of the fully connected layer. The output of the fully connected layer is a one-dimensional vector, and the mathematical model of the fully connected layer is described as follows.
f ( x ) = σ ( w * x + b )
where σ ( ) is the activation function, x is the input of a fully connected layer, f ( x ) denotes the output of a fully connected layer, and w, b are the bias and the weight.

2.2. SVM Classifier Parameters Tuning

2.2.1. Support Vector Machine

A SVM is a widely used machine learning method based on the minimization principle of structural risk [23]. It has many unique advantages in solving small sample, multi-class, non-linear and high-dimensional pattern recognition problems [5]. In this paper, the proposed method mainly uses a nonlinear multi-class SVM as the backend multi-class classifier to solve the classification problem [28]. Given a group of sample set S = { ( x i , y i ) i = 1 n | x i R N , y i { 1 , 1 } } , where x i presents the N dimensional input feature of the i-sample, and y i is the corresponding label for the input sample. If the hyperplane equation reaches the criterion of the optimal separating plane, the optimal separation plane can be solved by the following objective functions and constraints.
{ min   1 2 w 2 + c i = 1 n ξ i ( ξ i 0 ,   c 0 ) s . t . y i ( w x i + b ) 1 ξ i ( i = 1 , 2 , , n )
where w denotes the weight vector and b depicts the bias vector, ξ i is the relaxing factor and c is a constant named as the penalty factor and it can relieve contradiction between algorithm complexity and classification accuracy. Finally, the classification problem of SVM are transformed to the following dual optimization problem.
{ max L = i = 1 n μ i 1 2 i , j = 1 n μ i μ j y i y j K ( x i , x j ) s . t . i = 1 n μ i y i = 0 ( i = 1 , 2 , , n , μ i 0 )
where μ i are the Lagrangian multipliers, and K ( x i , x j ) represents a kernel function that meets K ( x i , x j ) = < ϕ T ( x i ) ϕ ( x j ) > , where the < > is the inner product operation. The final SVM classification function is:
f ( x ) = sgn ( i = 1 n a i y i < K ( x i , x j ) > + b )

2.2.2. Particle Swarm Optimization Algorithm

In the original PSO algorithm, there are N particles in the population, and each particle i has a position vector X i = ( x i 1 , x i 2 , , x i D ) and a velocity vector V i = ( v i 1 , v i 2 , , v i D ) to calculate its current state, where D is the dimensions of the solution space. First, the initial positions and velocities of the particles are randomly initialized in the search area, where the position of each particle represents a solution. The fitness value f i can be calculated by substituting the vector X i into the target function. Then by comparing the fitness value of each particle, the position with the best fitness value is the optimal position of the current population. At last, in the t + 1 iteration of the particle, the positions and velocities of the particles are updated according to Equations (8) and (9).
V g + 1 , i = ω V g , i + c 1 r 1 ( P g , i X g , i ) + c 2 r 2 ( P g X g , i )
X g + 1 , i = X g , i + V g + 1 , i
where w is inertia weight to control the impact of the previous velocity. c 1 ,   c 2 are the learning factors, and r 1 ,   r 2 are random numbers in [0, 1] for denoting remembrance ability. V g , i , X g , i represent the g-th generation velocity and position of the i-th particle respectively, P g , i is the optimal position that the i-th particle has searched so far, and P g is the historical optimal value searched by the whole population.

3. Fault Diagnosis Method Based on 1DCNN and INPSO-SVM

3.1. Overall Framework of the Proposed Method

To improve the fault diagnosis accuracy of rolling bearings, a novel hybrid fault diagnosis scheme which combines the 1DCNN and INPSO-SVM is proposed in this paper. Figure 1 shows the flowchart of the proposed method, which can be divided into three major steps (i.e., data collection, feature extraction and fault classification). The raw fault data from multiple sensors is directly input into the improved 1DCNN model for the training of a CNN model. The outputs of the fully connected layer are taken as the feature vectors which are input into the SVM classifier for fault classification. Since the classification performance of an SVM classifier is susceptible to the values of parameters c and g, the proposed INPSO is adopted to tune the parameters of the SVM.

3.2. Feature Extraction Based on 1DCNN

The training process of the convolutional neural network can be divided into a forward propagation stage and a backward propagation stage. In the process of forward propagation, the network firstly initializes the weights, and then convolution kernels filter each small region of the data to obtain the eigenvalues of the small region. The features obtained from the convolution layer are passed into the lower sampling layer, where the dimension and quantity of trainable parameters are reduced to avoid overfitting by pooling. Finally, the features extracted from the lower sampling layer are transferred to the fully connected layer and the softmax layer for classification, and the classification model is obtained to get the final fault diagnosis results. When the output results of the convolutional neural network are inconsistent with the real labels, the errors are propagated from the high level to the low level in the backward propagation stage. According to the ratio of each layer of errors to the total errors, the weights of each layer are updated. In the training stage of CNN, dropout technology, data sliding window expansion technology and batch normalization technology are used to prevent over-fitting and speed up the training process. The structure of the proposed 1DCNN network is shown in Figure 2.
The overall structure of proposed 1DCNN network is composed of four convolution layers, four maximum pooling layers and two fully connected layers. Details about the architecture of 1DCNN can be found in Table 1. The input signals are the normalized time-domain signals. Features can be extracted directly from the raw time domain signals by using one-dimensional convolution kernel and pooling kernel, which reduces the work of signal preprocessing. Compared with traditional CNN structure, the main advantage is that the convolution kernel of the first layer is larger in size to better suppress the high-frequency noise, and the following convolutional kernels are small with 3 × 1, which is helpful to obtain a good representation of the input signals and improve the network performance.

3.3. Fault Identification Based on INPSO-SVM Algorithm

A support vector machine (SVM) is an efficient machine learning tool developed by Vapnik [29], but the penalty factor c and kernel parameter g of a SVM classifier mainly rely on manual selection, which has a great impact on the recognition accuracy of the SVM classifier. In general, the penalty factor c represents the nonlinear fitting ability. If the parameter c is extremely large, SVM will lose the generalization ability [23]. The size of the kernel parameter g affects the kernel mapping distribution of the data samples. Therefore, parameter tuning is essential for the SVM to get a good classification performance. In view of this, an improved new PSO variant is proposed to optimize the SVM parameters in this study.

3.3.1. An Improved New PSO Algorithm (INPSO)

In order to overcome the problem of trapping into local minimums and facilitate the convergence speed of the basic PSO algorithm, a novel population initialization method based on Tent map and an enhanced particle position update approach based on Lévy flights was developed. The modified method can reduce the population degradation and enrich the population diversity. Moreover, the premature convergence is avoided by searching the neighborhood of the particles that have fallen into local optimums.

Chaos Based Initialization

In the standard PSO algorithm, the initial positions of the particles are initialized randomly, which affects the search efficiency of the algorithm and limits the optimization ability of the PSO algorithm. Since the 1990s, with the study of chaos theory, chaotic maps have been widely applied in various fields. Chaos is a typical nonlinear phenomenon in nature, a chaotic system can go through every state in a certain range without repeating according to its own rules. Adding chaos operator into the optimization algorithm can improve the global search ability of the algorithm [30]. Among the chaotic systems, the structure of Tent map is simple, and its ergodic performance produces uniform distribution chaotic sequences, and the iteration process is suitable for computer operation [31]. The mathematical model of Tent chaos map is defined as follows.
x i + 1 = { 2 x i + r a n d ( 0 , 1 ) * 1 N , 0 x 1 2 2 ( 1 x i ) + r a n d ( 0 , 1 ) * 1 N , 1 2 x 1  
where, N is the population size, and rand (0, 1) is a random number within the range of (0, 1). The introduction of random variable not only ensures the randomness, ergodicity and regularity of the Tent chaotic map, but also effectively avoids iterations falling into the points of small period and unstable period.

An Enhanced Particle Position Updating Strategy

Lévy flight, proposed by Paul Lévy, is a non-gaussian random process [32]. The short distance walk of Lévy flight mechanism improves the diversity of the population, and the directional variability of the occasional long jump ensures that the population searches the surrounding area in detail. The combination of a short distance walk and a long distance jump can fully search the solution domain and greatly improve the global optimization ability of the algorithm.
In the proposed INPSO algorithm, this paper eliminates the influence of the velocity term, and the particle swarm position is updated by the following equation.
X n e w , i = ω X g , i + c 1 r 1 ( P g , i X g , i ) + c 2 r 2 ( P g X g , i ) + c 3 r 3 ( P r X g , i )
where, P r is a random particle in the current set of particles, c 3 represents the learning factor of random particle, r 3 is drawn from the uniform distribution range in [0, 1]. The fourth term is a random learning mechanism, which can increase the diversity of the population and overcome premature convergence.
In order to make a good balance between local and global search capability, the particle swarm optimization algorithm is combined with an improved Lévy flight strategy to obtain a new generation of particle swarm. The updating equation of group position is defined by Equation (12).
X n e w , i = ω X g , i + c 1 r 1 ( P g , i X g , i ) + c 2 r 2 ( P g X g , i ) + c 3 r 3 ( P r X g , i ) * α L e v y ( λ )
where, α is random step size, represents the dot product, L e v y ( λ ) is a random search path that conforms to Lévy distribution and meets the following constraints [33].
L e v y ( λ ) φ * u | v | 1 / λ
where, u , v follow a standard normal distribution, λ = 1.5, and φ is defined as follows.
φ = [ Γ ( 1 + λ ) sin ( π λ / 2 ) Γ [ ( 1 + λ ) / 2 ] * λ * 2 ( λ 1 ) / 2 ] 1 / λ
Combined with the improved Lévy flight, the new algorithm has a better coordination between exploratory and exploitative patterns. At the same time, the greedy selection is added into the algorithm, which helps the PSO to only preserve the better-quality solutions with regard to the fitness values. The implementation of this process is described as follows.
X g + 1 , i = { X g + 1 , i , f ( X n e w , i ) > f ( X g + 1 , i ) X n e w , i ,   e l s e
The specific implementation steps of the proposed INPSO algorithm are presented as follows:
Step 1: Set the size of the population to N, dimension D, maximum number of iterations t max and initialize the w , r 1 , r 2 values.
Step 2: Initialize population { X i , i = 1 , 2 , , N } using Tent mapping defined by Equation (10).
Step 3: Calculate the individual fitness value { f i , i = 1 , 2 , , N } .
Step 4: Select the global and local optimal solution.
Step 5: Use Equation (12) to update the location of particles, then recalculate the fitness values and update the values of the particles.
Step 6: Select the particles according to the greed principle described by Equation (15).
Step 7: Judge whether t reaches t max value, if reached, the fitness value of α is output, that is, the best solution. Else go to Step 3.

3.3.2. Optimization of SVM Parameters Based on INPSO

The position coordinates corresponding to the optimal solution of the particle swarm are the best values of SVM parameters c and g. The flow chart of the SVM parameters tuning process based on INPSO is illustrated in Figure 3.

4. Validation of the Proposed Method

4.1. INPSO Algorithm for Numerical Function Optimization

In order to strictly investigate the capabilities of the INPSO, eight classical benchmark functions were selected for simulation. The equations of selected functions and their brief descriptions are provided in Table 2, where F 1 ~ F 5 are the unimodal benchmark functions, and F 6 ~ F 8 are the multimodal functions. To substantiate the efficiency of INPSO, the SCA, CS and PSO were utilized to optimize these benchmark tasks and the experimental results were compared with those obtained by INPSO. For fair comparisons, the common parameter settings of the four algorithms were the same: population size N = 30, dimensions d = 30 and the maximum number of iterations was 500. For SCA, the parameters were set to be: a = 1, r 1 , r 2 , r 3 and r 4 were random numbers. The parameters of CS were set to be pa = 0.25. For PSO, the parameters were set to be: w max = 0.9 , w min = 0.2 , and c 1 = c 2 = 2 . Then we separately simulated 30 times and recorded the average values and standard deviations. The average values of the experiments were used to reflect the convergence accuracy and speed of the algorithms with a given number of iterations. The standard deviations were used to reflect the stability of the algorithms. The results are shown in Table 3.
To further illustrate the optimization performance of the proposed INPSO, the convergence curves are exhibited in Figure 4. The curves shown in these figures were the average fitness values achieved from 30 independent experiments. As can be seen from Table 3 and Figure 4, in terms of F 1 ~ F 4 , the INPSO algorithm was superior to the other algorithms and could obtain solutions quite close to the global optimums. The solution accuracy of F 1 function even reached the level of 5.97 × 10−63, which showed that the algorithm had obvious superiorities and made up for the deficiencies of the basic PSO algorithm. On two multimodal functions ( F 6 and F 7 ), though the INPSO algorithm could not get the minimum standard deviations under the specified maximum number of iterations, the best and mean of the objective values performed better than the other four algorithms. For the convergence curves of F 5 and F 8 , its convergence speed was obviously faster than the other three algorithms during the initial stage, and the curves showed that the INPSO algorithm could reveal a promising exploitation potential. To sum up, all simulation results indicated that the proposed INPSO had better performances in search accuracy and speed.

4.2. Case 1: Experiment Results on the CWRU Bearing Dataset and Performance Analysis

4.2.1. Data Collection

In this section, vibration signals of different fault states collected by the motor experimental bench of Case Western Reserve University [34] were used as experimental data to verify the effectiveness of the proposed method in rolling bearings fault diagnosis. As shown in Figure 5, the experimental platform mainly consisted of a torque transducer, a load motor and a three-phase induction motor. In the test, vibration signals collected from the drive end (DE) were treated as the research target, where the sample frequency was 12 kHz and the rotation speed was 1730 r/min under the rated load of 3 hp. The experimental device adopted electric spark machining technology to simulate three fault states of rolling bearings: ball fault, inner race fault and out race fault, and the depth of the faults was 0.011 inches. Each fault state included three degrees of fault with width of 0.007 inch, 0.014 inch and 0.021 inch, respectively.
In this experiment, the data set contained 1000 groups of training data, and each training sample contains 864 sampling points. The standardized fault samples under each fault state were randomly divided into training data set and test data set with the scale of 7:3. Since different fault diameters at the same position were regarded as different fault types, so there were ten categories of health conditions in total. We selected the same amount of data under each operating condition, 70 samples were used as the training set and 30 samples were used as the test set. In this way, this method ensured the balance of data and facilitated the diagnosis model to get correct diagnosis results.

4.2.2. 1DCNN Feature Learning Verification and Analysis

In order to demonstrate the superiority of the proposed 1DCNN feature learning method, three other mainstream feature extraction methods, including wavelet packet transform (WPT), convolutional auto-encoder (CAE) and long short-term memory (LSTM), were applied to extract feature vectors for comparison using the same fault dataset. Firstly, the feature extractions of the raw vibration signals were conducted using the above four feature extraction methods. Then the SVM was adopted as the classifier to accomplish the fault classification task. To avoid randomicity, the grid search method and 3-fold cross validation were adopted for the parameter optimization of SVM, and the parameters c and g were set in the ranges of [0.001, 0.1] and [1, 100], respectively.
The main parameters of the other methods are briefly described as follows: (1) WPT: the energy spectra were calculated after the vibration signals were decomposed by three-layer wavelet packet decomposition. (2) CAE: using two layers of the convolutional layer, two layers of the pooling layer and two layers of the full-connected layer. The convolutional kernels were 64 × 1 and 8 × 1, the length of each pooling layer was 2, the Adam optimization algorithm was used and the activation function was ReLu. (3) LSTM: the model consisted of two LSTM layers, two fully connected layers, and one softmax layer. The dropout rate was 0.2, the batch size was 128, and the iteration epoch was 30.
As the dataset and the classifier used in the contrast experiments were the same, the differences in fault diagnosis performance were totally caused by the fault extraction processes. The precision rates and the recall rates of the four methods (i.e., 1DCNN-SVM, WPT-SVM, CAE-SVM and LSTM-SVM) are summarized in Table 4. To illustrate the differences more directly, F1-measure values (percentage form) of the four methods are presented in Figure 6.
From Table 4 and Figure 6, it can be observed that the average precision rate of SVM with manual feature extraction (WPT) was only 80.8%, while the deep learning models (LSTM-SVM, CAE-SVM) had better performances (about 10% improvement in precision rate) compared with traditional WPT feature extraction method. The 1DCNN-SVM model achieved the best average precision rate (95%) using a four-layer convolutional neural network. The accuracy of the proposed 1DCNN-SVM algorithm was obviously better than other competitors, which meant that the 1DCNN had advantages in feature learning of vibration signals compared with other intelligent feature extraction methods, e.g., CAE, LSTM. All four algorithms had high diagnosis accuracies for the normal state (condition 0), however the performance of 1DCNN-SVM algorithm was significantly better than those of the other three algorithms for the micro-fault identifications (condition 1 and condition 3). In summary, compared with WPT-SVM, CAE-SVM, and LSTM-SVM, the 1DCNN-SVM algorithm had a more superior overall performance both in diagnosis accuracy and stability, which indicated that 1DCNN was a more promising feature extraction tool for rolling bearings fault diagnosis. In addition, 1DCNN could extract the features of the raw vibration signals directly without a complicated signal preprocessing process, which had advantages over traditional methods.
Generally, the CNN was considered as a black box, and its inner working mechanism was difficult to understand [35]. In this paper, we tried to explore the internal operation process of the proposed 1DCNN model with a visual method. In order to explore the network feature extraction ability of each layer of 1DCNN, a t-Distributed Stochastic Neighbor Embedding(t-SNE) method was adopted in this paper to visualize the outputs of each network layer by mapping the high-dimensional feature vectors into three-dimensional space, and the data feature distributions of each layer after dimension reduction are shown in Figure 7. Figure 7a indicates that the original data sampled under 10 running states were overlapped seriously. However, it can be seen clearly from Figure 7b–f that the data sampled under different fault conditions were gradually separated due to the feature extraction ability of each layer being gradually enhanced. Finally, after the fully connected layer, the final obtained features were easily distinguishable which meant that the 1DCNN network had a good feature learning capability.

4.2.3. Bearing Fault Diagnosis Experiment

Fault Diagnosis Results and Evaluation of the Proposed INPSO Optimization Method

During the optimization stage, the proposed INPSO approach was used to optimize the penalty factor c and the kernel parameter g, where the number of particles was set to 20, the number of iterations was 100, and the search ranges of c and g were set to [0.001,10]. Fitness values were calculated by threefold cross validation for the training samples. A group of samples was randomly selected as the test set, and the proposed 1DCNN was applied for feature extraction. The extracted feature vectors were respectively fed into three classifiers (SVM, PSO-SVM and INPSO-SVM), where the parameter settings of the PSO and INPSO algorithms were the same. Each method was performed 10 times and the diagnosis accuracy of three classifiers are listed in Table 5. As can be seen from Table 5, in the standard SVM, the default parameters c and g of SVM respectively were 1 and 1 / n , where n was the data dimension. The fault diagnosis accuracy of the test set was 93.6% without the parameters tuning process of SVM. As a contrast, the PSO algorithm was used to optimize SVM parameters, and the average fault diagnosis accuracy of PSO-SVM was 90.56% with a standard deviation of 10.52. This means that the diagnosis accuracy was not stable and the parameter optimization process of the PSO was easily trapping into local optimums. In order to further evaluate the effectiveness of the proposed INPSO method, with the same test set, SVM optimized by INPSO was applied to do fault diagnosis experiments and the experimental results revealed that the average accuracy of the test set could reach 96.97%, and the standard deviation was only 0.69. Therefore, we could conclude that the diagnosis accuracy and robustness of INPSO-SVM was better than PSO-SVM and SVM method, which meant that the proposed INPSO overcame the shortcomings of original PSO, e.g., trapping into local optimums, and had a stronger global optimization ability.
Figure 8 and Figure 9 depict the convergence curves of INPSO algorithm compared with the original PSO algorithm in the first trial. The optimal fitness value of PSO and INPSO algorithm in the first iteration were both 70.72%. With the progress of iteration, the accuracies of the SVM optimized by two algorithms rose rapidly, while the particles fell into local optimums in the second iteration. At this point, the fitness achieved 80.87%. After a couple of iterations, the optimal fitness value of PSO algorithm did not change, while Lévy flight helped the particles jump out of the local optimums and find a better global solution in the proposed INPSO algorithm. The accuracy of the proposed INPSO-SVM method was higher than that of the PSO-SVM, reaching up to 100.0% in the training process. The training results also showed that the proposed INPSO approach had a better performance in the parameters tuning process and INPSO-SVM achieved a higher diagnosis accuracy.

Comparisons Among Different Fault Diagnosis Methods

In this subsection, comparative studies between the WPT-INPSO-SVM and the proposed 1DCNN-INPSO-SVM were performed to verify the effectiveness and superiorities of the proposed algorithm. In the comparative experiments, there were three points that need to be emphasized. Firstly, the 8-dimensional vectors extracted by the three-layer wavelet packet were used as the inputs of the INPSO-SVM classifier for fault diagnosis, and the parameters of the proposed INPSO algorithm were consistent in the contrast experiments. Secondly, for the fault diagnosis comparison test, two algorithms processed the same ten group of data (A-J) which were randomly selected from the original signal. Thirdly, the average values of the two models were taken as the final experimental results after 10 trials for each group of data, where particles iterated 100 times to find the optimal value. Diagnosis results are listed in Table 6. There were some identical recognition results, such as the 80% accuracy for the first set of data. Because the raw data used in the ten experiments were the same, and WPT used the same three-layer wavelet packet to extract 8-dimensional feature vectors, which meant that the extracted features by WPT were the same in the ten times of experiments. In addition, the INPSO algorithm might find the same optimal parameter values for SVM. Therefore, the raw data, the extracted features and the models of classifiers were all the same in this case, and the experimental accuracies for the first set of data were all 80%. For datasets B, C, D, E, F, G and I, the INPSO algorithm found different optimal parameters for SVM, which meant the classifier models were different, we could see that the average testing accuracies of the 1DCNN-INPSO-SVM method were higher than those of the WPT-INPSO-SVM in all 10 groups of comparative experiments, which indicated that 1DCNN could extract richer invariant features compared with WPT.

4.3. Case 2: Experiment Results on the Jiang Nan University Bearing Dataset and Performance Analysis

4.3.1. Data Collection

The Jiang Nan University dataset was collected from a centrifugal fan system rolling bearing fault diagnosis test rig [36]. Figure 10 displays the experimental setup for rolling bearing fault diagnosis. The experimental bench used a PCB MA352A60 acceleration sensor to collect the vibration signals in the vertical direction. The speed was 1000 r/min, and the sampling frequency was set to 50 kHz.
This study considered four bearing health conditions, including outer-race fault, inner-race fault, roller element fault and normal. Bearings under each health condition were operated at 800 rpm rotating speeds. These fault bearings are shown in Figure 11. In this experiment, the data set contained 2000 groups of training data, and each training sample contained 2400 sampling points. The standardized fault samples under each fault state were randomly divided into training data set and test data set with the scale of 7:3. We selected the same amount of data under each operating condition, 350 samples were used as the training set and 150 samples were used as the test set. In this way, this method ensured the balance of data and facilitated the diagnosis model to get correct diagnosis results.

4.3.2. Bearing Fault Diagnosis Experiment

Fault Diagnosis Results and Evaluation of the Proposed INPSO Optimization Method

During the optimization stage of SVM parameters (penalty factor c and the kernel parameter g), the proposed INPSO approach was adopted to tune appropriate parameters to improve fault diagnosis accuracy. The parameter settings of the algorithm were the same as in case 1. A group of samples was randomly selected as the test set, and the proposed 1DCNN was applied for feature extraction. Then the extracted feature vectors were respectively fed into the INPSO-SVM classifier, the SVM and PSO-SVM classifiers which were implemented as comparisons in this study. Each method was performed 10 times to reduce the effect of randomness and the diagnosis accuracy of the three classifiers are listed in Table 7. From Table 7, it can be seen that the fault diagnosis accuracy of SVM was only 91.98% without the parameters tuning process. In addition, the PSO-SVM performed better than SVM, where the average fault diagnosis accuracy was 93.3% with a standard deviation of 0.13. With the same test set, the proposed INPSO-SVM achieved the best performance in each trial and the average fault diagnosis could reach 94.186%. The standard deviation was 0.06, which indicated that the INPSO-SVM had good stability. Therefore, it could be concluded that the proposed algorithm had a stronger global optimization ability and played an important role in improving fault diagnosis accuracy.

Comparisons Among Different Fault Diagnosis Methods

In order to further prove that the proposed 1DCNN-INPSO-SVM could extract robust features and improve fault diagnosis accuracy, the WPT-INPSO-SVM method was used for comparative analysis. In the comparative experiments, the points to emphasize were the same as in case 1. The classification accuracies were evaluated and compared in Table 8. Table 8 lists the results obtained using the random 10 sets of data, where each set of data was used to test the 1DCNN-INPSO-SVM and the WPT-INPSO-SVM ten times respectively. The last column of Table 8 indicates that the proposed method achieved better performance, with the highest accuracy of 97.186%, while the competitor utilizing the features extracted by WPT obtained a lower accuracy.

5. Conclusions

As a promising machine learning algorithm, the convolutional neural network has been successfully applied in the field of intelligent fault diagnosis. In this paper, a novel hybrid fault diagnosis approach based on 1DCNN and INPSO-SVM is proposed for the fault diagnosis of rolling bearings. Firstly, the raw fault data is directly adopted as the inputs of the 1DCNN model to complete the training of CNN model, which can extract the invariant features of the original vibration signals. Moreover, in order to overcome the shortcomings of the basic PSO algorithm, e.g., trapping in the local optimums, a novel PSO variant INPSO is proposed, which combines two efficient strategies: Lévy flight and Chaos based initialization strategy. Finally, INPSO is used to tune the parameters of SVM and the features extracted by 1DCNN are used as the input of the INPSO-SVM classifier for fault classification. The proposed model can directly process the original vibration signals without any time-consuming manual feature extraction process. In addition, this method combines the excellent classification performance of SVM for small samples and the strong feature learning abilities of the CNN network, which avoids the problems of low representativeness and poor discrimination of traditional manual feature extraction, and improves the fault diagnosis accuracy. The INPSO algorithm is proposed to realize the automatic optimization of SVM parameters, which facilities the efficiency of SVM parameter optimization.
Extensive experiments were carried out to illustrate the effectiveness of the proposed method. First of all, numerical optimization experiments were conducted based on 8 widely applied benchmark functions and the simulation results indicated that the proposed PSO variant had a better performance in searching accuracy and convergence speed. Then, the feature learning ability of the 1DCNN was compared with three other feature extraction methods, including WPT, CAE and LSTM. The feature extraction results, which were presented in a novel visual manner, showed that the 1DCNN achieved a better performance. Finally, two cases of rolling bearing fault diagnosis experiments were carried out and experimental results showed that the proposed 1DCNN-INPSO-SVM hybrid framework was capable of diagnosing with high precision for rolling bearings and superior to some traditional fault diagnosis methods.
However, the proposed 1DCNN-INPSO-SVM does not use the temporal features of the vibration signals. The future work will focus on the extraction and utilization of the spatial-temporal characteristics. In addition, how to reduce the computational load and speed up the training process still deserve further study.

Author Contributions

Conceptualization, Y.S. (Yang Shao) and X.Y.; investigation, Y.S. (Yang Shao), X.Y. and Y.S. (Yang Song); methodology, Y.S. (Yang Shao) and X.Y.; software, Y.S. (Yang Shao); supervision, X.Y. and C.Z.; validation, Y.S. (Yang Shao), X.Y. and Q.X.; writing–original draft, Y.S. (Yang Shao) and X.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (Grant No. 61803227, 61973184, 61773242, 61603214), the National Key R & D Program of China (Grant No.2017YFB1302400), Independent Innovation Foundation of Shandong University (Grant No. 2082018ZQXM005).

Acknowledgments

We would like to thank the editors and the anonymous reviewers for their insightful comments and constructive suggestions.

Conflicts of Interest

The authors declare that there is no conflict of interests regarding the publication of this paper.

References

  1. Wang, S.; Selesnick, I.; Cai, J.; Feng, Y.; Sui, X.; Chen, X. Nonconvex sparse regularization and convex optimization for bearing fault diagnosis. IEEE Trans. Ind. Electron. 2018, 65, 7332–7342. [Google Scholar] [CrossRef]
  2. Jia, F.; Lei, Y.R.; Lu, N.; Xing, S. Deep normalized convolutional neural network for imbalanced fault classification of machinery and its understanding via visualization. Mech. Syst. Signal Process. 2018, 110, 349–367. [Google Scholar] [CrossRef]
  3. Jiang, W.; Cheng, C.; Zhou, B.; Ma, G.; Yuan, Y. A Novel GAN-based Fault Diagnosis Approach for Imbalanced Industrial Time Series. arXiv 2019, arXiv:1904.00575. [Google Scholar]
  4. Lei, Y.; Jia, F.; Zhou, X.; Lin, J. A deep learning-based method for machinery health monitoring with big data. J. Mech. Eng. 2015, 51, 49–56. [Google Scholar] [CrossRef]
  5. Sun, Y.; Gao, H.; Hong, X.; Song, H.; Liu, Q. Fault Diagnosis for Rolling Bearing Based on Deep Residual Neural Network. In Proceedings of the 2018 International Conference on Sensing, Diagnostics, Prognostics, and Control (SDPC), Xi’an, China, 15–17 August 2018. [Google Scholar]
  6. Zhao, M.; Kang, M.; Tang, B.; Pecht, M. Deep residual networks with dynamically weighted wavelet coefficients for fault diagnosis of planetary gearboxes. IEEE Trans. Ind. Electron. 2018, 65, 4290–4300. [Google Scholar] [CrossRef]
  7. Bi, F.; Ma, T.; Wang, X. Development of a novel knock characteristic detection method for gasoline engines based on wavelet-denoising and EMD decomposition. Mech. Syst. Signal. Process. 2019, 117, 517–536. [Google Scholar] [CrossRef]
  8. Tong, Z.; Li, W.H.; Zhang, B.; Li, B. Bearing fault diagnosis based on domain adaptation using transferable features under different working conditions. Shock. Vib. 2018, 2018, 1–12. [Google Scholar] [CrossRef]
  9. Chen, H.; Jiang, B.; Chen, W.; Yi, H. Data-Driven detection and diagnosis of incipient faults in electrical drives of high-speed trains. IEEE Trans. Ind. Electron. 2019, 66, 4716–4725. [Google Scholar] [CrossRef]
  10. Pan, L.; Zhu, D.; Shen, S.; Song, A.; Shi, X.; Duan, S. Gear fault diagnosis method based on wavelet-packet independent component analysis and support vector machine with kernel function fusion. Adv. Mech. Eng. 2018, 10, 1–10. [Google Scholar] [CrossRef] [Green Version]
  11. Tong, Z.; Li, W.; Zhang, B.; Jiang, F.; Zhou, G. Online bearing fault diagnosis based on a novel multiple data streams transmission Scheme. IEEE Access 2019, 7, 66644–66654. [Google Scholar] [CrossRef]
  12. He, Z.; Cheng, J.L.; Yang, Y. Linear maximum margin tensor classification based on flexible convex hulls for fault diagnosis of rolling bearings. Knowl. Based Syst. 2019, 173, 62–73. [Google Scholar] [CrossRef]
  13. Imani, M.; Dougherty, E.R.; Braga-Neto, U. Boolean Kalman filter and smoother under model uncertainty. Automatica 2020, 111, 108609. [Google Scholar] [CrossRef]
  14. Imani, M.; Ghoreishi, S.F. Bayesian Optimization Objective-Based Experimental Design; American Control Conference (ACC): Denver, CO, USA, 2020. [Google Scholar]
  15. Shao, H.; Jiang, H.; Lin, Y.; Li, X. A novel method for intelligent fault diagnosis of rolling bearings using ensemble deep auto-encoders. Mech. Syst. Signal. Process. 2018, 102, 278–297. [Google Scholar] [CrossRef]
  16. Gong, W.; Chen, H.; Zhang, Z.; Zhang, M.; Wang, R.; Guan, C.; Wang, Q. A novel deep learning method for intelligent fault diagnosis of rotating machinery based on improved CNN-SVM and multichannel data fusion. Sensors 2019, 19, 1693. [Google Scholar] [CrossRef] [Green Version]
  17. Chen, P.; Yuan, L.; He, Y. Deep learning. Neurocomputing 2015, 521, 436–444. [Google Scholar]
  18. Hinton, G.E.; Salakhutdunov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [Green Version]
  19. Chen, P.; Yuan, L.; He, Y. Gearbox fault identification and classification with convolutional neural networks. Shock. Vib. 2016, 211, 202–211. [Google Scholar] [CrossRef] [Green Version]
  20. Wang, Q.; Zhao, B.; Ma, H.; Chang, J.; Mao, G. Fault diagnosis method based on FFT-RPCA-SVM for cascaded-multilevel inverter. ISA Trans. 2015, 2015, 1–10. [Google Scholar] [CrossRef]
  21. Janssens, O.; Slavkovikj, V.; Vervisch, B. Convolutional neural network based fault detection for rotating machinery. J. Sound Vib. 2016, 337, 331–345. [Google Scholar] [CrossRef]
  22. Fu, W.; Tan, J.; Zhang, X.; Chen, T.; Wang, K. Blind parameter identification of MAR model and mutation hybrid GWO-SCA optimized SVM for fault diagnosis of rotating machinery. Complexity 2019, 51, 1–17. [Google Scholar] [CrossRef]
  23. Yan, X.; Jia, M. A novel optimized SVM classification algorithm with multi-domain feature and its application to fault diagnosis of rolling bearing. Neurocomputing 2018, 313, 47–64. [Google Scholar] [CrossRef]
  24. Li, H.; He, C.; Malekian, R.; Li, Z. Weak defect identification for centrifugal compressor blade crack based on pressure sensors and genetic algorithm. Sensors 2018, 18, 1264. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Hao, X.; Zhang, G.; Ma, S. Deep Learning. Proc. Int. J. Semant. Comput. 2016, 10, 417–439. [Google Scholar] [CrossRef] [Green Version]
  26. Zhang, W.; Peng, G.; Li, C.; Chen, Y.; Zhang, Z. A new deep learning model for fault diagnosis with good anti-noise and domain adaptation ability on raw vibration signals. Sensors 2017, 17, 425. [Google Scholar] [CrossRef] [PubMed]
  27. Xia, M.; Li, T.; Xu, L.; Liu, L.; Silva, C. Fault diagnosis for rotating machinery using multiple sensors and convolutional neural networks. IEEE/ASME Trans. Mechatron. 2018, 23, 101–110. [Google Scholar] [CrossRef]
  28. Zheng, J.; Pan, H.; Cheng, J. Rolling bearing fault detection and diagnosis based on composite multiscale fuzzy entropy and ensemble support vector machines. Mech. Syst. Signal Process. 2017, 85, 746–759. [Google Scholar] [CrossRef]
  29. Vapnik, V.; Liu, L. The Nature of Statistical Learning Theory; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  30. Teng, Z.; Lv, J.; Guo, L. An improved hybrid grey wolf optimization algorithm. Soft Comput. 2019, 23, 6617–6631. [Google Scholar] [CrossRef]
  31. Oliveira, J.; Oliveira, P.M.; Cunha, J.B.; Pinho, T. Chaos-based grey wolf optimizer for higher order sliding mode position control of a robotic manipulator. Nonlinear Dyn. 2017, 90, 1353–1362. [Google Scholar] [CrossRef]
  32. Yuan, X.; Jin, P.; Zhou, G. An improved QPSO algorithm integrating social learning with levy flights. Syst. Sci. Control. Eng. 2019, 6, 362–373. [Google Scholar]
  33. Zhang, W.; Peng, G.; Li, C.; Chen, Y. Antlion optimization algorithm integrating with Levy flight and golden sine. Appli. Res. Comput. 2018, 37, 1–6. [Google Scholar]
  34. Loparo, K. Case Western Reserve University Bearing Data Center, Cleveland, OH, USA, Tech. Rep. 2012. Available online: http://csegroups.case.edu/bearingdatacenter/pages/download-data-file (accessed on 28 September 2018).
  35. Zhang, W.; Li, C.; Peng, G.; Chen, Y.; Zhang, Z. A deep convolutional neural network with new training methods for bearing fault diagnosis under noisy environment and different working load. Mech. Syst. Signal Process. 2018, 100, 439–453. [Google Scholar] [CrossRef]
  36. Li, K. School of Mechanical Engineering, Jiangnan University. 2019. Available online: http://mad-net.org:8765/explore.html?t=0.5831516555847212 (accessed on 12 September 2019).
Figure 1. The framework of the proposed one-dimensional convolutional neural network improved novel particle swarm optimization support vector machine (1DCNN-INPSO-SVM) intelligent fault diagnosis method.
Figure 1. The framework of the proposed one-dimensional convolutional neural network improved novel particle swarm optimization support vector machine (1DCNN-INPSO-SVM) intelligent fault diagnosis method.
Applsci 10 04303 g001
Figure 2. Structure of the proposed 1DCNN model.
Figure 2. Structure of the proposed 1DCNN model.
Applsci 10 04303 g002
Figure 3. INPSO-SVM flow chart.
Figure 3. INPSO-SVM flow chart.
Applsci 10 04303 g003
Figure 4. Convergence trends for INPSO versus other optimizers: (a) Sphere function, (b) Schwefel 2.22 function, (c) Schwefel 1.2 function, (d) Schwefel 2.21 function, (e) Rosenbrock function, (f) Rastrigin function, (g) Griewank function, (h) penalized function.
Figure 4. Convergence trends for INPSO versus other optimizers: (a) Sphere function, (b) Schwefel 2.22 function, (c) Schwefel 1.2 function, (d) Schwefel 2.21 function, (e) Rosenbrock function, (f) Rastrigin function, (g) Griewank function, (h) penalized function.
Applsci 10 04303 g004aApplsci 10 04303 g004b
Figure 5. The rolling bearing fault data acquisition experimental bench.
Figure 5. The rolling bearing fault data acquisition experimental bench.
Applsci 10 04303 g005
Figure 6. The F1-measures of the bearing data using different methods.
Figure 6. The F1-measures of the bearing data using different methods.
Applsci 10 04303 g006
Figure 7. Scatter plots of the feature visualization by t-SNE: (a) feature distribution of raw data, (b) feature distribution of convolutional layer 1, (c) feature distribution of convolutional layer 2, (d) feature distribution of convolutional layer 3, (e) feature distribution of convolutional layer 4, and (f) feature distribution of the fully connected layer.
Figure 7. Scatter plots of the feature visualization by t-SNE: (a) feature distribution of raw data, (b) feature distribution of convolutional layer 1, (c) feature distribution of convolutional layer 2, (d) feature distribution of convolutional layer 3, (e) feature distribution of convolutional layer 4, and (f) feature distribution of the fully connected layer.
Applsci 10 04303 g007aApplsci 10 04303 g007b
Figure 8. Diagnosis accuracy of PSO-SVM algorithm.
Figure 8. Diagnosis accuracy of PSO-SVM algorithm.
Applsci 10 04303 g008
Figure 9. Diagnosis accuracy of INPSO-SVM algorithm.
Figure 9. Diagnosis accuracy of INPSO-SVM algorithm.
Applsci 10 04303 g009
Figure 10. Experimental setup for the rolling bearing fault diagnosis.
Figure 10. Experimental setup for the rolling bearing fault diagnosis.
Applsci 10 04303 g010
Figure 11. Bearing fault: (a) outer-race fault, (b) inner-race fault, and (c) roller element fault.
Figure 11. Bearing fault: (a) outer-race fault, (b) inner-race fault, and (c) roller element fault.
Applsci 10 04303 g011
Table 1. Parameters of 1DCNN algorithm applied in the experiments.
Table 1. Parameters of 1DCNN algorithm applied in the experiments.
No.Layer TypeKernel SizeStridesKernel NumberOutput Size (Width × Depth)
1Conv164 × 116 × 116128 × 16
2Pool12 × 12 × 11664 × 16
3Conv23 × 11 × 13264 × 32
4Pool22 × 12 × 13232 × 32
5Conv33 × 11 × 16432 × 64
6Pool32 × 12 × 16416 × 64
7Conv43 × 11 × 16416 × 64
8Pool42 × 12 × 1648 × 64
9F300 1300 × 1
10F10 110 × 1
Table 2. Benchmark functions.
Table 2. Benchmark functions.
Function ExpressionNamedSearch Interval f min
F 1 ( x ) = i = 1 n x i 2 Sphere30[−100, 100]n0
F 2 ( x ) = i = 1 n | x i 2 | + i = 1 n | x i | Schwefel 2.2230[−10, 10]n0
F 3 ( x ) = i = 1 n ( j = 1 i x j ) 2 Schwefel 1.230[−100, 100]n0
F 4 ( x ) = max i { | x i | , 1 i n } Schwefel 2.2130[−100, 100]n0
F 5 ( x ) = i = 1 n 1 [ 100 ( x i + 1 x i 2 ) 2 + ( x i 1 ) 2 ] Rosenbrock30[−30, 30]n0
F 6 ( x ) = i = 1 n [ x i 2 10 cos ( 2 π x i ) + 10 ] Rastrigin30[−5.12, 5.12]n0
F 7 ( x ) = 1 4000 i = 1 n x i 2 i = 1 n cos ( x i i ) + 1 Griewank30[−50, 50]n0
F 8 ( x ) = π n { 10 sin ( π y 1 ) + i = 1 n 1 ( y i 1 ) 2 [ 1 + 10 sin 2 ( π y i + 1 ) ] + ( y n 1 ) 2 } + i = 1 n u ( x i , 10 , 100 , 4 ) Penalized30[−50, 50]n0
Table 3. Optimization results on unimodal and multimodal functions.
Table 3. Optimization results on unimodal and multimodal functions.
FunctionIndicatorsAlgorithms
SCACSPSOINPSO
F 1 Mean38.941411.53631.26 × 10−45.97 × 10−63
S.D.72.50093.98370.00023.27 × 10−62
Best0.08665.72005.61 × 10−61.49 × 10−84
F 2 Mean0.03645.91927.04561.34 × 10−54
S.D.0.05081.96149.51276.91 × 10−54
Best0.00123.25510.00845.94 × 10−67
F 3 Mean8023.48892475.3582.43090.0010
S.D.4265.1060607.63340.60340.0038
Best867.62011310.470891.93552.58 × 10−6
F 4 Mean36.210111.12601.07807.63 × 10−9
S.D.10.79711.76310.22704.12 × 10−8
Best17.37977.79000.57663.04 × 10−15
F 5 Mean30779.15623.5256177.535726.5895
S.D.44845.7354.2101551.94380.1765
Best17.3797242.610521.643326.1215
F 6 Mean36.2101100.6785102.862013.4779
S.D.38.172814.017727.396724.7228
Best0.008877.239147.61990
F 7 Mean0.91521.10880.00800.0041
S.D.0.36640.03890.00880.0226
Best0.09711.04708.27 × 10−70
F 8 Mean100,0993.88710.06220.0231
S.D.417,1030.88500.08430.0818
Best0.06891.92528.91 × 10−73.44 × 10−7
Table 4. The precision rate and recall rate of different methods.
Table 4. The precision rate and recall rate of different methods.
Bearing ConditionMethods
1DCNN-SVMWPT-SVMCAE-SVMLSTM-SVM
P(%)R(%)P(%)R(%)P(%)R(%)P(%)R(%)
Condition 010010096808597100100
Condition 18680746770876677
Condition 2971008183100100100100
Condition 396906983889359100
Condition 410083917097978660
Condition 594100100979497100100
Condition 69197524310067100100
Condition 7100100100100100100100100
Condition 88610010010097976930
Condition 91009745679683100100
Average value9594.780.87992.791.88886.7
Table 5. The accuracy of different fault diagnosis methods.
Table 5. The accuracy of different fault diagnosis methods.
Run NumberAccuracy
SVMPSO-SVMINPSO-SVM
194.6775.3395.33
293.679797
39175.3397.33
495.3397.3397
594.6797.3397
694.6797.3396.33
79496.3397.67
891.6797.3397
994.6775.3397.33
1091.679797.67
Table 6. Recognition results of ten trials using different methods.
Table 6. Recognition results of ten trials using different methods.
DataModel12345678910Average
A1DCNN-INPSO-SVM92.339292.3392.33929292.339292.3392.3392.198
WPT-INPSO-SVM8080808080808080808080
B1DCNN-INPSO-SVM93.3393.3392.6793.3393.3392.6793.339192.6793.6792.933
WPT-INPSO-SVM83.6783.678483.6783.6783.6783.6783.6783.6783.6783.703
C1DCNN-INPSO-SVM93.3393.6793.6793.339493.6793.679493.6793.6793.668
WPT-INPSO-SVM8584.6785858584.6784.678584.678584.868
D1DCNN-INPSO-SVM94.339394.3394.3393.6791.6794.33929191.3392.999
WPT-INPSO-SVM8585.3385.338585.3385.3385.3385.3385.3385.3385.264
E1DCNN-INPSO-SVM95.6791.33969695.3395.3395.3395.3395.339695.165
WPT-INPSO-SVM8382.678383838382.6783838382.934
F1DCNN-INPSO-SVM9088.6789.339089.6790.33909090.3390.3389.866
WPT-INPSO-SVM83.6783.6783.6783.3383.3383.6783.6783.6783.3383.6783.568
G1DCNN-INPSO-SVM97.6797.6797.3397.679897.339897.6797.6797.6797.668
WPT-INPSO-SVM82.3382.3382.678282.3382.3382.6783838382.566
H1DCNN-INPSO-SVM92.3392.3392.3391.6792.3392.3392.3392.3392.3392.3392.264
WPT-INPSO-SVM8585858585858585858585
I1DCNN-INPSO-SVM95959595.6795.339695.33969595.3395.366
WPT-INPSO-SVM83.338883.3383.3383.3383.3383.3383.3383.3383.3383.797
J1DCNN-INPSO-SVM92.6792.679292.6792.6792.3392.3392.6792.6792.3392.501
WPT-INPSO-SVM79.6779.6779.6779.6779.6779.6779.6779.6779.6779.6779.67
Table 7. The influence of parameter optimization on fault diagnosis accuracy.
Table 7. The influence of parameter optimization on fault diagnosis accuracy.
Run NumberAccuracy
SVMPSO-SVMINPSO-SVM
19293.1794.17
293.3393.1794.17
39293.3394.17
491.593.3394.17
593.3393.1794.17
690.3393.1794.17
791.1693.3394.33
892.6693.594.17
992.1693.594.17
1091.3393.3394.17
Table 8. Recognition results of ten trials using different methods.
Table 8. Recognition results of ten trials using different methods.
DataModel12345678910Average
A1DCNN-INPSO-SVM94.1794.1794.1794.1794.1794.1794.3394.1794.1794.1794.186
WPT-INPSO-SVM84848484.1783.178483.8384.17848483.934
B1DCNN-INPSO-SVM97.1797.1797.1797.1797.1797.1797.1797.3397.1797.1797.186
WPT-INPSO-SVM86.8387.1786.8386.8386.838786.8386.8386.838786.898
C1DCNN-INPSO-SVM94.3394.3394.3394.3394.3394.3393.594.3394.3394.3394.247
WPT-INPSO-SVM8584.6784.6784.678584.6784.6784.6784.6784.6784.736
D1DCNN-INPSO-SVM90.3390.3390.590.3390.3390.3390.590.590.3390.3390.381
WPT-INPSO-SVM86.3386.3386.3386.3386.586.586.3386.1786.6786.3386.382
E1DCNN-INPSO-SVM93.33939393.3393.3393.3393.3393.3393.3392.8393.214
WPT-INPSO-SVM85.6785.3385.3385.3385.3385.3385.3385.3385.3385.3385.364
F1DCNN-INPSO-SVM94.6794.6795.3394.8394.8394.8394.8394.839594.6794.849
WPT-INPSO-SVM84.1784.1784.1784.1784.1784.1784.1784.1784.1784.1784.17
G1DCNN-INPSO-SVM95.595.595.595.6795.595.595.595.595.595.595.517
WPT-INPSO-SVM83.6783.6783.6783.1783.6783.6783.6783.6783.6782.6783.52
H1DCNN-INPSO-SVM94.594.594.594.594.594.594.594.594.594.3394.483
WPT-INPSO-SVM84.3384.3384.3384.3384.3384.3384.3384.1784.1784.3384.298
I1DCNN-INPSO-SVM93.593.593.593.593.6793.593.6793.6793.593.6793.568
WPT-INPSO-SVM85.6785.6785.6785.585.6785.6785.6785.6785.585.6785.636
J1DCNN-INPSO-SVM9292929292929292929292
WPT-INPSO-SVM83.583.583.583.583.583.583.583.583.583.3383.483

Share and Cite

MDPI and ACS Style

Shao, Y.; Yuan, X.; Zhang, C.; Song, Y.; Xu, Q. A Novel Fault Diagnosis Algorithm for Rolling Bearings Based on One-Dimensional Convolutional Neural Network and INPSO-SVM. Appl. Sci. 2020, 10, 4303. https://doi.org/10.3390/app10124303

AMA Style

Shao Y, Yuan X, Zhang C, Song Y, Xu Q. A Novel Fault Diagnosis Algorithm for Rolling Bearings Based on One-Dimensional Convolutional Neural Network and INPSO-SVM. Applied Sciences. 2020; 10(12):4303. https://doi.org/10.3390/app10124303

Chicago/Turabian Style

Shao, Yang, Xianfeng Yuan, Chengjin Zhang, Yong Song, and Qingyang Xu. 2020. "A Novel Fault Diagnosis Algorithm for Rolling Bearings Based on One-Dimensional Convolutional Neural Network and INPSO-SVM" Applied Sciences 10, no. 12: 4303. https://doi.org/10.3390/app10124303

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop