Next Article in Journal
3D Measurement Method for Saturated Highlight Characteristics on Surface of Fuel Nozzle
Previous Article in Journal
IoT and Satellite Sensor Data Integration for Assessment of Environmental Variables: A Case Study on NO2
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Rolling Bearing Fault Diagnosis Based on Conditional Depth Convolution Countermeasure Generation Networks under Small Samples

1
School of Computer, Hunan University of Technology, Zhuzhou 412007, China
2
School of Automation, Central South University, Changsha 410083, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(15), 5658; https://doi.org/10.3390/s22155658
Submission received: 24 June 2022 / Revised: 25 July 2022 / Accepted: 26 July 2022 / Published: 28 July 2022
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

:
Aiming at the problems of low fault diagnosis accuracy caused by insufficient samples and unbalanced data sample distribution in bearing fault diagnosis, this paper proposes a fault diagnosis method for rolling bearings referencing conditional deep convolution adversarial generative networks (C−DCGAN) for efficient data augmentation. Firstly, the concept of conditional constraints is used to guide and improve the sample generation process of the original generative adversarial network, and specific constraints are added to the data generation model to perform a balanced expansion of muti-category fault data for small sample data sets. Secondly, aiming at the phenomena of training instability, gradient disappearance and gradient explosion in the imbalanced sample set, it is proposed to optimize the structure of the generative network by using the structure of self-defined skip connections and spectral normalization, while using the Wasserstein distance with penalty term instead of cross entropy. The function is used as the loss function of the generative adversarial network to improve the stable feature extraction ability of the generative network and the effect of the training process; in this way, simulation sample data with only a small variation from the real data distribution can be generated. Finally, the complete fault data set (after mixing the original data with sufficient fault category and sample number) and the generated data are input into the one-dimensional convolution neural network for fault diagnosis of rolling bearing. The experiment’s results show that the diagnosis method in this paper can improve the fault classification effect of rolling bearings by generating balanced and sufficient sample data.

1. Introduction

Mechanical equipment is composed of many parts, and the bearing is an important part of heavy industry; over the long-term operation of mechanical equipment, the bearings may often be easily damaged. In daily operation, when the mechanical failure of bearings occurs, it risks accidents, even ones resulting in losses to factories and enterprises [1,2,3]. Therefore, if a specific fault diagnosis is made for a rolling bearing in the process of operation, the fault types of the bearing can be identified in time, and the faults of the bearing can be repaired or replaced to different degrees; such precautions can effectively ensure the operation of the mechanical equipment, and avoid accidents [4,5].
In terms of industrial machinery fault diagnosis, there are mainly two types of fault diagnosis and detection methods. The first is to use the model to compare the output value of the model with the actual signal data during the operation of the machine, so as to distinguish different potential failures caused by the machine, and to infer and classify these predictions according to the failure phenomenon. The second is a data-driven fault diagnosis method, which usually collects a large number of vibration signals during the actual production process when the machinery is running, describing the state information of the machinery operation, and then performs feature learning from the data to classify the faults. The deep learning model has powerful feature extraction capabilities based on its own structural characteristics, and can autonomously extract hidden features, which can be used to deal with a large number of complex fault features. Therefore, in recent years, deep learning-based fault diagnosis methods have been highly applicable to the field of faults diagnostics. Common methods used are auto-encoder (AE) [6], long-short term memory network (LSTM) [7] and Convolutional Neural Network (CNN) [8]. CNN is an artificial neural network based on the connection of multiple convolutional layers. Because the design of its frame structure uses the convolution kernel in the convolutional layer to automatically extract hidden features from sample data, CNN is mainly suitable for use in the fields of model recognition and fault diagnosis and detection [9,10,11]. However, the above three common fault diagnosis methods are all supervised learning methods, and their training process requires a large amount of effective sample data [12]. Due to the lack of fault data in the actual operation of mechanical equipment and the fact that the majority of data collected is normal data, the collected bearing fault data set risks the hazards of insufficient or unbalanced sample data [13,14]. Generally, previous efforts have been unable to train deep learning models with high classification accuracy. In the process of testing, the model ignores some fault categories with less sample data, and as a result, the fault cannot be recognized in time [15]. In view of the problem of small sample data which is therefore unbalanced and insufficient, past efforts have mainly used the following several solutions.
The first method is statistical learning theory [16]. The theory uses the small sample method of machine learning to enhance data, but it has some disadvantages, such as poor generalization and slow computation. The second type of method is image enhancement [17,18]. The basic idea of this method is to adjust and transform the data set of small scale to increase the number of samples, and improve the sample diversity for example, the image is switched to intercept and get different sample images. Although this method has good generalization capacities and can be widely used in the study of one-dimensional or two-dimensional small sample problems, because this method is based on the modification of original sample data, it is easy to cause the data distribution to be too unitary, and the sample repetition rate is high. The third kind of method is the Generative Adversarial Networks (GAN) [19]; the generation structure module and the discriminative structure module in the network are used to generate sample data through adversarial gameplaying in order to solve the problems of insufficient samples or unbalanced samples [20].
For the third type of method, Liu et al. [21] put forward a data set enhancement algorithm which is based on the original sample set. This algorithm divides the original sample data into groups based on several special monomers, and makes use of the combination of different monomers to change the data, thereby enhancing the data from small samples; these changes can be verified by subsequent experiments to improve the performance of learning data features and classification. Meng et al. [22] proposed a rotating machinery fault diagnosis framework based on GANs and multi-sensor data fusion. Xue and others [23] proposed a diagnosis algorithm for sample set enhancement using Wasserstein Generative Adversarial Networks (WGAN), which balanced the data distribution of small sample data sets through WGAN to solve the problem of poor generalization in the training process of fault diagnosis caused by serious imbalances of data. Radford et al. [24] proposed a new generation network, namely the deep convolution generative adversarial networks (DCGAN). On the basis of GAN, the step-size convolution is used to improve both the stability of the network training process and the validity of the generated data. However, DCCAN can only be used for small scale data generation, not for large scale sample data enhancement. Gauthier et al. [25] proposed conditional generative adversarial networks (CGAN), which synchronously inputted the input value and the specific constraint into the network structure and generated the sample data for the specific constraint. However, the problem of instability of the training process in the original GAN cannot be solved. Gao et al. [26] proposed to combine Wasserstein generative adversarial nets (WGAN) and convolutional neural network to construct WG−CNN, so as to carry out bearing fault diagnosis through data generation.
In summary, this paper proposes a bearing fault diagnosis method based on conditional deep convolution generative adversarial networks (C-DCGAN) on the basis of DCGAN and extra constraint condition C. The problem of the unbalanced distribution of generated data is solved by defining constraint condition C which is classification labels to constrain the generating process. According to the experimental results from the Case Western Reserve University (CWRU) Bearing Data Center, this method can effectively enhance muti-category data from small sample sets and improve the accuracy of bearing fault diagnosis while increasing the training stability of the generating network.

2. Related Works

2.1. DCGAN

A GAN consists of a generator and a discriminator. Random data is used as the input value of the generator, and the output G (z) and the real sample value are input into the discriminator at the same time to discriminate the generated data. Through the game between the two, the difference between the data distribution of the generated sample and that of the real sample is gradually reduced to a certain extent, which then permits the inference that the generation model has trained convergence; the generator in the model can then be used to generate data which will be similar to the data distribution of real sample data.
The differences between DCGAN and GAN are that: (1) in the discriminator of DCGAN, the pool layer is removed and the convolution layer is added to the structure; (2) by adding batch normalization, the problem that all fault samples converge to the same point is solved; (3) all the fully connected layers in the GAN structure are removed and replaced by convolution layers; (4) the activation functions of each layer of the DCGAN generator and discriminator are modified and replaced, and the instability and gradient disappearance in the original GAN are solved by using the characteristics and definitions of different activation functions. DCGAN essentially maximizes the discriminator’s accuracy with respect to the data source of the input value and minimizes the difference between the generator’s output and the real sample. This game process can be seen as a maximization and minimization problem, expressed in Formula (1):
min G   max D   V ( G , D ) = E x ~ P d a t a ( x ) [ log D ( x ) ] + E z ~ P z ( z ) [ log ( 1 D ( G ( z ) ) ) ]
Among them, Pdata is the data distribution of the real sample value, and Pz is the data distribution of the random noise

2.2. CGAN

In 2014, based on GAN, condition C was added to each input of its network structure, and a conditional generative adversarial network (CGAN) was proposed [27]. Condition C can be any kind of auxiliary information, such as a specific set of sample input values.
Taking random noise Z and constraint C as input conditions, sample G (Z|C) is generated by the generator. According to constraint C, the discriminator can distinguish whether the input value is the real sample or not. The objective function V (D, G) is:
min G   max D   V ( D , G ) = E x ~ p d a t a ( x ) [ log D ( x | c ) ] + E x ~ P z ( z ) [ log ( 1 D ( G ( z | c ) ) ) ]
CGAN makes use of the constraint condition C to constrain the training process of the generating model, and outputs G(Z|C), which conforms to the specific constraint condition, while the discriminator uses the real sample x and the specific constraint condition C, which is the probability of discriminating the input value as the real sample value.

3. C-DCGAN for Bearing Fault Diagnosis

In this paper, a C-DCGAN network structure is proposed, in which the bearing fault category label is defined as specific constraint C. The fault category label is processed by one-hot encoding, and the one-bit valid code of the original label of the fault data is then defined as the constraint C in this paper. There are 10 categories in total, and the generated data can be restricted. Addressing the problems of training instability, gradient disappearance and gradient explosion of the original network model, solutions are summarized as follows:
(1)
Using custom residual blocks, the regularization effect of the generating network is optimized, the feature extraction ability of the network is enhanced, and the authenticity of the generated simulation fault data is improved.
(2)
In the generator and discriminator, spectral normalization is added to improve the stability of the network model training in this experiment, and to solve the gradient explosion problem.
(3)
The loss function is set as W distance with penalty term, which effectively solves the problem of gradient disappearance.

3.1. Conditional Deep Convolution Antagonism Generation Network Structure

In this paper, C−DCGAN uses the iterative training process to study the relation between the constraint condition C and the mapping relationship between the random noise and the original fault sample data, and also to supervise the generator whether in accordance with condition C; the problem of uneven distribution and insufficient number of bearing fault samples is solved by generating bearing fault data which obey the distribution of real fault samples.
The structure of the C−DCGAN network is shown in Figure 1.
In C-DCGAN, the objective function is:
min G   max D   V ( D , G ) = E x ~ p d a t a ( x ) [ log D ( x ) ] + E x ~ P z ( z ) [ log ( 1 D ( G ( z | c ) ) ) ]
E (·) is the expectation of the objective function, the distribution of the real fault data is Pdata, and the distribution of the input noise data is Pz.
In the generator, the constraint condition C and the random noise Z are input into the generator synchronously as input values, and the five transposed convolution layers are connected to each other; between each transposed convolution layer, the network structure is deepened by the residual structure [28], that is, the transposed convolution layer, the activation function layer, the transposed convolution layer, the summation operation h(x) and the activation function layer are superposed one by one to form a self-defined residual block; using the jump connection in the residual block, the mapping F(x) and the input value x are superimposed for output to reduce additional calculations and parameter values. In the output layer, the Tanh function functions as the activation function, which improves the convergence speed and reduces the number of iterations, while in the input layer and other intermediate layers, the Relu function is chosen as the activation function, and the calculation process is thereby simplified and the problem of gradient disappearance is solved. The network structure deepened in this experiment is only composed of two residual blocks, because excessive network depth may lead to problems such as unstable training and loss of data features, as shown in Figure 2.
In the discriminator, the constraint condition C, the generator output G(Z|C) and the real sample value x are input first, and then the feature is extracted through the defined five-layer convolution layer, and a residual block structure similar to the generator is added between each convolution layer. Finally, 0 and 1 are used as true and false judgment values to output the discrimination results of the generated samples. The Leaky Relu function is defined as the activation function of all layers, which solves the phenomenon of neuron necrosis in the original activation function, as shown in Figure 3.

3.2. Structural Improvement of the C-DCGAN Model

3.2.1. Spectral Normalization

In the generative adversarial networks, the number of parameters increases exponentially with the number of generating layers, which leads to an increase in the amplitude of parameter variation and increases the probability of gradient explosion. By adding spectral normalization to the generator and discriminator respectively [29], the gradient upper bound of the function is restricted to make the function smoother, improve the stability of parameter variation and reduce the probability of gradient explosion.
The relationship between the output and input of the nth convolution layer of C−DCGAN network presented in this paper is shown in Formula (4):
x n = D n W n x n 1
Dn refers to the diagonal matrix, and Wn is a parameter matrix.
From Formula (4), the output function in C-DCGAN is:
f ( x ) = D n W n D n 1 W n 1 D 1 W 1 X 0
The upper bound of the gradient of the function is restricted by using the Lipschitz constraint [30]. In Formula (5), the gradient of the function is:
x ( f ( x ) ) 2 = D n W n D n 1 W n 1 D 1 W 1 2 D n 2 W n 2 W 1 2
W‖ is the spectral norm of the W matrix, which can also be written as:
x ( f ( x ) ) 2 i = 1 N σ ( W i )
σ ( W ) is the maximum singular value of the W matrix.
The Formula (7) is normalized as follows:
x ( f ( x ) ) 2 = D n W n σ ( W n ) D n 1 W n 1 σ ( W n 1 ) D 1 W 1 σ ( W 1 ) 2 i = 1 N σ ( W i ) σ ( W i ) = 1
The essential process of spectral normalization is SVD layer by layer, which is solved one by one by power iteration because of time constraints and the cost of computing resources.
After the spectral normalization is added, the parameters of the generated network and the discriminated network are shown in the Table 1 and Table 2.

3.2.2. Loss Function

KL divergence or JS divergence is usually used to calculate the distance in the original GAN. However, because of the asymmetry of the KL divergence, different loss values will be obtained for the same difference value when different calculation methods are used. However, the distance calculated by JS divergence is 0 when there is no overlap part in the probability distribution, and over-training will result in a loss value that converges to a fixed value and cannot yield a gradient value.
The KL divergence and JS divergence are similar to distance W. W can be understood as the minimum cost under an optimal path planning, and it can be used to compute the distance between two probability distributions. If there is no overlap between the two probability distributions, or the ratio of overlap is small, W distance can still measure the distance between the distributions. Therefore, W distance can not only provide the distance of distribution, but also guarantee the reliability of gradient information. By defining W as the loss function of this experiment, the calculation can, to a large extent, solve the problems of training instability and gradient disappearance.
The W distance is defined as follows:
W ( P d a t a , P z ) = inf γ ( P d a t a , P z ) E ( x , y ) ~ y [ x y ]
The definition of the W distance is derived by a transformation of the given formula, as follows:
W ( P d a t a , P z ) = sup f L 1 E x ~ P d a t a [ f ( x ) ] E x ~ p z [ f ( x ) ]
It is indicated that the training of the network needs to satisfy the Lipschitz constraint condition, that is D l 1 , the amplitude of the parameter change in the training process must be restricted within a certain range.
In this paper, we modify the loss function of the network by adding a gradient penalty term [31], which satisfies the Lipschitz condition. The expression of the gradient penalty term is as follows:
H = λ E x P x ^ [ ( x D ( x ^ ) 1 ) 2 ]
The regular term coefficient represents the random sampling value between the real sample value x and the generated sample value.
According to the above contents and formulas, the loss functions of the generator and discriminator can be obtained, which are Formulas (12) and (13) respectively:
L G = E x ~ P z [ D ( G ( z | c ) | c ) ]
L D = E x ~ P z [ D ( G ( z | c ) | c ) ] E x ~ P d a t a [ D ( x | c ) ] + λ E x ~ P x ^ [ ( x D ( x ^ ) 1 ) 2 ]

3.3. Training Process of C-DCGAN

The optimization and improvement points of C-DCGAN have been described in detail in Section 2.1 and Section 2.2 of this paper. Algorithm 1 is the pseudo-code of the improved C-DCGAN algorithm.
Algorithm 1: C-DCGAN training process
Input: Noise fault data with constraints
Output: Enhanced bearing fault sample data
1. Initialization generator and discriminator.
2. while i do:
   For step do:
Sample mini batch of n noise samples from noise Pz(z)
Sample mini batch of n examples from data generating distribution Pdata(x)
Add constraints C to the generator
Noise sample input generator to obtain generated data
Update the discriminator by its stochastic gradient:
1 n i = 1 n [ log D ( x ( i ) ) + log ( 1 D ( G ( z ( i ) | c ) ) ) ]
End for
3. Sample mini batch of n noise samples from noise Pz(z)
4. Add constraints C to the discriminator
5. Update the generator by its stochastic gradient:
1 n i = 1 n log ( 1 D ( G ( z ( i ) | c ) ) )
6. end while
   I is the maximum number of iterations in the training process. Step is the training times of the discriminator.

3.4. One-Dimension Convolutional Neural Network

One-Dimension Convolutional Neural Network (1−D−CNN) is a deep feedback neural network. Since the bearing fault data is one-dimensional data, 1−D−CNN can be used for one-dimensional sequence processing and convolution operations.
Since the convolutional neural network can achieve muti-category classification effects, this paper uses 1−D−CNN to classify the faults of rolling bearings and mixes the original fault data and the generated sample data into the 1−D−CNN network for classification and diagnosis.
In this paper, the bearing fault sample data is input into multiple convolutional layers in the 1−D−CNN model, the input data is convolved through the convolutional layer, and the Relu function is used to activate and add nonlinear features to the feature pool and then use the maximum pooling layer to reduce the input data length of each layer and reduce the probability of over fitting [32]. Finally, the fault data features are processed through the fully connected layer and input to the softmax layer to calculate the probability of fault category and output classification result. The 1−D−CNN is literately trained, and the parameters of the model are literately updated according to the value of the cross-entropy loss function calculated by iteration until the end of the training.
The 1−D−CNN parameters in this paper are shown in Table 3.

3.5. Fault Diagnosis Algorithm

Because the fault diagnosis classification model needs sufficient robust data set to ensure its effectiveness, rich data categories and other characteristics will improve its performance. At the same time, it is necessary to maintain the stability of the training process of the countermeasure generation network. In order to optimize the solutions to the problems mentioned above, a fault diagnosis method based on C−DCGAN network structure is proposed in this paper, and its main flow is shown in Figure 4.

4. Experiment and Result Analysis

4.1. Data Set

For this paper, the experimental data from the rolling bearing on the inner and outer rings and the rolling body was recorded, after the use of electric spark technology on the rolling bearing had caused different degrees of single-point bearing fault. Figure 5 is the test stand; the diameters of the fault damage were 0.18 mm, 0.36 mm, 0.54 mm, and 0.71 mm, respectively. Then the vibration signals were collected under the conditions of 0HP, 1HP, 2HP and 3HP, respectively; the sampling frequency was 12 kHz.
In this experiment, faults are divided into 10 categories. In order to ensure that the statistical characteristic distribution of fault samples follows that of the distribution of a large number of overall fault characteristics, 1024 consecutive sample points are intercepted on the original vibration signal for each small number of fault samples and a large number of overall fault samples are generated, with a total of three data sets A, B and C. Data set A is mainly used to train C−DCGAN to generate high-quality fault sample data, so it mainly contains the original fault sample data of various types of faults in the original fault data set. Data set B is mainly used to train the classification model for fault classification and diagnosis. The simulated fault samples generated from data set A are mixed into the original data set A, and the mixed fault data is used as the training set. Data set C is a test set to test the training convergent classification model. The number of experimental samples is shown in Table 4.

4.2. Experimental Results and Comparative Analysis

In order to verify the feasibility of the fault diagnosis method based on the C-DCGAN extended data set, the comprehensive indexes of positive case accuracy and negative case accuracy are used to evaluate the method.
“Accuracy” is the ratio of the number of correctly classified failure samples to the total number of failure samples when testing against a failure test set.
“Recall” is the ratio of the number of failure samples from all positive cases to the number of failure samples from the actual positive cases in the correct classification.
“Specificity” refers to the ratio of all false samples to all actual false samples in the predicted accurate fault samples.
G-mean, an index obtained by combining the accuracy of positive cases and negative cases, is usually used to evaluate the classification effect when the data distribution is unbalanced.
A c c u r a c y = T P + T N T P + F N + F P + T N
R e c a l l = T P T P + F N
S p e c i f i c i t y = T N T N + F P
G m e a n = R e c a l l S p e c i f i c i t y
TP refers to the number of real samples judged as true sample categories; TN refers to the number of false samples judged as false sample categories; FP refers to the number of false samples judged as true sample categories, that is, the number of samples that are mistakenly classified; FN refers to the number of real samples judged as false sample categories, that is, the number of real samples omitted.
The accuracy and the value of loss are shown in Figure 6 and Figure 7.
Figure 8a–c shows the time domain diagrams of the original fault samples of the rolling element, inner ring and outer ring, and the corresponding time domain diagrams of the samples generated by C−DCGAN. It can be seen from Figure 8 that the fault sample data generated by C−DCGAN is not exactly the same as the original data, but its overall distribution is similar to the original sample data. At the same time, the generated data has expanded the diversity of the real sample data, which proves that the generation network can be used to effectively expand the unbalanced data set, so as to solve the problem of data imbalance in the data set.
As shown in Figure 9, the data set of original unbalanced fault and the expanded fault are used for fault diagnosis by the same classifier, and T−SNE visualization is performed. Figure 9a is the diagnostic classification result of the original unbalanced fault data set. It can be seen that there are many overlapping cases of different types of faults. Figure 9b shows the diagnostic classification results after the data set is balanced and expanded by the generated model structure in this paper. It can be seen that C−DCGAN can generate muti-category fault sample data and mix the generated simulated fault data into the original fault data set. Use of the mixed fault data set for fault diagnosis and classification can improve the clustering effect of each type of fault sample data, and effectively improve the accuracy of fault diagnosis and classification. It can be seen from the classification effect in Figure 9 that the sample values of bearing faults of the same category are gathered together, while the fault data of different categories can be clearly separated according to the fault category, and most of the sample data can be correctly classified.
According to the definition and formula, the value range of G−mean is (0, 1), and values close to 1 indicate that the classification effect of G−mean is better. According to Figure 10, the original unbalanced fault data of bearing inputted into the 1−D−CNN network for fault diagnosis and the classification result reveals that most of the values are too small, indicating that the original unbalanced small fault data cannot train 1−D−CNN to a high accuracy rate, so the classification effect is poor. CGAN performs supervised data generation under constraints. The simulated fault sample data generated after the constraint guidance is mixed into the original fault data, and input into the classifier for training. After the classifier is trained to Nash equilibrium, the classification is used. It can be seen from the G−mean value that generating valid fault data through CGAN and expanding the data set can reduce the impact of uneven distribution of data samples. The performance of C−DCGAN under ten fault categories is the strongest of the three, and the average value returned exceeds 0.8, which can prove that the network structure described in this paper can better balance the expansion of the original fault sample data set. By balancing the fault data set, the classifier is trained to achieve high-accuracy fault classification and diagnosis, reducing the impact of classification errors caused by uneven data distribution.
In order to verify the validity and authenticity of the fault sample data generated by the model proposed in this paper, the maximum mean discrepancy (MMD) is used for evaluation. This metric evaluates the true rows of the generated simulated sample data by calculating the probability distribution distance between the simulated sample data and the original sample data; the calculation formula is shown in Formula (18):
M M D ( X , Y ) = 1 m i = 1 m ϕ ( x i ) 1 n j = 1 n ϕ ( y j ) K 2
where K indicates that the distance between the original and generated data set, and which is mapped to the regenerated Hilbert space by the function of Gaussian kernel Randomly selected 0, 1, 4, and 7 categories of fault data, as shown in Figure 11, in the process of generating data for these four different fault categories, as the number of training iterations increases, the maximum mean difference overall shows a downward trend, and the probability distribution between the original fault data and the generated fault samples gradually decreases. Scaled down, this verifies the authenticity of the simulated fault samples generated by the adversarial generative network proposed in this paper.
In order to verify that the C−DCGAN fault diagnosis model proposed in this paper is suitable for fault diagnosis of bearings in the case of unbalanced sample data, the model is compared with the following two different diagnostic models of adversarial generative networks, with a 1:20 imbalance. The proportion of the fault data set is defined, and the highly unbalanced data set is input into three fault diagnosis models trained to Nash equilibrium, the trained diagnostic model is then used to classify the fault diagnosis of the test set. The classification results are classified into a confusion matrix, as shown in Figure 12. Under the premise of unbalanced sample distribution, the data generation quality of the generative model can be reflected according to the classification accuracy of fault diagnosis. In Figure 12, the fault diagnosis model proposed in this paper evinces high accuracy and can therefore be applied to bearing fault diagnosis scenarios with small sample fault data sets.
In order to further verify that the conditional deep convolutional adversarial generation network proposed in this paper can improve fault diagnosis classification accuracy by balanced expansion of the small sample fault dataset, common fault diagnostic methods, such as C−DCGAN+SVM, C−DCGAN+LSTM, C−DCGAN+1-D−CNN, infoGAN+1−D−CNN, and CGAN+1−D−CNN were adopted for a comparison of rolling bearing fault diagnosis. LSTM is a type of long short-term memory network that detects and classifies faults by learning the temporal information between fault features. C−DCGAN+SVM is a type of process in which, after expanding the small sample data set through the confrontation generation network proposed in this paper, the data set is input to the SVM network for classification training, and the trained classifier is used to perform fault diagnosis and classification on the test set. infoGAN+1-D−CNN is based on an information generation adversarial network for data generation, and it then uses one-dimensional convolutional network for fault feature extraction and fault classification. CGAN is based on GAN and adds constraints for data generation.
According to Table 5, it can be seen that the adversarial generation model proposed in this paper can effectively expand the fault data set and use a variety of classifiers for fault diagnosis, and the classification accuracy can reach 90%. The results show that the fault diagnosis method proposed in this paper can effectively improve the fault diagnosis accuracy of rolling bearings compared with several common fault diagnosis methods based on data augmentation.

5. Conclusions

In this paper, a conditional deep convolution generative adversarial network is proposed, one which can effectively expand the data set of small samples, improve the data imbalance of small samples, and improve the precision of bearing fault diagnosis. Compared with other fault diagnosis methods, the demonstrated improved loss functions and network structure can improve the stability of C-DCGAN and reduce the probability of gradient vanishing and gradient explosion. The experimental results show that the fault diagnosis model designed in this paper can effectively improve the classification accuracy of these 10 types of faults and better diagnose the more complex types of faults in the operation of mechanical equipment. However, there are still some remaining concerns. If the fault category is increased or the mechanical operation environment is disturbed, the classification accuracy of the classifier will be reduced to a certain extent, and the model training will be difficult to converge to the ideal state; Future research will aim at the construction of the fault diagnosis model in order to realize multiple fault classifications.

Author Contributions

Conceptualization, C.P. and S.Z.; methodology, C.L. and C.P.; software, C.P.; validation, C.P.; formal analysis, C.P. and C.L.; investigation, C.P.; resources, S.Z.; writing—original draft preparation, S.Z.; writing—review and editing, C.P.; visualization, S.Z.; supervision, C.L. project administration, C.P.; funding acquisition, C.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by Natural Science Foundation of China (No. 61871432), the Natural Science Foundation of Hunan Province (No. 2022JJ30231).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lei, Y.G.; Yang, B.; Du, Z.J.; Lu, N. A deep migration diagnosis method for mechanical equipment faults under big data. Chin. J. Mech. Eng. 2019, 55, 1–8. [Google Scholar]
  2. Wang, Z.Y.; Yao, L.G. Generalized fine compound muti-scalesample entropy combined with manifold learning for rolling bearing fault diagnosis method. China Mech. Eng. 2020, 31, 2463–2471. [Google Scholar]
  3. Zhang, M.; Lu, J.; Ma, J. Fault Diagnosis of Rolling Bearing Based on Multi Scale Convolution Strategy CNN. J. Chongqing Univ. Technol. (Nat. Sci.) 2020, 34, 102–110. [Google Scholar]
  4. Gao, Y.; Yu, D. Intelligent fault diagnosis for rolling bearings based on graph shift regularization with directed graphs. Adv. Eng. Inform. 2021, 47, 101253. [Google Scholar] [CrossRef]
  5. Liu, Z.; Wang, H.; Liu, J.; Qin, Y.; Peng, D. Multi-task Learning Based on Lightweight 1DCNN for Fault Diagnosis of Wheelset Bearings. IEEE Trans. Instrum. Meas. 2020, 70, 1–11. [Google Scholar] [CrossRef]
  6. Zhu, H.; Cheng, J.; Zhang, C.; Wu, J.; Shao, X. Stacked pruning sparse denoising autoencoder based intelligent fault diagnosis of rolling bearings. Appl. Soft Comput. 2020, 88, 106060. [Google Scholar] [CrossRef]
  7. Wang, Z.; Ma, H.; Chen, H.; Yan, B.; Chu, X. Performance degradation assessment of rolling bearing based on convolutional neural network and deep long-short term memory network. Int. J. Prod. Res. Taylor Fr. J. 2020, 58, 3931–3943. [Google Scholar] [CrossRef]
  8. Zhiyi, H.; Haidong, S.; Xiang, Z.; Yu, Y.; Junsheng, C. An intelligent fault diagnosis method for rotor-bearing system using small labeled infrared thermal images and enhanced CNN transferred from CAE. Adv. Eng. Inform. 2020, 46, 101150. [Google Scholar] [CrossRef]
  9. Shao, X.; Kim, C.-S. Unsupervised Domain Adaptive 1D-CNN for Fault Diagnosis of Bearing. Sensors 2022, 22, 4156. [Google Scholar] [CrossRef]
  10. Tang, H.; Gao, S.; Wang, L.; Li, X.; Li, B.; Pang, S. A Novel Intelligent Fault Diagnosis Method for Rolling Bearings Based on Wasserstein Generative Adversarial Network and Convolutional Neural Network under Unbalanced Dataset. Sensors 2021, 21, 6754. [Google Scholar] [CrossRef]
  11. Chen, Z.; Wang, Y.; Wu, J.; Deng, C.; Jiang, W. Wide Residual Relation Network-Based Intelligent Fault Diagnosis of Rotating Machines with Small Samples. Sensors 2022, 22, 4161. [Google Scholar] [CrossRef]
  12. Yuan, C.; Qian, L.; Ting, Z.H. Generation of malicious domain training data based on generative adversarial network. Appl. Res. Comput. 2019, 36, 1540–1568. [Google Scholar] [CrossRef]
  13. Deng, W.; Yao, R.; Zhao, H.; Yang, X.; Li, G. A novel intelligent diagnosis method using optimal LS-SVM with improved PSO algorithm. Soft Comput. 2019, 23, 2445–2462. [Google Scholar] [CrossRef]
  14. Ding, Y.; Ma, L.; Ma, J.; Wang, C.; Lu, C. A Generative Adversarial Network-Based Intelligent Fault Diagnosis Method for Rotating Machinery under Small Sample Size Conditions. IEEE Access 2019, 7, 149736–149749. [Google Scholar] [CrossRef]
  15. Zhang, T.; Chen, J.; Li, F.; Zhang, K.; Lv, H.; He, S.; Xu, E. Intelligent fault diagnosis of machines with small & imbalanced data: A state-of-the-art review and possible extensions. ISA Trans. 2022, 119, 152–171. [Google Scholar] [CrossRef] [PubMed]
  16. Ha, M.; Li, J.; Tian, J.; Wang, X. The key theorem of learning theory on g/sub /spl lambda//measure spaces. Proceedings of 2004 International Conference on Machine Learning and Cybernetics (IEEE Cat. No.04EX826), Shanghai, China, 26–29 August 2004; Volume 3, pp. 1904–1907. [Google Scholar] [CrossRef]
  17. Antoniou, A.; Storkey, A.; Edwards, H. Data Augmentation Generative Adversarial Networks. arXiv 2017, arXiv:1711.04340. [Google Scholar] [CrossRef]
  18. Yin, H.; Li, Z.; Zuo, J.; Liu, H.; Yang, K.; Li, F. Wasserstein Generative Adversarial Network and Convolutional Neural Network (WG-CNN) for Bearing Fault Diagnosis. Math. Probl. Eng. 2020, 1–16. [Google Scholar] [CrossRef]
  19. Lan, G.; Shen, H.F.; Wang, Y. Dataset enhancement method based on improved deep convolutional generative adversarial networks. J. Comput. Appl. 2021, 1–11. [Google Scholar] [CrossRef]
  20. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Bangkok, Thailand, 23–27 November 2020; MIT Press: Cambridge, MA, USA, 2020; pp. 2672–2680. [Google Scholar]
  21. Liu, Q.; Ma, G.; Cheng, C. Data Fusion Generative Adversarial Network for Multi-Class Imbalanced Fault Diagnosis of Rotating Machinery. IEEE Access 2020, 8, 70111–70124. [Google Scholar] [CrossRef]
  22. Meng, Z.; Guo, X.; Pan, Z.; Sun, D.; Liu, S. Data Segmentation and Augmentation Methods Based on Raw Data Using Deep Neural Networks Approach for Rotating Machinery Fault Diagnosis. IEEE Access 2019, 7, 79510–79522. [Google Scholar] [CrossRef]
  23. Xue, Z.Z.; Man, J.F.; Peng, C.; Deng, H. Research on Bearing Fault Diagnosis Based on WGAN and GAPCNN under unbalanced data. Appl. Res. Comput. 2020, 37, 3681–3685. [Google Scholar] [CrossRef]
  24. Radford, A.; Metz, L.; Chintala, S. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
  25. Gao, X.; Deng, F.; Yue, X. Data augmentation in fault diagnosis based on the Wasserstein generative adversarial network with gradient penalty. Neurocomputing 2020, 396, 487–494. [Google Scholar] [CrossRef]
  26. Mirza, M.; Osindero, S. Conditional Generative Adversarial Nets. Comput. Sci. 2014, 2672–2680. [Google Scholar] [CrossRef]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  28. Jiang, M.; Zhi, M.; Wei, L.; Yang, X.; Zhang, J.; Li, Y.; Wang, P.; Huang, J.; Yang, G. FA-GAN: Fused attentive generative adversarial networks for MRI image super-resolution. Comput. Med. Imaging Graph. 2021, 92, 101969. [Google Scholar] [CrossRef] [PubMed]
  29. Tan, H.W.; Zhou, L.Y.; Wang, G.D. Instability analysis for generative adversarial networks and its solving techniques. Sci. Sin. 2021, 51, 602–617. [Google Scholar] [CrossRef]
  30. Ming, Z.; Li, T.; Zhu, R.; Tang, Y.; Tang, M.; Lin, L.; Ma, Z. Conditional Wasserstein generative adversarial network-gradient penalty-based approach to alleviating imbalanced data classification. Inf. Sci. 2020, 512, 1009–1023. [Google Scholar] [CrossRef]
  31. Jalayer, M.; Orsenigo, C.; Vercellis, C. Fault detection and diagnosis for rotating machinery: A model based on convolutional LSTM, Fast Fourier and continuous wavelet transforms. Comput. Ind. 2021, 125, 103378. [Google Scholar] [CrossRef]
  32. Li, Y.; Zhang, Y.; Miao, Z.; Xu, Y. Convolutional encoding image retrieval method based on t-SNE. Appl. Res. Comput. 2017, 34, 1244–1248, 1264. (In Chinese) [Google Scholar]
Figure 1. Conditional Deep Convolution Generative Adversarial Networks.
Figure 1. Conditional Deep Convolution Generative Adversarial Networks.
Sensors 22 05658 g001
Figure 2. C-DCGAN generator.
Figure 2. C-DCGAN generator.
Sensors 22 05658 g002
Figure 3. C-DCGAN discriminator.
Figure 3. C-DCGAN discriminator.
Sensors 22 05658 g003
Figure 4. Overall flow chart.
Figure 4. Overall flow chart.
Sensors 22 05658 g004
Figure 5. Test stand.
Figure 5. Test stand.
Sensors 22 05658 g005
Figure 6. Accuracy.
Figure 6. Accuracy.
Sensors 22 05658 g006
Figure 7. Loss.
Figure 7. Loss.
Sensors 22 05658 g007
Figure 8. Time domain diagram of the fault.
Figure 8. Time domain diagram of the fault.
Sensors 22 05658 g008
Figure 9. Visual diagram of fault classification: (a) CWRU(original)+1-D−CNN; (b) proposed method.
Figure 9. Visual diagram of fault classification: (a) CWRU(original)+1-D−CNN; (b) proposed method.
Sensors 22 05658 g009
Figure 10. G−mean value under ten categories.
Figure 10. G−mean value under ten categories.
Sensors 22 05658 g010
Figure 11. MMD.
Figure 11. MMD.
Sensors 22 05658 g011
Figure 12. Confusion matrix: (a) CGAN; (b) C−DCGAN (without SN); and (c) C−DCGAN.
Figure 12. Confusion matrix: (a) CGAN; (b) C−DCGAN (without SN); and (c) C−DCGAN.
Sensors 22 05658 g012
Table 1. Network parameters for generator.
Table 1. Network parameters for generator.
Network LayerConvolution NucleusStep LengthActivation FunctionLearning RateSN
Input4*40Relu N
Deconv15*52Relu0.001Y
Deconv25*52Relu0.001Y
Deconv35*52Relu0.001Y
Deconv45*52Relu0.001Y
Deconv55*52Relu0.001Y
Output5*52Tanh N
Table 2. Network parameters of discriminator.
Table 2. Network parameters of discriminator.
Network LayerConvolution NucleusStep LengthActivation FunctionLearning RateSN
Input5*52Leaky Relu N
Conv15*52Leaky Relu0.001Y
Conv25*52Leaky Relu0.001Y
Conv35*52Leaky Relu0.001Y
Conv45*52Leaky Relu0.001Y
Conv55*52Leaky Relu0.001Y
Output4*40Leaky Relu N
Table 3. Network parameters for 1−D−CNN.
Table 3. Network parameters for 1−D−CNN.
Network LayerKernel CountKernel SizeStridePadding
Conv1321*911
BN
Maxpool321*520
Conv2641*511
BN
Maxpool641*520
Conv31281*511
BN
Maxpool1281*520
Conv42561*511
BN
Maxpool2561*520
Flatten
FC1
FC2
Softmax
Table 4. Number of experimental samples.
Table 4. Number of experimental samples.
Fault LocationInnerOuter BallNormal
category0123456789
diameter0.180.360.540.180.360.540.180.360.540.00
Data set A500500500500500500500500500500
Data set B1500150015001500150015001500150015001500
Data set C600600600600600600600600600600
Table 5. Comparative experimental data.
Table 5. Comparative experimental data.
ExperienceAccuracy (%)Standard Deviation
C-DCGAN+SVM92.51±0.75
C-DCGAN+LSTM97.28±0.33
C-DCGAN+1-D-CNN99.01±0.19
infoGAN+1-D-CNN98.17±0.41
CGAN+1-D-CNN97.82±0.33
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Peng, C.; Zhang, S.; Li, C. A Rolling Bearing Fault Diagnosis Based on Conditional Depth Convolution Countermeasure Generation Networks under Small Samples. Sensors 2022, 22, 5658. https://doi.org/10.3390/s22155658

AMA Style

Peng C, Zhang S, Li C. A Rolling Bearing Fault Diagnosis Based on Conditional Depth Convolution Countermeasure Generation Networks under Small Samples. Sensors. 2022; 22(15):5658. https://doi.org/10.3390/s22155658

Chicago/Turabian Style

Peng, Cheng, Shuting Zhang, and Changyun Li. 2022. "A Rolling Bearing Fault Diagnosis Based on Conditional Depth Convolution Countermeasure Generation Networks under Small Samples" Sensors 22, no. 15: 5658. https://doi.org/10.3390/s22155658

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop