Next Article in Journal
Demonstration of Coherent Interference between Acoustic Waves Using a Fiber Ring Resonator
Next Article in Special Issue
Combination of VMD Mapping MFCC and LSTM: A New Acoustic Fault Diagnosis Method of Diesel Engine
Previous Article in Journal
Analysis of Radon Measurements in Relation to Daily Seismic Activity Rates in the Vrancea Region, Romania
Previous Article in Special Issue
Metalearning-Based Fault-Tolerant Control for Skid Steering Vehicles under Actuator Fault Conditions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Wide Residual Relation Network-Based Intelligent Fault Diagnosis of Rotating Machines with Small Samples

1
School of Mechanical Science and Engineering, Huazhong University of Science and Technology, Wuhan 430074, China
2
China Electronic Product Reliability and Environmental Testing Research Institute, Guangzhou 510610, China
3
School of Naval Architecture and Ocean Engineering, Huazhong University of Science and Technology, Wuhan 430074, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(11), 4161; https://doi.org/10.3390/s22114161
Submission received: 3 May 2022 / Revised: 25 May 2022 / Accepted: 28 May 2022 / Published: 30 May 2022
(This article belongs to the Special Issue Artificial Intelligence Enhanced Health Monitoring and Diagnostics)

Abstract

:
Many existing fault diagnosis methods based on deep learning (DL) require numerous fault samples to train the diagnosis model. However, in industrial applications, rotating machines (RMs) operate in normal states for most of their service life with fault events being rare and thus failure samples are very limited. To solve the problem above, a novel wide residual relation network (WRRN) is proposed for intelligent fault diagnosis of the RMs. Specifically, the WRRN is trained by performing a series of learning tasks in RMs with sufficient samples to obtain knowledge about how to diagnose, and then it is directly transferred to realize fault task of the RM with small samples. In this method, a wide residual network-based feature extraction module is used to generate representative fault features from input samples, and a relation module is designed to calculate the relation score between the sample pairs so as to determine their categories. Extensive experiments are conducted on two RMs to validate the WRRN method. The results demonstrate that the WRRN can accurately identify the fault types of the RMs with only small samples or even one sample. The WRRN significantly outperforms the existing popular methods in diagnostic performance.

1. Introduction

As a multi-disciplinary research field, rotating machine (RM) fault diagnosis has been explored by studies from mechanical engineering, machine learning, artificial intelligence, fault tolerance schemes, and so on [1,2,3,4,5]. Many machine-learning (ML) methods have been employed and modified for fault diagnosis. The ML-based fault diagnosis methods generally include feature extraction and classifier [6,7,8]. The widely used feature extraction methods include empirical mode decomposition [9], Fourier transform [10], continuous wavelet transform (CWT) [11], and so on. The classifier-based ML contain support vector machine [12], Bayesian [13], ensemble learning [14], and so on. However, these ML-based fault diagnosis methods require manually extracted features, which cannot provide an end-to-end diagnosis.
In recent decades, intelligent fault diagnosis methods based on deep learning (DL) have become widely applied. Many DL methods for fault diagnosis in industrial machines has been received attention due to their ability to automatically extract fault features from monitoring signals and deliver reliable diagnostic results [15,16,17]. For instance, Li et al. [18] incorporated Bayesian Gaussian mixture and convolutional neural network (CNN) to perform bearing fault diagnosis. A bearing dataset and a gearbox dataset are used to test the efficiency of the proposed method. Chen et al. [19] combined continuous wavelet transform and local binary CNN to provide end–end fault diagnosis of RMs. Two experimental studies are conducted to verify the stability and reliability of the proposed method, including bearing fault diagnosis and gearbox compound fault diagnosis. Zhao et al. [20] developed a deep network based on a residual shrinkage network. A soft threshold is inserted into the network to eliminate unimportant features, resulting in improved feature learning ability from highly noised signals and high fault diagnostic accuracy. Most existing DNN-based fault diagnosis methods aim to improve diagnostic accuracy given enough fault samples, while restricted fault samples are rarely considered [21,22,23]. However, industrial applications, RMs operate under the normal state in most of their service life, whereas failure events seldom happen. Thus, it is difficult to collect sufficient fault samples to meet the training purposes of DL models.
Transfer learning (TL) can transfer the diagnostic knowledge learned from source domain to apply it to a related but new target domain for fault diagnosis [24,25,26]. Many researchers have found that the transfer diagnostic model is formed on RM operated in lab environment (RMLE) and then transferred to specific machines with small fault samples for fault diagnosis. Yang et al. [27] developed a feature-based transfer neural network (FTNN) for bearing fault diagnosis. The FTNN can learn diagnostic knowledge from other machines to diagnose the health of the machine. Shao et al. [28] used CWT to transform time-frequency maps from raw vibration signals. Then, the TL model based on DL has been built and high diagnostic accuracy obtained on three datasets including gearboxes, motors, and bearings. Guo et al. [29] proposed a transfer relation network and employed multikernel maximum mean discrepancy to improve the transfer performance. The effectiveness of the method is verified by four datasets, including three lab datasets and one practical dataset. Other transfer learning-based tasks were also investigated, such as motor fault diagnosis [30] and tool remaining useful life prediction [31]. Those TL-based fault diagnosis methods required a certain correlation between the source domain and the target domain, and the data from the target domain involved in training. However, it is hard to find an appropriate dataset as source domain. Since failure events are uncommon in real-world industrial scenarios, it is difficult to ensure that the target machine dataset has a certain number of fault samples.
For the problem of small data, few-shot learning methods have been proven to be an effective solution by many researchers [32,33]. The few learnings provide much practical value and have recently received a lot of attention from researchers in the field of computer vision. The few-shot methods are able to learn classifiers in source domains with enough labeled data and then perform a classification task on target domains with little labeled data of each class. Li et al. [34] developed a hierarchical Bayesian model to learn visual concepts with just one example. Gregory et al. [35] developed a Siamese network for one shot learning. This network used a similarity algorithm to measure the similarity between samples. Sung et al. [36] designed a relation network to calculate the relation score between the sample pairs to determine their types.
Inspired by the above-mentioned methods for few-shot learning methods, a novel wide residual relation network (WRRN) is proposed in this paper for solving the few sample problems in intelligent fault diagnosis of RMs. The method mainly includes a feature extraction module and a relation module. The wide residual network-based feature extraction module is used to generate representative fault features from input samples. The relation module calculates the relation score between the sample pairs to determine their types. The main contributions of this paper are summarized below.
  • A WRRN method is first proposed to exploit the fault knowledge learned from the lab machine for fault diagnosis in several real-case machines with small fault data, whereas only lab machine datasets are used for training.
  • The built wide residual network can generate more representative fault features from input samples compared to traditional CNN methods.
  • The relation module can reveal the similarity relations between the sample pairs to determine their categories, which can improve diagnostic performance.
The remainder of the paper is organized as follows. Section 2 describes problem definition, the proposed WRRN method and the optimization objective of the WRRN. Section 3 presented a fault diagnosis procedure based on WRRN. In Section 4 discusses the experimental results. Finally, Section 5 summarizes the proposed method.

2. Proposed Method

2.1. Problem Formulation

The dataset for the WRRN method mainly includes training dataset D t r a i n and the test dataset D t e s t , where the dataset D t r a i n is a relatively large labeled dataset from RMLE and the dataset D t e s t is only a small dataset from RM operated in real-world environments (RMRE). The purpose of the WRRN is to can utilize the transferable diagnostic model trained on the dataset D t r a i n to the test datasets D t e s t for fault diagnosis. Both training dataset and the test datasets include support set S and query set . The D t r a i n and the D t e s t can be redefined as D t r a i n = S t r a i n , t r a i n and D t e s t = S t e s t , t e s t , respectively, where the S represents labeled dataset, and the represents unlabeled dataset. The support set S = x i , y i i = 1 m consists of C health conditions, each health conditions with K labeled samples. For each episode, the total number of S samples have m = K × C . The few-shot learning method is to diagnose the health conditions of the based on the S . This setting can be called C -way K -shot diagnosis.
For each training episode, a certain number of samples are randomly selected from the D t r a i n to construct a C -way K -shot setting as follows: T = S t r a i n t r a i n where S t r a i n = x i , y i i = 1 m and   t r a i n = x j , y j j = 1 n . T = m + n is the total number of the samples in the task T . In the training process, the WRRN model F S , ; ϑ is learned on the S t r a i n labeled to minimize the predictions loss of t r a i n .
F S , ; ϑ : S t r a i n , t r a i n C t r a i n
where C t r a i n represents the health conditions of the t r a i n . θ is indicated as
ϑ = argmin θ D t r a i n S S t r a i n , t r a i n
where S · represents the similarity between the S t r a i n and the t r a i n .
For test process, the WRRN model ϑ is transferred to diagnose the test dataset D t e s t = S t e s t , t e s t , where the S t e s t represents small-labeled samples, and the t e s t needs to be diagnosed.
F S , ; ϑ : S t e s t , t e s t C t e s t
In this paper, the assumptions are given as:
  • The different RMs have the same machine health states.
  • The training dataset comes from a RMLE. The test dataset is from a RMRE, which is not required to be involved in the training process.

2.2. Wide Residual Relation Network

As illustrated in Figure 1, the WRRN consists of a CWT module W δ , feature extractor module F φ and relation module R θ . The CWT module W δ is adopted to convert time-frequency maps from raw signals. The feature extractor F φ can mine time-frequency maps to generate representative fault features. Then the features are fed into the relation module R θ . The relation module R θ calculates the similarity relations between the sample pairs to determine their categories. The WRRN is described in detail below.
CWT module: The CWT can reveal fault information at low-frequency and high-frequency information, preserving the effective signal features. In the CWT module, wavelet time-frequency maps are formed by calculating the inner product of the time-domain data x t and the wavelet basis function φ u , s t , which are expressed as
W δ u , s = 1 s x t ψ t u s d t
where s is the scale factor of CWT, and u is the time shift factor. Then, the support set S = x i , y i i = 1 m and the query set = x j , y j j = 1 n   are converted to wavelet time-frequency maps by CWT and fed to the feature extractor module.
Feature extractor: The feature extractor F φ is adopted to mine useful information and extract high-level features from the wavelet time-frequency maps. The specific structural parameters of the feature extractor F φ are shown in Table 1. The feature extractor F φ adoptes a wide residual network (WRN), which consists of four wide residual blocks and one pooling block. The operation of wide residual block is formulated as
W l x i l = F c l x i l + b l
where W l x i l is convolutional operation. c l and b l represent the convolutional kernel and bias at layer l , respectively. The operation of pooling block is defined as
P l x i l = A v g x a , b x i l
where   a and b denote the length and width of the pool window, respectively. For inputs of the support set S = x i , y i i = 1 m and the query set = x j , y j j = 1 n , the corresponding outputs of the feature extractor F φ , are described as high-level features f S = f i i = 1 m and level features f = f j j = 1 n , respectively.
To explicitly indicate the feature relations between support set and query set, support–query pairs are constructed. The support–query pairs are represented as
G m f S , f = f i , f j
where f i f S = f i i = 1 m , g j f = f j j = 1 n . · , · is the concatenation operation.
Relation module: The relation module R θ is composed of WRN and two fully connected layers. The WRN is composed of two wide residual blocks, which can mine relational features of support–query pairs. The specific structural parameters of the relation module R θ are shown in Table 1. The corresponding operation is shown in Equation (5). For inputs of the support–query pairs G m , the corresponding outputs of the WRN are described as relational features f G = f i i = 1 m . The relational features f G = f i i = 1 m are fed into connected layers to the relation score about support–query pairs. The sizes of two fully connected blocks is 1 × 8 and 1 × 1 , respectively. The output of the fully connected blocks in l is expressed as
G l x i l = f ο l x i l + ρ l
where ο l and ρ l represent the weight and bias of fully connected blocks at layer l . The relation module R θ can calculate the relation score about feature map pairs G m to preform relationship learning. The relation score r i , j is a scalar between 0 and 1 and represents the similarity between the support set and the query set. This means that the higher relation score belongs to the same category, while the lower relation score belongs to a different category. Thus, the output of the relation module R θ is defined as
r i , j = R θ ( S G n f S , f ,   i = 1 , 2 , , C

2.3. Optimization Objective of the WRRN

Considering that the WRRN is a similarity score regression task, mean square error (MSE) is adopted to calculate loss function of the WRRN. The loss function is formulated as
L M S E = i = 1 m j = 1 n ( r i , j 1 · y i = = y j ) 2
If y i and y j are in the same category, the label is 1; otherwise, the label is 0.
Suppose θ F φ and θ R θ are the parameters of the feature extractor module F φ and the relation module R θ , respectively. The (8) is rewritten as
θ F φ * , θ R θ * argmin θ F φ , θ R θ , L M S E θ F φ , θ R θ
where θ F φ * and θ R θ * are optimal parameters. The training pipeline of the WRRN in an epoch is described in Algorithm 1.
Algorithm 1. Mini-batch training algorithm for the WRRN. b and epochs denote the batch size and the number of iterations
Input: support set S t r a i n = x i , y i i = 1 m ; query set t r a i n = x j , y j j = 1 n ;
Feature extractor module F φ ; relation module R θ .
  • for i = 1 to epochs do
  •   Randomly sample K support set S t r a i n = x i , y i i = 1 m and N query set t r a i n = x j , y j j = 1 n from each category
      of D t r a i n to construct b ;
  •   Forward update L M S E θ F φ , θ R θ
  •   Backward update F φ and R θ ;
  • end for
  • return  F φ and R θ for the classification of test datasets

3. Fault Diagnosis Procedure Based on WRRN

As shown in Figure 2, the WRRN includes the training process and test process. In the training process, the WRRN model is trained on the dataset from the RMLE. In the test process, the trained WRRN model is transferred to diagnose the health conditions of the RMLE. These two processes are described below.
In the training process, fault simulation experiments are conducted on RMLE to generate fault data of different fault types. The simulated fault types by the RMLM need to include all fault types occurring in the RMRM, but do not require the same type of machine as the RMRM. The RMLE dataset is used as the training dataset D t r a i n . Next, the dataset D t r a i n are converted by CWT into wavelet time–frequency maps with a size of 28 × 28. Correspondingly, the support set S t r a i n = x i , y i i = 1 m and the query set t r a i n = x j , y j j = 1 n are constructed. The built WRRN model is trained on the S t r a i n and t r a i n . The WRRN model is trained in such way that the final loss L M S E is minimized, and the training process is completed. This trained WRRN model is directly used to diagnose the health conditions of the RMLEs.
In the test process, the test dataset D t e s t is collected from the RMRE, where the RMRE has a very small quantity of labeled samples. The test datasets D t e s t are converted by CWT into wavelet time–frequency maps with a size of 28 × 28. The support set S t e s t = x i , y i i = 1 m = K × C and the query set t e s t = x j j = 1 n are constructed from the test dataset D t e s t . The S t e s t = x i , y i i = 1 m represents the small-labeled data from the RMRE. The t e s t = x j j = 1 n represents the diagnosed data. Both sets are fed into the trained WRRN model. The WRRN model can calculate the similarity score between the t e s t and the labeled S t e s t to figure out the health conditions of the t e s t .

4. Experimental Studies

4.1. Experimental Setup and Dataset Description

The WRRN method is validated by RMs from two different fields, including a shafting machine and a steam turbine.
The shafting machine is a self-built testbed to obtain large amounts of labeled data by simulating failure experiments. The shafting machine consists of three intermediate bearings, a magnetic powder brakes, a flange, and a drive motor, as shown in Figure 3a. The shafting machine dataset has three machine health conditions: misalignment (MS), imbalance (IB), and normal (N). The shafting machine operates under five operating conditions controlled by the speed of the shaft. Variation signals are collected for 2 min at 2000 Hz. A total of 1000 samples are obtained each with 1024 data points for each health condition. Table 2 shows a detailed description of the dataset.
As shown in Figure 3b, the steam turbine consists of a speed increasing gearbox, a rotor mechanism, a coupling, a bearing based, an electric motor, and an oil pump. The rotation seep of motor is 6680 r/min, and the flow rate of the oil circuit system is 1300 L/min. The steam turbine has three health conditions, including N, IB, and MS. Variation signals are collected at 20 kHz. A total of 1000 samples each with 1024 data points is collected for each health condition. Table 2 shows a detailed description of the steam turbine dataset.
Considering that the shafting machine dataset is recorded from the RMLE, and the steam turbine dataset is from the RMRE, the transfer experiments of shafting machine to steam turbine are carried out. The shafting machine dataset is collected under five operating conditions. Thus, Table 3 lists five transfer tasks: A1, A2, A3, A4, and A5. For instance, the task A1 represents the shafting machine data from the L1 operating condition as the training dataset and the steam turbine dataset as the test dataset.

4.2. Results and Discussion

To explore the impact of the WRN as feature extractors on the diagnostic performance of the WRRN, a comparative experiment with the CNNRN method using the CNN as feature extractors is carried out. The experiment settings of 1-shot, 3-shot, 5-shot, 8-shot, and 10-shot are carried out to investigate the impact of the WR as feature extractors on the diagnostic performance of the WRRN method. For instance, 1-shot represents that one sample from each health state in the shafting machine is taken as the support set S t r a i n for the training process. For the test process, the 1-shot indicates that only one sample from each health state in the steam turbine is labeled and taken as the support set S t e s t for test process. Taking task A1 as an example, the impact of the fault sample size on the diagnostic performance of the WRRN method is explored. For each experiment setting, the training strategy of the WRRN method follows the usual zero-shot learning way by episode-based training. These two methods adopt Adam with an initial learning rate 10 3 and half annealing every 1000 sets for end-to-end training.
Figure 4 displays the diagnostic mean accuracies and standard deviations (Std) of the WRRN method and the CNNRN method, where both methods are conducted in ten trials. The results reveal that the diagnostic accuracies of the WRRN method are significantly higher than those of the CNNRN method by at least 5% in the five diagnostic tasks. From the Std perspective, it is also seen that the diagnostic performance of the WRRN method is significantly more stable than that of the CNNRN method. This can prove that the WRN has superior feature extraction capability, which enables the relation module to better discriminate the similarity relations between sample pairs, so as to improve the diagnostic performance of the WRRN. Furthermore, as the fault sample size from the steam turbine increases, the diagnostic accuracy improves. The WRRN method achieves almost 100% diagnostic accuracy, and the lowest Std in task 10-shot. The conclusion from the results is that increasing the fault sample size significantly improves the diagnostic performance of the WRRN method.
Figure 5 shows the Pareto charts of diagnostic performance of the WRRN method for five settings under task A1. Each plot in Figure 5 represents the number of misidentified testing samples in all health conditions from largest to smallest, and the cumulative frequency of misidentified samples. The results also show that most of the fault samples with incorrect diagnoses are related to the inner race fault samples, while the misdiagnosis rate between the health samples and the outer race fault samples is much smaller. This implies that the WRRN method can accurately diagnose the health state and fault state, which is a critical need in real-world engineering applications. These results demonstrate that the WRRN trained on the shafting machine can be directly transferred to the steam turbine for fault diagnosis and achieve superior diagnosis performance.
It can be seen from Figure 4 that the diagnostic accuracy of the WRRN method is relatively low on the 1-shot and 3-shot settings, especially the diagnostic accuracy of the WRRN method on the 1-shot is only 87.1%. Therefore, it is explored to improve the diagnostic performance of the WRRN method by increasing the training data size when the fault samples of the steam turbine are relatively small. Figure 6 displays the diagnostic mean accuracy and standard deviations (Std) for the different settings. It can be shown from Figure 6 that as the training dataset size increases, the diagnostic mean accuracy increases, and the Std decreases accordingly. The WRRN method achieves almost 100% diagnostic accuracy and the lowest Std in task A5. It can be concluded that the diagnostic performance of the WRRN method for the steam turbine can be improved by increasing the training data size when there is only one sample of each fault type from the steam turbine. In addition, it can be found that the diagnostic accuracy of the WRRN method under 10-shot setting is as high as 100% when the task A1. It can also be concluded that when there is a certain amount of each fault type from the steam turbine, the WRRN method can also achieve effective diagnostic performance.
Figure 7 displays the distribution of similar scores between each health state. It can be concluded from the Figure 7 that the similarity scores of each category increase gradually with the increase of the training data size. The similarity scores for each category in task A5 are concentrated at 0.9. This verifies that the diagnostic accuracy of the WRRN method is as high as 100% in task A5. Thus, the diagnostic performance of the WRRN method can be improved by increasing the amount of training data from the shafting machine when fault samples from the steam turbine are severely insufficient.
Table 4 displays the classification time for each sample under the different setting. All experiment methods are performed on a NVIDIA GeForce GTX 1660, a computer (Intel Core (TM) 3.6 GHz processor with 8 GB of RAM), and a windows version of the PyTorch platform. As can be seen from Table 4, as the number of supporting samples increases, the classification time for each sample increases accordingly. If the diagnosis task requires fast classification time, the number of supporting samples can be reduced, and the training dataset can be increased to ensure effective fault diagnosis performance.

4.3. Comparative Analysis

To show the superiority of the WRRN method even more clearly, several advanced methods for comparison, including DL methods, TL methods, and few-shot learning methods, are used to demonstrate the superiority of the WRRN method. The DL adopts the WRN as feature extractor like the WRRN method. For the WRRN method, the training dataset includes a shafting machine dataset and a small amount of fault data from the steam turbine. For TL methods, VGG-11 [37] and Resnet-18 [38] are used as backbone networks for knowledge transfer. The two networks are pre-trained on the data from the shafting machine, and then fine-tuned on small data from the RM operating real-world experiment. For the few-shot learning methods, the matching network [39] is employed in the comparative experiments, using the same feature extractors as the WRRN method. For a fair comparison, the C -way K -shot setting is used for all the above comparing methods.
The task A5 is chosen as a comparison experiment because the fault data from the shafting machine could be simulated. The Figure 8 shows the fault diagnosis performance of different methods in the steam turbine with different fault sample sizes. Similarly, each method is tested in ten trials, and the diagnostic accuracies and standard errors are obtained for six methods. It can be found from the Figure 8 that the diagnostic accuracy of all methods increases with the increase of the fault samples from the steam turbine. The diagnostic performance of these few-shot learning approaches is significantly better than that of these DL and TL methods. In terms of diagnostic performance, the WRRN method beats all other methods, with diagnostic accuracy approaching 100% in a variety of fault sample sizes. This is because the wide residual network can generate more representative fault features from input samples. Then, RM can reveal the similarity relations between the features pairs to determine their categories, which can improve diagnostic performance. These findings further show that the WRRN method can accurately diagnose health conditions of the steam turbine when just few fault samples or even only one fault sample is provided.

5. Conclusions

In this paper, a novel WRRN is proposed to diagnose the health conditions of RM with insufficient fault data. Specifically, the WRRN is trained by performing a series of learning tasks in RMs with sufficient samples to obtain knowledge about how to diagnose, and then, it is directly transferred to diagnose the RM with small samples. The method mainly includes the feature extraction module and the relation module. The wide residual network-based feature extraction module is used to generate representative fault features from input samples. The relation module calculates the relation score between the sample pairs so as to determine their health states. Extensive experiments are conducted on two RMs to validate the WRRN method. The results show that the WRRN model trained on a RMLE can properly diagnose the health conditions of RMRE with just a few fault samples, where the RMRE and the RMLE can come from different machine domains. Furthermore, the impact of RMRE fault sample size and training dataset size on diagnostic performance have been investigated. The results show that increasing the RMRM fault sample size or the RMLE training dataset size can improve the diagnostic performance of the WRRN significantly. Finally, the comparative experiments demonstrate that the WRRN outperforms state-of-the-art methods for fault diagnosis in RMs with very limited fault data circumstances.

Author Contributions

Methodology: Z.C.; algorithm design: Z.C. and J.W.; validation: J.W., C.D. and W.J.; formal analysis: Z.C., C.D. and Y.W.; writing—original draft preparation: Z.C., J.W. and C.D.; writing—review and editing: J.W. and C.D.; visualization: Z.C., J.W. and W.J.; supervision: J.W. and Y.W.; project administration: J.W. and C.D.; funding acquisition: J.W., C.D. and Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

The work was supported in part by the National Natural Science Foundation of China under Grant No. 51875225, in part by the Ministry of Industry and Information Technology of China under Grant No. TC210804R-1, and in part by Hubei Provincial Natural Science Foundation for Innovation Groups under Grant No. 2021CFA026.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, G.Q.; Deng, C.; Wu, J.; Xu, X.B.; Shao, X.Y.; Wang, Y.H. Sensor data-driven bearing fault diagnosis based on deep convolutional neural networks and s-Transform. Sensors 2019, 19, 2750. [Google Scholar] [CrossRef] [Green Version]
  2. Wu, H.; Ma, X.; Wen, C.L. Multilevel fine fault diagnosis method for motors based on feature extraction of fractional fourier transform. Sensors 2022, 22, 1310. [Google Scholar] [CrossRef] [PubMed]
  3. Lang, X.; Steven, C.; Paolo, P. Rolling element bearing diagnosis based on singular value decomposition and composite squared envelope spectrum. Mech. Syst. Signal Process. 2021, 148, 107174. [Google Scholar]
  4. Kanizo, Y.; Rottenstreich, O.; Segall, I.; Yallouz, J. Designing optimal middlebox recovery schemes with performance guarantees. IEEE J. Sel. Area Commun. 2018, 36, 2373–2383. [Google Scholar] [CrossRef]
  5. Kim, D.; Nelson, J.; Ports, D.R.K.; Sekar, V.; Seshan, S. RedPlane: Enabling fault-tolerant stateful in-switch applications. In Proceedings of the 2021 ACM SIGCOMM 2021 Conference, Virtual Event, 23–27 August 2021; pp. 223–244. [Google Scholar]
  6. Lei, Y.G.; Yang, B.; Jiang, X.W.; Jia, F.; Li, N.P.; Nandi, A.K. Applications of machine learning to machine fault diagnosis: A review and roadmap. Mech. Syst. Signal Process. 2020, 138, 106587. [Google Scholar] [CrossRef]
  7. Cheng, Y.W.; Zhu, H.P.; Wu, J.; Shao, X.Y. Machine health monitoring using adaptive kernel spectral clustering and deep long short-term memory recurrent neural networks. IEEE Trans. Ind. Inform. 2019, 15, 987–997. [Google Scholar] [CrossRef]
  8. Stetco, A.; Dinmohammadi, F.; Zhao, X.Y.; Robu, V.; Flynn, D.; Barnes, M.; Keane, J.; Nenadic, G. Machine learning methods for wind turbine condition monitoring: A review. Renew. Energy 2019, 133, 620–635. [Google Scholar] [CrossRef]
  9. Wu, J.; Wu, C.Y.; Cao, S.; Or, S.W.; Deng, C.; Shao, X.Y. Degradation data-driven time-to-failure prognostics approach for rolling element bearings in electrical machines. IEEE Trans. Ind. Electron. 2019, 66, 529–539. [Google Scholar] [CrossRef]
  10. Wang, Y.X.; Xiang, J.W.; Markert, R.; Liang, M. Spectral kurtosis for fault detection, diagnosis and prognostics of rotating machines: A review with applications. Mech. Syst. Signal Process. 2016, 66–67, 679–698. [Google Scholar] [CrossRef]
  11. Chen, Z.; Wang, Y.; Wu, J.; Deng, C.; Hu, K. Sensor data-driven structural damage detection based on deep convolutional neural networks and continuous wavelet transform. Appl. Intell. 2021, 51, 5598–5609. [Google Scholar] [CrossRef]
  12. Wu, J.; Guo, P.; Cheng, Y.; Zhu, H.; Wang, X.-B.; Shao, X. Ensemble generalized multiclass support-vector-machine-based health evaluation of complex degradation systems. IEEE-ASME Trans. Mechatron. 2020, 25, 2230–2240. [Google Scholar] [CrossRef]
  13. Jiang, Q.C.; Yan, X.F.; Huang, B.A. Performance-driven distributed PCA process monitoring based on fault-relevant variable selection and bayesian inference. IEEE Trans. Ind. Electron. 2016, 63, 377–386. [Google Scholar] [CrossRef]
  14. Zheng, J.D.; Pan, H.Y.; Cheng, J.S. Rolling bearing fault detection and diagnosis based on composite multiscale fuzzy entropy and ensemble support vector machines. Mech. Syst. Signal Process. 2017, 85, 746–759. [Google Scholar] [CrossRef]
  15. Li, G.Q.; Wu, J.; Deng, C.; Xu, X.B.; Shao, X.Y. Deep reinforcement learning-based online Domain adaptation method for fault diagnosis of rotating machinery. IEEE-ASME Trans. Mechatron. 2021. [Google Scholar] [CrossRef]
  16. Pang, X.Y.; Xue, X.Y.; Jiang, W.W.; Lu, K.B. An Investigation into fault diagnosis of planetary gearboxes using a bispectrum convolutional neural network. IEEE-ASME Trans. Mechatron. 2021, 26, 2027–2037. [Google Scholar] [CrossRef]
  17. Cheng, Y.W.; Wu, J.; Zhu, H.P.; Or, S.W.; Shao, X.Y. Remaining useful life prognosis based on ensemble long short-term memory neural network. IEEE Trans. Instrum. Meas. 2021, 70, 3503912. [Google Scholar] [CrossRef]
  18. Li, G.Q.; Wu, J.; Deng, C.; Chen, Z.Y.; Shao, X.Y. Convolutional neural network-based bayesian gaussian mixture for intelligent fault diagnosis of rotating machinery. IEEE Trans. Instrum. Meas. 2021, 70, 3517410. [Google Scholar] [CrossRef]
  19. Cheng, Y.W.; Lin, M.X.; Wu, J.; Zhu, H.P.; Shao, X.Y. Intelligent fault diagnosis of rotating machinery based on continuous wavelet transform-local binary convolutional neural network. Knowl.-Based Syst. 2021, 216, 106796. [Google Scholar] [CrossRef]
  20. Zhao, M.H.; Zhong, S.S.; Fu, X.Y.; Tang, B.P.; Pecht, M. Deep residual shrinkage networks for fault diagnosis. IEEE Trans. Ind. Inform. 2020, 16, 4681–4690. [Google Scholar] [CrossRef]
  21. Chen, Z.; Wu, J.; Deng, C.; Wang, C.; Wang, Y. Residual deep subdomain adaptation network: A new method for intelligent fault diagnosis of bearings across multiple domains. Mech. Mach. Theory 2022, 169, 104635. [Google Scholar] [CrossRef]
  22. Li, X.; Zhang, W.; Xu, N.X.; Ding, Q. Deep learning-based machinery fault diagnostics with domain adaptation across sensors at different places. IEEE Trans. Ind. Electron. 2020, 67, 6785–6794. [Google Scholar] [CrossRef]
  23. Wu, Z.H.; Jiang, H.K.; Zhao, K.; Li, X.Q. An adaptive deep transfer learning method for bearing fault diagnosis. Measurement 2020, 151, 107227. [Google Scholar] [CrossRef]
  24. Li, X.; Zhang, W.; Ding, Q.; Li, X. Diagnosing rotating machines with weakly supervised data using deep transfer learning. IEEE Trans. Ind. Inform. 2020, 16, 1688–1697. [Google Scholar] [CrossRef]
  25. Xu, G.W.; Liu, M.; Jiang, Z.F.; Shen, W.M.; Huang, C.X. Online fault diagnosis method based on transfer convolutional neural networks. IEEE Trans. Instrum. Meas. 2020, 69, 509–520. [Google Scholar] [CrossRef]
  26. Wen, L.; Gao, L.; Li, X.Y. A new deep transfer learning based on sparse auto-encoder for fault diagnosis. IEEE Trans. Syst. Man Cybern. Syst. 2019, 49, 136–144. [Google Scholar] [CrossRef]
  27. Yang, B.; Lei, Y.G.; Jia, F.; Xing, S.B. An intelligent fault diagnosis approach based on transfer learning from laboratory bearings to locomotive bearings. Mech. Syst. Signal Process. 2019, 122, 692–706. [Google Scholar] [CrossRef]
  28. Shao, S.Y.; McAleer, S.; Yan, R.Q.; Baldi, P. Highly Accurate Machine Fault Diagnosis Using Deep Transfer Learning. IEEE Trans. Ind. Inform. 2019, 15, 2446–2455. [Google Scholar] [CrossRef]
  29. Guo, L.; Lei, Y.G.; Xing, S.B.; Yan, T.; Li, N.P. Deep convolutional transfer learning network: A new method for intelligent fault diagnosis of machines with unlabeled data. IEEE Trans. Ind. Electron. 2019, 66, 7316–7325. [Google Scholar] [CrossRef]
  30. Xiao, D.Y.; Huang, Y.X.; Zhao, L.J.; Qin, C.J.; Shi, H.T.; Liu, C.L. Domain adaptive motor fault diagnosis using deep transfer learning. IEEE Access 2019, 7, 80937–80949. [Google Scholar] [CrossRef]
  31. Sun, C.; Ma, M.; Zhao, Z.B.; Tian, S.H.; Yan, R.Q.; Chen, X.F. Deep transfer learning based on sparse autoencoder for remaining useful life prediction of tool in manufacturing. IEEE Trans. Ind. Inform. 2019, 15, 2416–2425. [Google Scholar] [CrossRef]
  32. Li, L.J.; Han, J.W.; Yao, X.W.; Cheng, G.; Guo, L. DLA-MatchNet for few-shot remote sensing image scene classification. IEEE Trans. Geosci. Remote 2021, 59, 7844–7853. [Google Scholar] [CrossRef]
  33. Li, Y.; Chao, X.W. Semi-supervised few-shot learning approach for plant diseases recognition. Plant Methods 2021, 17, 68. [Google Scholar] [CrossRef] [PubMed]
  34. Li, F.F.; Fergus, R.; Perona, P. One-shot learning of object categories. IEEE Trans. Pattern Anal. 2006, 28, 594–611. [Google Scholar]
  35. Gregory, K.; Zemel, R.; Salakhutdinov, R. Siamese Neural Networks for One-Shot Image Recognition. Master’s Thesis, University of Toronto, Toronto, ON, Canada, 2015. [Google Scholar]
  36. Sung, F.; Yang, Y.X.; Zhang, L.; Xiang, T.; Torr, P.H.S.; Hospedales, T.M. Learning to compare: Relation network for few-shot learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 1199–1208. [Google Scholar]
  37. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  38. He, K.M.; Zhang, X.Y.; Ren, S.Q.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, San Juan, PR, USA, 17–19 June 2016; pp. 770–778. [Google Scholar]
  39. Vinyals, O.; Blundell, C.; Lillicrap, T.; Kavukcuoglu, K.; Wierstra, D. Matching networks for one shot learning. In Proceedings of the Advances in Neural Information Processing Systems 29, Barcelona, Spain, 5–10 December 2016. [Google Scholar]
Figure 1. Structure illustration of the proposed TRPGN model.
Figure 1. Structure illustration of the proposed TRPGN model.
Sensors 22 04161 g001
Figure 2. Fault diagnosis pipeline based on the WRRN method.
Figure 2. Fault diagnosis pipeline based on the WRRN method.
Sensors 22 04161 g002
Figure 3. Test bench for RMs: (a) the shafting machine operated in lab environment; (b) the steam turbine operated in real-world environments.
Figure 3. Test bench for RMs: (a) the shafting machine operated in lab environment; (b) the steam turbine operated in real-world environments.
Sensors 22 04161 g003
Figure 4. Diagnostic performance comparison on methods CNNRN and WRRN.
Figure 4. Diagnostic performance comparison on methods CNNRN and WRRN.
Sensors 22 04161 g004
Figure 5. Diagnostic performance of WRRN method for five settings under task A1.
Figure 5. Diagnostic performance of WRRN method for five settings under task A1.
Sensors 22 04161 g005
Figure 6. Diagnostic performance of the WRRN method under different training dataset size for the different setting.
Figure 6. Diagnostic performance of the WRRN method under different training dataset size for the different setting.
Sensors 22 04161 g006
Figure 7. Distribution of similar scores between each health state.
Figure 7. Distribution of similar scores between each health state.
Sensors 22 04161 g007
Figure 8. Accuracy comparison results of different methods on task A5.
Figure 8. Accuracy comparison results of different methods on task A5.
Sensors 22 04161 g008
Table 1. Architecture of feature extractor and relation module.
Table 1. Architecture of feature extractor and relation module.
ModuleGroup NameBlock Type = B (3,3)
Feature extractor
F φ
Conv_1 3 × 3 ,   16
Conv_2 3 × 3 , 16 × 2 3 × 3 , 16 × 2 × N
Conv_3 3 × 3 , 16 × 3 3 × 3 , 16 × 3 × N
Conv_4 3 × 3 , 16 × 4 3 × 3 , 16 × 4 × N
Avg-pool 5 × 5
Relation module
R θ
Conv_1 3 × 3 , 16 × 4 3 × 3 , 16 × 4 × N
Conv_2 3 × 3 , 16 × 4 3 × 3 , 16 × 4 × N
Avg-pool 1 × 1
FC 18
FC 21
Table 2. Introduction to datasets of lab machines and real-case machines.
Table 2. Introduction to datasets of lab machines and real-case machines.
DatasetsHealth StateOperating ConditionsNumber of Samples
Shafting machineN
I
B
MS
L1–200 r/min3 × 1000
L2–250 r/min3 × 1000
L3–300 r/min3 × 1000
L4–350 r/min3 × 1000
L5–400 r/min3 × 1000
Steam turbineN
IB
MS
6680 r/min
1300 L/min
3 × 1000
Table 3. Description of fault diagnosis task.
Table 3. Description of fault diagnosis task.
TaskTraining Dataset from Shafting MachineTesting Dataset
A1L1Steam turbine
A2L1, L2Steam turbine
A3L1, L2, L3Steam turbine
A4L1, L2, L3, L4Steam turbine
A5L1, L2, L3, L4, L5Steam turbine
Table 4. Classification time for each sample under the different setting.
Table 4. Classification time for each sample under the different setting.
TaskClassification Time (ms)
1-shot4.1
3-shot19.5
5-shot44.5
8-shot64.25
10-shot90.25
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, Z.; Wang, Y.; Wu, J.; Deng, C.; Jiang, W. Wide Residual Relation Network-Based Intelligent Fault Diagnosis of Rotating Machines with Small Samples. Sensors 2022, 22, 4161. https://doi.org/10.3390/s22114161

AMA Style

Chen Z, Wang Y, Wu J, Deng C, Jiang W. Wide Residual Relation Network-Based Intelligent Fault Diagnosis of Rotating Machines with Small Samples. Sensors. 2022; 22(11):4161. https://doi.org/10.3390/s22114161

Chicago/Turabian Style

Chen, Zuoyi, Yuanhang Wang, Jun Wu, Chao Deng, and Weixiong Jiang. 2022. "Wide Residual Relation Network-Based Intelligent Fault Diagnosis of Rotating Machines with Small Samples" Sensors 22, no. 11: 4161. https://doi.org/10.3390/s22114161

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop