Next Article in Journal
Discomfort Glare Perception by Drivers—Establishing a Link between Subjective and Psychophysiological Assessment
Next Article in Special Issue
Extension of DBSCAN in Online Clustering: An Approach Based on Three-Layer Granular Models
Previous Article in Journal
Simplify Belief Propagation and Variation Expectation Maximization for Distributed Cooperative Localization
Previous Article in Special Issue
Finite-Time Set Reachability of Probabilistic Boolean Multiplex Control Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Hybrid Model for the Prediction and Classification of Rolling Bearing Condition

1
Key Laboratory of Intelligent Control and Optimization for Industrial Equipment of Ministry of Education, Dalian University of Technology, Dalian 116024, China
2
School of Control Science and Engineering, Dalian University of Technology, Dalian 116024, China
3
Army Academy of Armored Forces, Changchun 130000, China
4
College of Chemical Process Automation, Shenyang University of Technology, Shenyang 110000, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(8), 3854; https://doi.org/10.3390/app12083854
Submission received: 21 February 2022 / Revised: 1 April 2022 / Accepted: 6 April 2022 / Published: 11 April 2022
(This article belongs to the Special Issue Statistical Learning: Technologies and Industrial Applications)

Abstract

:
Rotating machinery is a key piece of equipment for tremendous engineering operations. Vibration analysis is a powerful tool for monitoring the condition of rotating machinery. Furthermore, vibration signals have the characteristics of time series. Hence, it is necessary to monitor the condition of vibration signal series to avoid any catastrophic failure. To this end, this paper proposes an effective condition monitoring strategy under a hybrid method framework. First, we add variational mode decomposition (VMD) to preprocess the data points listed in a time order into a subseries, namely intrinsic mode functions (IMFs). Then the framework of the hybrid prediction model, namely the autoregressive moving average (ARMA)-artificial neural network (ANN), is adopted to forecast the IMF series. Next, we select the sensitive modes that contain the prime information of the original signal and that can imply the condition of the machinery. Subsequently, we apply the support vector machine (SVM) classification model to identify the multiple condition patterns based on the multi-domain features extracted from sensitive modes. Finally, the vibration signals from the Case Western Reserve University (CWRU) laboratory are utilized to verify the effectiveness of our proposed method. The comparison results demonstrate advantages in prediction and condition monitoring.

1. Introduction

In practical operation, machinery control technology [1,2,3,4], fault diagnosis and condition monitoring [5,6] have attracted wide research. Vibration analysis is a widely used tool for monitoring the condition of rotating machinery in industrial operation [7,8].
There is abundant information in the time or frequency domain reflecting the characteristics of a machine’s condition; however, some information is not suitable to apply directly. The signal decomposition method can decompose the original signal into a subseries to further comprehensively analyze the signal. Therefore, many researchers have been attracted to the field. A considerable number of techniques have been extensively carried out to fix this issue, such as wavelet decomposition (WT) [9,10]; however, wavelet decomposition aims to find wavelet basis functions, which generates a non-adaptive problem. This is the reason that the wavelet basis function is selected empirically. Different from the WT, empirical mode decomposition (EMD) was presented by Huang et al. [11,12]. Based on the inherent properties of the signal itself, the original signal was empirically decomposed by EMD. Therefore, the data from nonlinear and non-stationary processes can be decomposed by EMD, while a physically meaningful subseries can be obtained based on its adaptive characteristics [13]. Regrettably, the endpoint effect in EMD is still unresolved. Ensemble empirical mode decomposition (EEMD) was introduced by Wu and Huang [14] to alleviate the mode mixing problem occurred in EMD [15], which was a typical noise-assisted data analysis method and also has been widely applied in time-frequency analysis. Furthermore, Wang et al. [16] validated the superiority of EEMD over EMD. Both EMD and EEMD exhibit tremendous computational load and are not suitable for decomposing signals that contain large amounts of data based on their recursive decomposition [17]. In 2013, Dragomiretskiy and Zosso proposed a new algorithm for decomposing signals adaptively called variational mode decomposition (VMD) [18]. VMD is a fully non-recursive method with a more solid mathematical theoretical foundation [19]. Several studies have verified that VMD has better performance over WT, EMD, and EEMD in terms of time-frequency analysis [20,21,22]. Thus, the original signal is decomposed into several sub-series through VMD, which is targeted to enhance the accuracy and convenience of the downstream feature extraction and classification processing.
The vibration signal has the characteristics of a typical time series. It has been found that time series model-based techniques play a pivotal role in monitoring the condition of machinery based on vibration signals [23]. Therefore, an effective time series forecasting model is the key to monitoring the condition of industrial equipment, which can uncover potential statistical patterns through exploring the functional relationships. Furthermore, such patterns can give a valuable early warning. However, it is a challenge to acquire reliable information in practical operations. For time-series analysis and modeling, ARMA is one of the most popular approaches, which has made considerable achievements in various fields in terms of forecasting, such as finance [24,25], engineering [26], and many others [27,28]. In addition, ARMA is more accurate than some other popular machine learning methods in prediction, such as multi-layer perceptron, SVM, and long short-term memory [29,30]. Although ARMA has made great progress in coping with linear analyses based on the assumption that a linear functional relationship exists for current and past values and white noise, its limitations in nonlinear analysis hinder its application. Considerable research on ARMA has demonstrated that its performance is poor in modeling nonlinear real-world problems [31]. For nonlinear time series modeling, ANN is one of the most widely used algorithms [32,33], while the ANN model alone can not give full consideration to both linear and nonlinear patterns simultaneously [31,34]. In summary, any single model can not perform well in every situation. Consequently, considering the merits and drawbacks of ARMA in linear flexibility and ANN for nonlinear time series modeling, their synthetic prediction model, namely, ARMA-ANN, is constructed as a prediction model for subseries in this paper. The motivation for applying the hybrid model is demonstrated as follows [34]. First, instability and uncertainty are common in engineering applications. Second, the time series demonstrates complex characteristics in the real world. Third, each type of individual model rarely perform well compared to the hybrid technique, which combines the advantages of each type of model into an integration. Therefore, using the hybrid technique improves forecasting performance.
Various measurement indicators were considered in vibration signal analysis and fault feature extraction [35,36]. To comprehensively describe the condition information of the original signal, feature extraction plays an important role. Thus, the relevant studies have been widely studied [37,38]. The multi-domain features extracted from the sensitive modes contain the main information of the original signal. The motivation for selecting sensitive IMFs is to avoid the interference of redundant information [39]. SVM is a promising method that can solve the classification problem of small samples [40,41]. Therefore, SVM has been extensively used in mechanical fault classification based on its merits [39,42].
This work proposes an effective hybrid condition monitoring strategy that achieves state-of-the-art performance, which makes predictive maintenance come true [43]. First, we apply VMD to decompose the original signals into several subseries and predict the IMFs by ARMA-ANN. Then, the time-frequency domain (T-F) features are extracted from the sensitive IMFs. Finally, a SVM classifier is used to identify the condition patterns of the rolling bearing based on the T-F features. The performance of our method is verified through experimental signals from the CWRU Laboratory.
Our work mainly provides the following contributions:
First, the key information can be sufficiently captured through VMD decomposing the original signal into subseries to improve the accuracy of feature extraction and classification. Then, the subseries are fed into the hybrid prediction model to provide a reliable trend of the original signal. Next, the excellent accuracy for pattern identification of rolling bearing is acquired based on the comprehensive T-F features. Finally, a sufficient comparable experimental analysis verifies that our method achieved excellent performance while simultaneously relieving time resource problems.
The rest of this study is organized as follows: Section 2 presents the theoretical background. Section 3 introduces the proposed framework in detail. Experimental results and comparisons are presented in Section 4, and conclusions are given in Section 5.

2. Theoretical Foundation

2.1. Variational Model Decomposition

VMD is an adaptive, completely non-recursive variational mode decomposition model. The core idea of VMD is to determine the decomposed K IMFs, their corresponding center frequencies, and bandwidth by iteratively searching the optimal solution of the variational model. The constraint model formula of VMD is as follows:
min u k , ω k k = 1 K t δ ( t ) + j π t * u k ( t ) e j ω k t 2 2 s . t . k = 1 K u k t = x t ,
where x t is the original signal and u k : = u 1 , u 2 , u K and ω k : = ω 1 , ω 2 , ω K represent the set of all modes and the corresponding center frequencies, respectively. δ ( t ) is the Dirac distribution, t is the partial with respect to t, and K is the number of modes to be decomposed (positive integer). To solve the optimal solution of the constraint model, the augmented Lagrange function method is introduced to transform the constraint problem into an unconstrained problem, and its formula is as follows:
L u k , ω k , λ : = α k = 1 K t δ ( t ) + j π t * u k ( t ) e j ω k t 2 2 + x ( t ) k = 1 K u k ( t ) 2 2 + λ ( t ) , x ( t ) k = 1 K u k ( t ) ,
where α is the quadratic penalty term and λ t is the Lagrange multipliers. The alternate direction method of multipliers is adopted to settle the optimal u k , ω k and λ t , which can be written as:
u ^ k n + 1 f ^ ( ω ) i < k u ^ i n + 1 ( ω ) i > k u ^ i n ( ω ) + λ ^ n ( ω ) 2 1 + 2 α ( ω ω k n ) 2 ,
ω k n + 1 0 ω u ^ k n + 1 ( ω ) 2 d ω 0 u ^ k n + 1 ( ω ) 2 d ω ,
λ ^ n + 1 ( ω ) λ ^ n ( ω ) + τ x ^ ω k u ^ k n + 1 ( ω ) .
where u ^ ω , ω ^ and λ ^ ω denote the Fourier transforms of each variable, and n is the number of iteration. Ref. [18] provided more details on the VMD algorithm.

2.2. Prediction Model

Two popular forecasting models, ARMA and ANN, will be introduced in this section, which also will be used to predict the IMFs in our research.

2.2.1. Autoregressive Moving Average Model

The ARMA model is designed for the variables for which past observations imply a linear function with the future value. ARMA is first used to predict the IMF series. Because the response of IMF Y l t at t is related to the response of its previous time Y t 1 , Y t 2 , , Y t p , it is also affected by q random errors, namely ε t 1 , ε t 2 , , ε t q . For the ARMA p , q model,
Y ^ l t = β 1 Y t 1 + β 2 Y t 2 + + β p Y t p + ε t θ 1 ε t 1 θ 2 ε t 2 + + θ q ε t q ,
provides high accuracy prediction, where β p p = 1 , 2 , and θ q q = 1 , 2 , are the auto-regression (AR) and moving average (MA) coefficients, respectively. p, q are the orders of the ARMA model, which are identified based on Akaike’s information criterion and Bayesian information criterion in this paper. When q = 0 , the model is reduced to an AR p model. If p = 0 , the model is reduced to a MA q model. Note that E ε t = 0 , V a r ε t = σ ε 2 .

2.2.2. Artificial Neural Network Model

ANN is a tool which can detect all possible interactions between the input and output from the provided training data. It also provides a flexible architecture for nonlinear modeling in fault detection, diagnosis [44], and prediction [45]. The framework of ANN depends on the characteristics of the data. Thus, the number layers and neurons are configured based on the essence of the input data. The widely used three-layer ANN model presented in Figure 1 is applied to model time series data, in which the neurons are acyclically linked.
The nonlinear function f of Y n t sequence from Y n t 1 to Y n t N is expressed as follows:
Y ^ n t = ω 0 + j = 1 H ω j f ω 0 j + i = 1 N ω i j Y n t i + e t ,
where Y ^ n t is the prediction result at any given time t, ω i j and H denote the hidden layer neuron weights and number, respectively. The corresponding input layers are ω j and N. f represents a “sigmoid” function in this research, while e t is a noise or error term. Additionally, to avoid the local minimum and overfitting, k-fold cross validation is applied to train the ANN. Finally, we evaluate the test data with the trained ANN [46].

2.2.3. Hybrid Prediction Model

We take the original time series data Y t as consisting of two components, the linear components Y l t and the nonlinear components Y n t . The residual data r t = Y t Y l t are considered as the distinction between the linear component and original data. Thus, the hybrid model is constructed as:
Y ^ l t = β 1 Y t 1 + β 2 Y t 2 + + β p Y t p + ε t θ 1 ε t 1 θ 2 ε t 2 + + θ q ε t q Y ^ t = Y ^ l t + Y ^ n t Y n t = Y t Y ^ l t Y ^ n t = ω 0 + j = 1 H ω j f ω 0 j + i = 1 N ω i j Y n t i + e t ,
where Y ^ t is the final prediction result, Y ^ l t is the ARMA prediction result, Y n t is the residual of the ARMA model, and Y ^ n t is the prediction result of the ANN model.

2.3. Feature Extraction

Both time-domain and frequency-domain indicators are affected by the equipment’s operating condition [19]. Table 1 presents the features extracted from the sensitive IMFs, which are the subseries that contain 90% of the energy of the original signal. The energy of each IMF can refer to the Ref. [47].

2.4. Support Vector Machine

SVM is one of the most celebrated and popular classification algorithms in the field of multi-classification [39,40,48]. The prime purpose of SVM is to maximize the margin of between-class, meanwhile narrowing the distance between the hyperplane focuses. The training set of SVM is denoted by
S = x i , y i i = 1 n | x i R N , y i 1 , 1 , i = 1 , 2 , , n ,
where x i and y i are the sample data and category, respectively. The hyperplane can separate the data correctly with the maximum interval. The solution of the optimal separating plane is described as:
min 1 2 ω 2 s . t . y i ω x i + b 1 ,
where b is the bias vector, ω denotes the weight vector, and ξ i 0 i = 1 , 2 , , n represents a relaxing factor, which ensures the accuracy of classification. We rewrite the problem (8) as the following minimization problem:
min 1 2 ω 2 + λ i = 1 n ξ i ξ 0 s . t . y i ω x i + b 1 ξ i i = 1 , 2 , , n λ 0 ,
where λ is the penalty factor, which is used to relieve the SVM classifier dilemma between accuracy and complexity. The classical hinge loss function is applied in this research. Thus, the dual optimization problem of SVM is described as:
max L α i = i = 1 n α i 1 2 i = 1 n j = 1 n α i α j y i y j K x i , x j s . t . i = 1 n α i y i = 0 0 α i λ , i = 1 , 2 , , n ,
where K x i , x j denotes a kernel function, K x i , x j = Φ x i · Φ x j , and · is the inner product operation. The SVM classifier with a radial basis function (RBF) kernel, K x i , x j = e x p γ x i x j 2 , γ > 0 is applied in this work, where γ is a hyper-parameter that affects the merits of the SVM. Thus, the classification is described as:
S x = s i g n i = 1 n α i y i Φ x i · Φ x j + b .

3. Proposed Architecture

An architecture for the condition monitoring of the rolling bearing is shown in Figure 2, which is mainly divided into four submodules (VMD decomposition, ARMA-ANN prediction, T-F features extraction and SVM classification).
The specific details are as follows.
(1)
The VMD algorithm decomposes the multi-component vibration signal of rolling bearing into several IMFs. The descriptions of the VMD algorithm are presented in Section 2.1;
(2)
Based on the established ARMA-ANN prediction model, the prediction of each IMF is conducted, and the sensitive IMFs are selected;
(3)
The multi-domain features set, i.e., T-F features set including time domain and frequency domain, are extracted as characteristic parameters from the sensitive IMFs. The specific features are presented in Section 2.3;
(4)
The classification of the condition is performed by a SVM classifier based on T-F features.

4. Experimental Analysis

4.1. Data Description

In this section, the motor bearing data set from the CWRU laboratory is utilized to demonstrate the advantages of our proposed method. More details on the experiment are referenced in [19,49].

4.2. Experimental Results and Analysis

In this paper, four classes of bearing operating conditions data sets, including normal (NS), inner race fault (IF), outer race fault (OF), and ball fault (BF) were applied, respectively. Each piece of the data set consisted of 12,000 points. The working speed, fault size, load and sampling frequency were set to 1730 r/min, 0.007 in, 3 HP, 12 kHz, respectively.

4.2.1. Decomposing the Vibration Signal Using VMD

According to the steps in Section 3, a group of IF data as an example was analyzed. Table 2 shows the VMD relevant parameters according to [19]. The time-waveform of the original signal and the modes obtained by VMD are presented in the left of Figure 3, and the corresponding frequency spectrums are presented in the right of this figure.

4.2.2. Prediction Based on Hybrid Prediction Model

Based on the aforementioned VMD, the original signal was decomposed into several IMF time series, using the hybrid prediction model to obtain the trend of subseries. First, the obtained six IMF series by VMD were subjected to ARMA. We used the AR model of order 12 (AR(12)) and MA model of order 16 (MA(16)) for IMF1. Then the three-layered 1 × 4 × 2 ANN architecture was used for the residuals. Figure 4, Figure 5 and Figure 6 present the expectations and predictions of IF IMF1 series for ARMA, ANN and ARMA-ANN, respectively. The predictions of the IMF1 for NS, BF, OF, using ARMA-ANN are presented in Figure 7, Figure 8 and Figure 9, respectively.
To further quantify the performance of the prediction models, the comparison of quantitative criteria is shown in Table 3, including the root-mean-square error (RMSE): 1 n t = 1 n e t 2 , the mean average percentage error (MAPE): 100 % n t = 1 n | e t Y t | , and the mean absolute error (MAE): 1 n t = 1 n | e t | , which are used to quantify the prediction accuracy, where e t denotes the error between expectation and the prediction.
We further compared the prediction of subseries with the original data series. The prediction of the IF original data series for the ARMA-ANN obtained by VMD is shown in Figure 10.
The quantitative evaluation criteria shown in Table 4 demonstrate that the prediction of subseries obtained by VMD achieves better performance than the original data series.
This result motivated us to make original data series into subseries to improve the accuracy of results. Furthermore, our hybrid prediction method with VMD could provide a superior scheme for machinery condition monitoring.

4.2.3. Feature Extraction

Next, the features were extracted from the sensitive IMFs. The power spectral density (PSD) distribution of the original signal are shown in Figure 11, which demonstrates that the sensitive modes are the 2nd IMF to the 6th IMF. However, the 1st is obviously redundant for the downstream feature extraction and classification processing. Then, we extracted the T-F features from the 2nd IMF to the 6th IMF to form a new feature set and also normalized the set.

4.2.4. Classification Based on SVM

In this subsection, the condition patterns are identified using SVM based on the normalized feature set. The performance of classification highly depended on the features set. Consequently, the time-domain features and the frequency-domain features were combined to construct the input data vectors. Then, a normalized 10-dimension input vector was fed into SVM to train for the downstream classification processing. Figure 12, Figure 13 and Figure 14 show the recognition results based on the features of different domains.
The quantitative classification performance of the SVM is compared in Table 5, which shows the classification results of the condition patterns using the features of different domains.
The classification accuracy of individual time-domain features or individual frequency-domain features is poor as compared to the time-frequency domain features. These results suggest that either time domain or frequency domain alone cannot fully capture the condition patterns. To tackle this problem, it would be effective to make use of each of the domain features to comprehensively describe all patterns.
In terms of the computational complexity [50,51] and the number of sample points N d with d dimensions to be trained, the computational complexity of the training phase required could be O N d 2 d . In the test phase, the computational complexity required O M N operations, where M and N are the kernel operations and support vectors, respectively. The computational complexity of SVM in this example, specifically the 400 points with 50 dimension T-F features to be trained, required the computational complexity to be O 400 2 × 50 . The RBF kernel was evaluated for 200 operations requiring 307 support vectors. Therefore, the computational complexity required O 200 × 307 for the test phase.

4.2.5. Discussion

Considerable achievements have been made in time series forecasting [52] and classification [53] for the deep learning-based methods; however, these methods highly depend on the input scale level. Meanwhile, it is worthy to notice that the deep learning-based methods show accuracy associated with the scale level, and the more scales contribute, the more accurate and stable the performance becomes. Therefore, the solution to acquire high performance is costly in terms of time to train and test. This fact indicates that the augmentation of computation becomes the problem caused by its scale level. Moreover, most of the deep learning-based methods cannot deal with a small sample of data; in other words, accuracy and stability cannot be guaranteed in this case.
This research uses 12,000 sample points to verify the performance of the prediction and classification. The biggest challenge is that the data used for training are much less. The reason for this is that our proposed method is conducted to monitor the rolling bear operating condition. Moreover, our strategy for condition monitoring is also effective while simultaneously relieving time resource problems. A detailed description of the comparative method’s parameters is given as follows. The LSTM [52] is utilized to forecast the future trend of the subseries, and “Sigmoid” and “Adam” are utilized as the activation function and optimizer, respectively. The network is updated by minimizing the mean square error for e t with a gradient descent algorithm. The configurations of the three one-dimensional CNN [53] for classifying layers are: Conv1D 16 , 64 , 16 , Conv1D 32 , 3 , 1 , Conv1D 64 , 3 , 1 . After the convolution operation, the Batch Normalization is carried out. “Relu” is applied as an activation function. Then, the convolution kernel 2 × 1 is employed in the maximum pooling operation. To more clearly illustrate the advantages of our proposed method on time series forecasting using the same small sample data set, a comparison of quantitative criteria for prediction results is presented in Table 6. Our proposed method outperforms the CNN-based method in classification, whose averaged accuracy is higher than 91.88 % . Therefore, both the performance of prediction and classification confirm the advantages of our proposed method.

5. Conclusions

In this paper, an effective condition monitoring scheme is designed for rolling bearings. First, the original signal can be sufficiently described by IMFs obtained by VMD. Next, each IMF is forecasted through the hybrid prediction model ARMA-ANN. Moreover, the sensitive IMFs, which contain over 90% of the energy of the original signal, are selected to provide downstream feature extraction and classification processing. Then, the T-F features, which can represent the comprehensive condition characteristics of the rolling bearing, are extracted and normalized. Finally, the normalized T-F features are fed into the SVM classifier to identify the operating patterns of the rolling bearing. The performance of the proposed condition monitoring method is verified by the motor-bearing data set of the CWRU laboratory.
Note that the scheme of the original signal first decomposed by VMD into a subseries provides the sufficiency and convenience of analyzing the operating condition. Secondly, the hybrid prediction model enhances the accuracy of prediction. Thirdly, the condition patterns are identified accurately using a SVM classifier based on the T-F features. Additionally, the proposed method balances the performance and time resources equally well. Thus, it can be concluded that our proposed method under the hybrid framework has some potential value for monitoring machinery condition.

Author Contributions

Conceptualization, A.W.; methodology, A.W.; software, A.W. and B.X.; validation, A.W. and B.X.; formal analysis, A.W.; investigation, A.W.; data curation, A.W. and B.X.; writing—original draft preparation, A.W. and B.X.; writing—review and editing, Y.L., Z.Y., C.Z. and Z.G.; funding acquisition, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the LiaoNing Revitalization Talents Program XLYC1903015.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available in a publicly accessible repository that does not issue DOIs Publicly available datasets were analyzed in this study. This data can be found here: (https://engineering.case.edu/bearingdatacenter, accessed on 5 April 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
VMDVariational mode decomposition
IMFsIntrinsic mode functions
ARMA         Autoregressive moving average
ANNArtificial neural network
ARMA-ANNAutoregressive moving average-Artificial neural network
SVMSupport vector machine
CWRUCase Western Reserve University
WTWavelet decomposition
EEMDEnsemble empirical mode decomposition
EMDEmpirical mode decomposition
ADMMAlternate direction method of multipliers
RBFRadial basis function
ARAuto-regression
MAMoving average
T-FTime-frequency
IFInner race fault
OFOuter race fault
BFBall fault
PSDPower spectral density
RMSERoot-mean-square error
MAPEMean average percentage error
MAEMean absolute error

References

  1. Shen, X.; Wu, Y.; Shen, T. Logical control scheme with real-time statistical learning for residual gas fraction in IC engines. Sci. China Inf. Sci. 2018, 61, 010203. [Google Scholar] [CrossRef]
  2. Wu, Y.; Shen, T. Policy iteration approach to control residual gas fraction in IC engines under the framework of stochastic logical dynamics. IEEE Trans. Control. Syst. Technol. 2016, 25, 1100–1107. [Google Scholar] [CrossRef]
  3. Zhang, Y.; Shen, X.; Wu, Y.; Shen, T. On-board knock probability map learning–based spark advance control for combustion engines. Int. J. Engine Res. 2019, 20, 1073–1088. [Google Scholar] [CrossRef]
  4. Wu, Y.; Guo, Y.; Toyoda, M. Policy iteration approach to the infinite horizon average optimal control of probabilistic Boolean networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 2910–2924. [Google Scholar] [CrossRef] [PubMed]
  5. Brito, L.C.; Susto, G.A.; Brito, J.N.; Duarte, M.A. An explainable artificial intelligence approach for unsupervised fault detection and diagnosis in rotating machinery. Mech. Syst. Signal Process. 2022, 163, 108105. [Google Scholar] [CrossRef]
  6. Liu, R.; Yang, B.; Zio, E.; Chen, X. Artificial intelligence for fault diagnosis of rotating machinery: A review. Mech. Syst. Signal Process. 2018, 108, 33–47. [Google Scholar] [CrossRef]
  7. Randall, R.B. Vibration-Based Condition Monitoring: Industrial, Automotive and Aerospace Applications; John Wiley & Sons: Hoboken, NJ, USA, 2021. [Google Scholar]
  8. Vishwakarma, M.; Purohit, R.; Harshlata, V.; Rajput, P. Vibration analysis & condition monitoring for rotating machines: A review. Mater. Today Proc. 2017, 4, 2659–2664. [Google Scholar]
  9. Wang, D.; Miao, Q.; Fan, X.; Huang, H.Z. Rolling element bearing fault detection using an improved combination of Hilbert and wavelet transforms. J. Mech. Sci. Technol. 2009, 23, 3292–3301. [Google Scholar] [CrossRef]
  10. Li, Z.; Feng, Z.; Chu, F. A load identification method based on wavelet multi-resolution analysis. J. Sound Vib. 2014, 333, 381–391. [Google Scholar] [CrossRef]
  11. Huang, N.E.; Shen, Z.; Long, S.R.; Wu, M.C.; Shih, H.H.; Zheng, Q.; Yen, N.C.; Tung, C.C.; Liu, H.H. The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis. Proc. R. Soc. Lond. Ser. A Math. Phys. Eng. Sci. 1998, 454, 903–995. [Google Scholar] [CrossRef]
  12. Huang, N.E.; Wu, Z. A review on Hilbert-Huang transform: Method and its applications to geophysical studies. Rev. Geophys. 2008, 46. [Google Scholar] [CrossRef] [Green Version]
  13. Huang, N.E. Introduction to the Hilbert–Huang transform and its related mathematical problems. In Hilbert–Huang Transform and Its Applications; World Scientific: Singapore, 2014; pp. 1–26. [Google Scholar]
  14. Wu, Z.; Huang, N.E. Ensemble empirical mode decomposition: A noise-assisted data analysis method. Adv. Adapt. Data Anal. 2009, 1, 1–41. [Google Scholar] [CrossRef]
  15. Lei, Y.; He, Z.; Zi, Y. Application of the EEMD method to rotor fault diagnosis of rotating machinery. Mech. Syst. Signal Process. 2009, 23, 1327–1338. [Google Scholar] [CrossRef]
  16. Wang, T.; Zhang, M.; Yu, Q.; Zhang, H. Comparing the applications of EMD and EEMD on time–frequency analysis of seismic signal. J. Appl. Geophys. 2012, 83, 29–34. [Google Scholar] [CrossRef]
  17. Zhang, X.; Sun, T.; Wang, Y.; Wang, K.; Shen, Y. A parameter optimized variational mode decomposition method for rail crack detection based on acoustic emission technique. Nondestruct. Test. Eval. 2021, 36, 411–439. [Google Scholar] [CrossRef]
  18. Dragomiretskiy, K.; Zosso, D. Variational mode decomposition. IEEE Trans. Signal Process. 2013, 62, 531–544. [Google Scholar] [CrossRef]
  19. Lei, Z.; Wen, G.; Dong, S.; Huang, X.; Zhou, H.; Zhang, Z.; Chen, X. An intelligent fault diagnosis method based on domain adaptation and its application for bearings under polytropic working conditions. IEEE Trans. Instrum. Meas. 2020, 70, 3505914. [Google Scholar] [CrossRef]
  20. Wang, Y.; Markert, R.; Xiang, J.; Zheng, W. Research on variational mode decomposition and its application in detecting rub-impact fault of the rotor system. Mech. Syst. Signal Process. 2015, 60, 243–251. [Google Scholar] [CrossRef]
  21. Wang, X.B.; Yang, Z.X.; Yan, X.A. Novel particle swarm optimization-based variational mode decomposition method for the fault diagnosis of complex rotating machinery. IEEE/ASME Trans. Mechatron. 2017, 23, 68–79. [Google Scholar] [CrossRef]
  22. Lian, J.; Liu, Z.; Wang, H.; Dong, X. Adaptive variational mode decomposition method for signal processing based on mode characteristic. Mech. Syst. Signal Process. 2018, 107, 53–77. [Google Scholar] [CrossRef]
  23. Chen, Y.; Liang, X.; Zuo, M.J. Sparse time series modeling of the baseline vibration from a gearbox under time-varying speed condition. Mech. Syst. Signal Process. 2019, 134, 106342. [Google Scholar] [CrossRef]
  24. Doroudyan, M.H.; Niaki, S.T.A. Pattern recognition in financial surveillance with the ARMA-GARCH time series model using support vector machine. Expert Syst. Appl. 2021, 182, 115334. [Google Scholar] [CrossRef]
  25. Zaw, T.; Kyaw, S.S.; Oo, A.N. ARMA Model for Revenue Prediction. In Proceedings of the 11th International Conference on Advances in Information Technology, Bangkok, Thailand, 1–3 July 2020; pp. 1–6. [Google Scholar]
  26. Zhang, Y.; Zhao, Y.; Kong, C.; Chen, B. A new prediction method based on VMD-PRBF-ARMA-E model considering wind speed characteristic. Energy Convers. Manag. 2020, 203, 112254. [Google Scholar] [CrossRef]
  27. Bang, S.; Bishnoi, R.; Chauhan, A.S.; Dixit, A.K.; Chawla, I. Fuzzy Logic based Crop Yield Prediction using Temperature and Rainfall parameters predicted through ARMA, SARIMA, and ARMAX models. In Proceedings of the 2019 Twelfth International Conference on Contemporary Computing (IC3), Noida, India, 8–10 August 2019; pp. 1–6. [Google Scholar]
  28. Ji, W.; Chee, K.C. Prediction of hourly solar radiation using a novel hybrid model of ARMA and TDNN. Sol. Energy 2011, 85, 808–817. [Google Scholar] [CrossRef]
  29. Makridakis, S.; Spiliotis, E.; Assimakopoulos, V. Statistical and Machine Learning forecasting methods: Concerns and ways forward. PLoS ONE 2018, 13, e0194889. [Google Scholar] [CrossRef] [Green Version]
  30. Moon, J.; Hossain, M.B.; Chon, K.H. AR and ARMA model order selection for time-series modeling with ImageNet classification. Signal Process. 2021, 183, 108026. [Google Scholar] [CrossRef]
  31. Aras, S.; Kocakoç, İ.D. A new model selection strategy in time series forecasting with artificial neural networks: IHTS. Neurocomputing 2016, 174, 974–987. [Google Scholar] [CrossRef]
  32. Chandran, V.; K Patil, C.; Karthick, A.; Ganeshaperumal, D.; Rahim, R.; Ghosh, A. State of charge estimation of lithium-ion battery for electric vehicles using machine learning algorithms. World Electr. Veh. J. 2021, 12, 38. [Google Scholar] [CrossRef]
  33. Anushka, P.; Upaka, R. Comparison of different artificial neural network (ANN) training algorithms to predict the atmospheric temperature in Tabuk, Saudi Arabia. Mausam 2020, 71, 233–244. [Google Scholar] [CrossRef]
  34. Zhang, G.P. Time series forecasting using a hybrid ARIMA and neural network model. Neurocomputing 2003, 50, 159–175. [Google Scholar] [CrossRef]
  35. Brusa, E.; Bruzzone, F.; Delprete, C.; Di Maggio, L.G.; Rosso, C. Health indicators construction for damage level assessment in bearing diagnostics: A proposal of an energetic approach based on envelope analysis. Appl. Sci. 2020, 10, 8131. [Google Scholar] [CrossRef]
  36. Alessandro Paolo, D.; Luigi, G.; Alessandro, F.; Stefano, M. Performance of Envelope Demodulation for Bearing Damage Detection on CWRU Accelerometric Data: Kurtogram and Traditional Indicators vs. Targeted a Posteriori Band Indicators. Appl. Sci. 2021, 11, 6262. [Google Scholar] [CrossRef]
  37. Zhang, X.; Miao, Q.; Zhang, H.; Wang, L. A parameter-adaptive VMD method based on grasshopper optimization algorithm to analyze vibration signals from rotating machinery. Mech. Syst. Signal Process. 2018, 108, 58–72. [Google Scholar] [CrossRef]
  38. Gao, Y.; Mosalam, K.M.; Chen, Y.; Wang, W.; Chen, Y. Auto-Regressive Integrated Moving-Average Machine Learning for Damage Identification of Steel Frames. Appl. Sci. 2021, 11, 6084. [Google Scholar] [CrossRef]
  39. Yan, X.; Jia, M. A novel optimized SVM classification algorithm with multi-domain feature and its application to fault diagnosis of rolling bearing. Neurocomputing 2018, 313, 47–64. [Google Scholar] [CrossRef]
  40. Cervantes, J.; Garcia-Lamont, F.; Rodríguez-Mazahua, L.; Lopez, A. A comprehensive survey on support vector machine classification: Applications, challenges and trends. Neurocomputing 2020, 408, 189–215. [Google Scholar] [CrossRef]
  41. Pang, H.; Wang, S.; Dou, X.; Liu, H.; Chen, X.; Yang, S.; Wang, T.; Wang, S. A Feature Extraction Method Using Auditory Nerve Response for Collapsing Coal-Gangue Recognition. Appl. Sci. 2020, 10, 7471. [Google Scholar] [CrossRef]
  42. Xiang, J.; Zhong, Y. A novel personalized diagnosis methodology using numerical simulation and an intelligent method to detect faults in a shaft. Appl. Sci. 2016, 6, 414. [Google Scholar] [CrossRef] [Green Version]
  43. Paolanti, M.; Romeo, L.; Felicetti, A.; Mancini, A.; Frontoni, E.; Loncarski, J. Machine learning approach for predictive maintenance in industry 4.0. In Proceedings of the 2018 14th IEEE/ASME International Conference on Mechatronic and Embedded Systems and Applications (MESA), Oulu, Finland, 2–4 July 2018; pp. 1–6. [Google Scholar]
  44. Mamuya, Y.D.; Lee, Y.D.; Shen, J.W.; Shafiullah, M.; Kuo, C.C. Application of machine learning for fault classification and location in a radial distribution grid. Appl. Sci. 2020, 10, 4965. [Google Scholar] [CrossRef]
  45. Long, B.; Wu, K.; Li, P.; Li, M. A Novel Remaining Useful Life Prediction Method for Hydrogen Fuel Cells Based on the Gated Recurrent Unit Neural Network. Appl. Sci. 2022, 12, 432. [Google Scholar] [CrossRef]
  46. Ren, Y.; Suganthan, P.; Srikanth, N. A comparative study of empirical mode decomposition-based short-term wind speed forecasting methods. IEEE Trans. Sustain. Energy 2014, 6, 236–244. [Google Scholar] [CrossRef]
  47. Bi, F.; Li, X.; Liu, C.; Tian, C.; Ma, T.; Yang, X. Knock detection based on the optimized variational mode decomposition. Measurement 2019, 140, 1–13. [Google Scholar] [CrossRef]
  48. Maldonado, S.; López, J. Dealing with high-dimensional class-imbalanced datasets: Embedded feature selection for SVM classification. Appl. Soft Comput. 2018, 67, 94–105. [Google Scholar] [CrossRef]
  49. Gu, R.; Chen, J.; Hong, R.; Wang, H.; Wu, W. Incipient fault diagnosis of rolling bearings based on adaptive variational mode decomposition and Teager energy operator. Measurement 2020, 149, 106941. [Google Scholar] [CrossRef]
  50. Burges, C.J. A tutorial on support vector machines for pattern recognition. Data Min. Knowl. Discov. 1998, 2, 121–167. [Google Scholar] [CrossRef]
  51. Razzak, I.; Zafar, K.; Imran, M.; Xu, G. Randomized nonlinear one-class support vector machines with bounded loss function to detect of outliers for large scale IoT data. Future Gener. Comput. Syst. 2020, 112, 715–723. [Google Scholar] [CrossRef]
  52. Shao, X.; Pu, C.; Zhang, Y.; Kim, C.S. Domain fusion CNN-LSTM for short-term power consumption forecasting. IEEE Access 2020, 8, 188352–188362. [Google Scholar] [CrossRef]
  53. Shao, X.; Kim, C.S.; Kim, D.G. Accurate multi-scale feature fusion CNN for time series classification in smart factory. Comput. Mater. Contin. 2020, 65, 543–561. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of ANN.
Figure 1. Schematic diagram of ANN.
Applsci 12 03854 g001
Figure 2. The architecture of the proposed condition monitoring strategy.
Figure 2. The architecture of the proposed condition monitoring strategy.
Applsci 12 03854 g002
Figure 3. Time-domain waveform and frequency spectrum of original signal and IMFs for inner race signal by VMD ( K = 6 , α = 2891 ).
Figure 3. Time-domain waveform and frequency spectrum of original signal and IMFs for inner race signal by VMD ( K = 6 , α = 2891 ).
Applsci 12 03854 g003
Figure 4. Prediction result of ARMA model for IF IMF1 time series.
Figure 4. Prediction result of ARMA model for IF IMF1 time series.
Applsci 12 03854 g004
Figure 5. Prediction result of ANN model for IF IMF1 time series.
Figure 5. Prediction result of ANN model for IF IMF1 time series.
Applsci 12 03854 g005
Figure 6. Prediction result of ARMA-ANN model for IF IMF1 time series.
Figure 6. Prediction result of ARMA-ANN model for IF IMF1 time series.
Applsci 12 03854 g006
Figure 7. Prediction result of ARMA-ANN model for NS IMF1 time series.
Figure 7. Prediction result of ARMA-ANN model for NS IMF1 time series.
Applsci 12 03854 g007
Figure 8. Prediction result of ARMA-ANN model for BF IMF1 time series.
Figure 8. Prediction result of ARMA-ANN model for BF IMF1 time series.
Applsci 12 03854 g008
Figure 9. Prediction result of ARMA-ANN model for OF IMF1 time series.
Figure 9. Prediction result of ARMA-ANN model for OF IMF1 time series.
Applsci 12 03854 g009
Figure 10. Prediction result of ARMA-ANN model for IF original time series.
Figure 10. Prediction result of ARMA-ANN model for IF original time series.
Applsci 12 03854 g010
Figure 11. The energy ratio distribution of subseries for IF data by VMD decomposition.
Figure 11. The energy ratio distribution of subseries for IF data by VMD decomposition.
Applsci 12 03854 g011
Figure 12. Recognition results of SVM classifier based on time-domain features.
Figure 12. Recognition results of SVM classifier based on time-domain features.
Applsci 12 03854 g012
Figure 13. Recognition results of SVM classifier based on frequency-domain features.
Figure 13. Recognition results of SVM classifier based on frequency-domain features.
Applsci 12 03854 g013
Figure 14. Recognition results of SVM classifier based on T-F features.
Figure 14. Recognition results of SVM classifier based on T-F features.
Applsci 12 03854 g014
Table 1. The expression of feature parameters.
Table 1. The expression of feature parameters.
Time-Domain
Parameter
Feature ExpressionFrequency-Domain
Parameter
Feature Expression
F1 | max x i min x i | F6 i = 1 N f i G f i i = 1 N G f i
F2 max x i 1 N i = 1 N | x i | F7 i = 1 N f i G f i i = 1 N f i
F3 max x i 1 N i = 1 N | x i | 2 F8 i = 1 N f i F 6 2 G f i i = 1 N G f i
F4 1 N i = 1 N x i μ 3 σ 3 F9 1 N i = 1 N G f i
F5 1 N i = 1 N x i μ 4 σ 4 F10 i = 1 N f i 2 G f i i = 1 N G f i * i = 1 N f i 4 G f i
Table 2. Parameter settings for VMD.
Table 2. Parameter settings for VMD.
AlgorithmParameterSetting
VMDK6
α 2891
DC0
tau0
init1
tol 10 7
Table 3. Comparison of performance criteria of three models for all dataset series.
Table 3. Comparison of performance criteria of three models for all dataset series.
ConditionModelRMSEMAPEMAE
IFARMA1.0279 × 10 1 1.1276 × 10 0 8.2175 × 10 2
ANN9.6119 × 10 2 2.9847 × 10 0 7.5156 × 10 2
ARMA-ANN3.6904 × 10 4 2.2367 × 10 2 2.6025 × 10 4
NSARMA1.4100 × 10 2 1.8500 × 10 0 1.1707 × 10 2
ANN1.1744 × 10 2 1.8433 × 10 0 9.4530 × 10 3
ARMA-ANN4.6082 × 10 4 3.0578 × 10 2 2.9941 × 10 4
BFARMA4.2184 × 10 2 1.3759 × 10 0 3.4340 × 10 2
ANN3.9973 × 10 2 4.0775 × 10 0 3.2100 × 10 2
ARMA-ANN8.1888 × 10 4 4.3402 × 10 2 2.6100 × 10 4
OFARMA1.2997 × 10 1 1.0736 × 10 0 1.0533 × 10 1
ANN1.2400 × 10 1 2.1900 × 10 0 9.9000 × 10 2
ARMA-ANN2.4100 × 10 3 2.9969 × 10 2 1.5100 × 10 3
Table 4. Comparison of performance criteria of ARMA-ANN model for the original series and subseries.
Table 4. Comparison of performance criteria of ARMA-ANN model for the original series and subseries.
ConditionSeriesRMSEMAPEMAE
IFOriginal5.2744 × 10 3 3.0023 × 10 1 4.1815 × 10 3
Subseries 3.6904 × 10 4 2.2367 × 10 2 2.6025 × 10 4
NSOriginal5.3683 × 10 3 3.1319 × 10 1 4.2680 × 10 3
Subseries 4.6082 × 10 4 3.0578 × 10 2 2.9941 × 10 4
BFOriginal5.2354 × 10 3 3.0846 × 10 1 4.1308 × 10 3
Subseries 8.1888 × 10 4 4.3402 × 10 2 2.6100 × 10 4
OFOriginal5.1080 × 10 3 3.0283 × 10 1 4.0719 × 10 3
Subseries 2.4100 × 10 3 2.9969 × 10 2 1.5100 × 10 3
Table 5. Comparison of the classification results by using different features.
Table 5. Comparison of the classification results by using different features.
DomainClassification AccuracyAverage Accuracy
IFBFOFNormal
Time domain94%86%98%96%93.5%
Frequency domain78%68%88%98%83%
Time-frequency domain98%88%98%96%95%
Table 6. Comparison of performance criteria of prediction results for ARMA-ANN and LSTM.
Table 6. Comparison of performance criteria of prediction results for ARMA-ANN and LSTM.
Operating ConditionMethodRMSEMAPEMAE
Inner RaceLSTM5.5412 × 10 2 7.0271 × 10 1 4.1816 × 10 2
ARMA-ANN5.2744 × 10 3 3.0023 × 10 1 4.1815 × 10 3
NormalLSTM1.2815 × 10 2 4.5444 × 10 1 9.6361 × 10 3
ARMA-ANN5.3683 × 10 3 3.1319 × 10 1 4.2680 × 10 3
BallLSTM2.8762 × 10 2 4.0346 × 10 1 2.1655 × 10 2
ARMA-ANN5.2354 × 10 3 3.0846 × 10 1 4.1308 × 10 3
Outer RaceLSTM1.0495 × 10 1 1.3158 × 10 0 7.7409 × 10 2
ARMA-ANN5.1080 × 10 3 3.0283 × 10 1 4.0719 × 10 3
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, A.; Li, Y.; Yao, Z.; Zhong, C.; Xue, B.; Guo, Z. A Novel Hybrid Model for the Prediction and Classification of Rolling Bearing Condition. Appl. Sci. 2022, 12, 3854. https://doi.org/10.3390/app12083854

AMA Style

Wang A, Li Y, Yao Z, Zhong C, Xue B, Guo Z. A Novel Hybrid Model for the Prediction and Classification of Rolling Bearing Condition. Applied Sciences. 2022; 12(8):3854. https://doi.org/10.3390/app12083854

Chicago/Turabian Style

Wang, Aina, Yingshun Li, Zhao Yao, Chongquan Zhong, Bin Xue, and Zhannan Guo. 2022. "A Novel Hybrid Model for the Prediction and Classification of Rolling Bearing Condition" Applied Sciences 12, no. 8: 3854. https://doi.org/10.3390/app12083854

APA Style

Wang, A., Li, Y., Yao, Z., Zhong, C., Xue, B., & Guo, Z. (2022). A Novel Hybrid Model for the Prediction and Classification of Rolling Bearing Condition. Applied Sciences, 12(8), 3854. https://doi.org/10.3390/app12083854

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop