Next Article in Journal
Performance of Lunar Regolith Shield under Meteoroid Impact: Uncertainties of a Numerical Prediction
Next Article in Special Issue
Composite Fault Diagnosis of Rolling Bearings: A Feature Selection Approach Based on the Causal Feature Network
Previous Article in Journal
Active Disturbance Rejection Control for Speed Control of PMSM Based on Auxiliary Model and Supervisory RBF
Previous Article in Special Issue
Experimental Analysis of the Current Sensor Fault Detection Mechanism Based on Cri Markers in the PMSM Drive System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Computation and Statistical Analysis of Bearings’ Time- and Frequency-Domain Features Enhanced Using Cepstrum Pre-Whitening: A ML- and DL-Based Classification

by
David Cascales-Fulgencio
1,
Eduardo Quiles-Cucarella
2,* and
Emilio García-Moreno
2
1
Escuela Técnica Superior de Ingeniería Industrial, Universitat Politècnica de València, Camino de Vera, s/n, 46022 Valencia, Spain
2
Instituto de Automática e Informática Industrial, Universitat Politècnica de València, Camino de Vera, s/n, 46022 Valencia, Spain
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(21), 10882; https://doi.org/10.3390/app122110882
Submission received: 30 September 2022 / Revised: 20 October 2022 / Accepted: 21 October 2022 / Published: 27 October 2022
(This article belongs to the Special Issue Machine Fault Diagnostics and Prognostics Volume III)

Abstract

:
Vibration signals captured with an accelerometer carry essential information about Rolling Element Bearings (REBs) faults in rotating machinery, and the envelope spectrum has proven to be a robust tool for their diagnosis at an early stage of development. In this paper, Cepstrum Pre-Whitening (CPW) has been applied to REBs’ signals to enhance and extract health-state condition indicators from the preprocessed signals’ envelope spectra. These features are used to train some of the state-of-the-art Machine Learning (ML) algorithms, combined with time-domain features such as basic statistics, high-order statistics and impulsive metrics. Before training, these features were ranked according to statistical techniques such as one-way ANOVA and the Kruskal–Wallis test. A Convolutional Neural Network (CNN) has been designed to implement the classification of REBs’ signals from a Deep Learning (DL) point of view, receiving raw time signals’ greyscale images as inputs. The different ML models have yielded validation accuracies of up to 87.6%, while the CNN yielded accuracy of up to 77.61%, for the entire dataset. In addition, the same models have yielded validation accuracies of up to 97.8%, while the CNN, 90.67%, where signals from REBs with faulty balls have been removed from the dataset, highlighting the difficulty of classifying such faults. Furthermore, from the results of the different ML algorithms compared to those of the CNN, frequency-domain features have proven to be highly relevant condition indicators combined with some time-domain features. These models can be potentially helpful in applications that require early diagnosis of REBs faults, such as wind turbines, vehicle transmissions and industrial machinery.

1. Introduction

One of the clean energy sources that has gained the most momentum in recent years is wind energy, whose global installed capacity grew by 53% year-on-year in 2020 to 743 GW, with more than 93 GW of newly installed capacity [1]. Wind energy is generated by wind turbines that convert the wind’s kinetic energy into electrical energy. The geometry of a wind turbine blade is designed to spin the rotor hub when positioned at a certain angle of attack with the wind. The torque is transmitted through a mechanical system consisting of shafts, couplings, REBs and gearboxes to an asynchronous machine specifically designed for this system: a doubly fed asynchronous machine.
The study of REBs failures from the point of view of vibration analysis is a field extensively researched by the international scientific community, and an essential subject of condition monitoring of rotating machines. Kiral & Karagülle developed a method based on finite element vibration analysis to detect single or multiple faults under the action of an imbalanced force on different components of the REB structure using time and frequency domain features [2]. Sawalhi & Randall conducted an in-depth study on the nature of vibration signals at different stages of rolling elements’ impacts with faults with applications in defect size estimation [3]. Smith & Randall thoroughly analysed the Case Western Reserve University (CWRU) signals with their benchmark method, based on pre-processing them with Discrete/Random Separation and using the envelope spectrum to identify faults [4].
In the scope of artificial intelligence [5,6], numerous papers have been published proposing different methods to train models for REBs diagnosis. In the field of ML, Dong et al. proposed a method based on SVM and the Markov model to predict the degradation process of REBs [7]. Pandya et al. extracted features based on Acoustic Emission (AE) analysis with the Hilbert–Huang Transform (HHT) and used them to train a k-NN model [8]. Piltan et al. proposed a non-linear observer-based technique called Advanced Fuzzy Sliding Mode Observer (AFSMO) to improve the average performance of fault identification with a DT model [9]. In the field of DL, Pan et al. used the second-generation wavelet transform to improve the robustness of DL-based fault diagnosis [10]. Peng et al. converted vibration signals from REBs into greyscale images and used them to extract features to serve as inputs to their proposed CNN [11]. Zhao et al. conducted a DL-based benchmark study evaluating four models: a multilayer perceptron, an auto-encoder, a CNN and an recurrent neural network [12]. Duong et al. suggested obtaining Defect Signature Wavelet Images (DSWIs) from REBs signals, which show the visualization of the discriminated pattern for different types of faults, to train a CNN [13].
Nevertheless, the known advantages offered by the cepstral domain to enhance cyclostationary components in the envelope spectrum are not sufficiently developed by the community in order to extract functional condition indicators that can train ML models to diagnose REBs faults automatically. In this paper, we prove the usefulness of CPW combined with the envelope spectrum to extract frequency-domain features from the (CWRU) signals database that, combined with time-domain features, have yield accuracies over 97% for some ML models. The former frequency-domain condition indicators are the impulse responses in the envelope spectrum, whose amplitudes are the maximum of all impulse responses’ amplitudes spaced at each fault’s characteristic frequency for every signal and their log ratios (Ball Pass Frequency of the Inner Race (BPFI) Amplitude, Ball Pass Frequency of the Outer Race (BPFO) Amplitude, Ball Spin Frequency (BSF) Amplitude, Log (BPFI Amplitude/BPFO Amplitude), Log (BPFI Amplitude/BSF Amplitude), Log (BSF Amplitude/BPFO Amplitude)). The latter time-domain indicators are basic statistics (mean, standard deviation, Root Mean Square (RMS) and shape factor), high-order statistics (kurtosis and skewness) and impulsive metrics (peak value, impulse factor, crest factor and clearance factor). In particular, the log ratio between the maximum amplitude of the impulse responses spaced at the BPFI and the maximum amplitude of the impulse responses spaced at the BPFO in the envelope spectrum has been proven a crucial feature.
The significance of the aforementioned condition indicators has been assessed by applying two statistical methods: one-way Analysis of Variance (ANOVA) and the Kruskal–Wallis test. The ML models trained with these features are Decision Tree (DT), Support Vector Machine (SVM), k-Nearest Neighbors (k-NN) and Naïve Bayes (NB), developed by the MATLAB®-powered Statistics and Machine Learning Toolbox 11.7™. Therefore, this paper focuses on the extraction, analysis and demonstration of feature validity, and not on developing application-specific ML models. Their performance has been compared with that of a CNN trained with greyscale images of raw time signals, as this is a simple method that has delivered outstanding results in previous studies. Furthermore, we provide our MATLAB® functions and processed datasets for a proper assessment of the merits of our method.
This manuscript is organised as follows: Section 2.1 summarises the CWRU’s experimental set-up. Section 2.2 explains the theoretical basis for frequency- and time-domain feature extraction. Section 2.3 covers the statistical analysis of these features and concludes by explaining the method for training ML models. Section 2.4 summarises the process for obtaining greyscale images from the raw CWRU’s time signals. Section 2.5 summarises the proposed CNN’s architecture and defines its hyper-parameters. Section 3.1 presents the statistical analysis results and the different ML algorithms’ validation accuracies. Section 3.2 presents CNN’s validation accuracies. Section 4 discusses the results, comparing them with some of the works referred to above. Finally, Section 5 concludes by presenting the main advantages and disadvantages of the proposed method.

2. Materials and Methods

The following section will cover the in-depth feature extraction for ML and DL models. It also details the method by which the different ML algorithms will be trained, as well as the statistical analysis to which the time- and frequency-domain features have been subjected in order to shed light on the importance of each one for the classification problem. Furthermore, the architecture of the proposed CNN and its different hyper-parameters are explained.

2.1. CWRU’s Experimental Setup Overview

The test stand layout is depicted in Figure 1, and the different REBs geometries and faults’ characteristic frequencies are detailed in [14].
In summary, the test rig consisted of a two-hp induction motor, a torque transducer and encoder and a dynamometer. Faults were implanted into the different REBs’ parts with an electro-discharge machine. Outer race faults were seeded at different positions relative to the REB load zone, since they are stationary, with the fault’s position having an essential effect on the motor/REB system’s vibration response. Vibration signals were captured with accelerometers. During each test, acceleration was measured at the drive-end REB housing (Drive End Accelerometer (DEA)), the fan end REB housing (Fan End Accelerometer (FEA)), and the motor support base plate (Base Accelerometer (BA)). Data were captured at 12 kHz and 48 kHz sampling rates using a 16-channel data acquisition card.
The faulty REBs vibration signals to be classified belong to the drive and fan end REBs tables [14], accompanied by healthy REBs vibration signals. These vibration signals have a commonality: they were collected at a sample rate of 12 kHz. Thus, four groups are derived from the dataset: healthy, inner race fault, outer race fault and ball fault. If vibration signals collected at 48 kHz had wanted to be included, they would have formed four different groups, taking into account the difference between sample rates. For simplification purposes, they have not been included in the dataset. As previously mentioned, each file contains FEA, DEA and BA data. Overall, there are 307 vibration signals, 8 corresponding to healthy REBs, 76 to inner race fault REBs, 76 to ball fault REBs and 147 to outer race fault REBs, as summarised in Table 1.

2.2. Machine Learning Condition Indicators

From a mechanical vibration point of view, each drive train component excites the entire system at its corresponding characteristic frequency in a wind turbine. When an accelerometer is placed at a certain point in the machine, the signal obtained is a sum of the convolutions of the excitations of the different components and the transfer path of these excitations to the measuring element. The latter can be modelled mathematically as a specific impulse response function, as shown by Barszcz in [15].
REBs are components that transmit the loads coming from the supported shaft. They consist of an outer race, an inner race, the rolling elements and a cage to maintain their relative position. When a spall appears on the surface of one of these parts due to fatigue, the rolling elements repetitively impact it. These impacts excite the system in the form of repetitive impulses at the corresponding fault’s characteristic frequency, a function of the shaft rotation frequency, the REB geometry, the number of rolling elements and the load angle, as shown in Table 2.
By computing the fast Fourier transform [16] of a time-domain signal captured by an accelerometer, the harmonics of such excitation will be observed in a specific bandwidth, where the impulse’s resonant frequency will be the fault’s characteristic frequency. Usually, the fault information will be masked in a complex signal such as that of a wind turbine’s drive train.
The envelope spectrum is a widely used tool for diagnosing REB faults in an early stage of development and limited size. Information about the different excitation sources is extracted from the spacing between impulse responses and not from excited frequencies. Therefore, the envelope spectrum consists of a series of impulse responses in the frequency domain spaced at the characteristic frequency of the different excitation sources, as shown in Figure 2. When obtaining the envelope spectrum of a time-domain signal for diagnosing REB faults, it is essential to demodulate the signal in the bandwidth where the fault information is present. For this purpose, a tool called the Kurtogram can be used.
As mentioned above, the rolling elements’ impacts on a spall excite the mechanical system in the form of repetitive impulses at the corresponding fault’s characteristic frequency. These impulses take the form of peaks in a time signal. Spectral Kurtosis (SK) is a remarkably sensitive indicator of a signal’s peakedness. SK is a statistical method to detect non-Gaussian components in a signal, i.e., very impulsive components. Randall and Antoni demonstrated the usefulness of SK in detecting faults in rotating machines [17]. The practical application of SK is the Kurtogram proposed by Antoni and Randall in the same paper, a 2D colour map representing the SK at different levels and bandwidths of a signal. The Kurtogram returns the signal’s most impulsive bandwidth and its centre frequency. Later, Antoni proposed an optimised version of the Kurtogram called the Fast Kurtogram [18]. This version reduces the number of variants for which the filter parameters are calculated without affecting the result’s accuracy by applying the filter bank approach. The main drawback of SK is that it is susceptible to non-Gaussian random components external to the mechanical system, such as noise or interference of various sorts.
Sawalhi & Randall’s CPW [19] can be applied to a raw time signal to enhance the presence of REB’s fault’s impulse responses in the envelope spectrum and remove both unwanted components’ harmonics and sidebands. Earlier in this section, REB fault signals have been referred to as repetitive and not periodic, as they are not strictly periodic: they are second-order cyclostationary. This means that their second-order statistic, the variance, is periodic but not the impulses themselves. The repetitive impulses of REB failures are not precisely periodic due to the rolling elements’ slightly random location in the cage-free space and the non-exact cage’s rotational speed due to slippage. This is why the REB’s fault information is not altered by applying CPW, as the cyclostationary components of a signal do not represent significant peaks in the cepstral domain. A detailed explanation of CPW can be found in the paper by Borghesani et al. [20].
In summary, when a signal is demodulated correctly, a series of impulse responses will appear in the envelope spectrum spaced at the fault’s characteristic frequency. When a fault affects a specific REB part, the impacts with the fault will excite a particular frequency band in the form of repetitive impulses. Suppose this band (including the transfer path from the place of impact to the sensor) is selected for demodulation. In that case, the impulse responses in the envelope spectrum spaced at the actual fault’s characteristic frequency will have a greater amplitude than those spaced at other fault’s characteristic frequencies, as depicted in Figure 2. Computing CPW to a raw time signal enhances cyclostationary components’ responses by filtering out periodic components’ harmonics and sidebands from the envelope spectrum. Furthermore, it sets all frequency components to the same magnitude (around × 10 4 for the CWRU’s signals), as shown in Figure 3.
The above defines a health-state condition indicator that can be calculated using MATLAB®. The idea is to set up a function [21,22] that computes and stores the impulse responses in the envelope spectrum, whose amplitudes are the maximum of all impulse responses’ amplitudes spaced at each fault’s characteristic frequency for every signal (BPFI Amplitude, BPFO Amplitude and BSF Amplitude). Then, by calculating the ratio of the different amplitudes, a quantified relationship between each magnitude and the others is obtained, reflecting their proportion in each case. The logarithm of these ratios is then computed to set them to the same scale. The above is known in the statistical jargon as the log ratio of two values.
L o g B P F I   A m p l i t u d e B P F O   A m p l i t u d e
L o g B S F   A m p l i t u d e B P F O   A m p l i t u d e
L o g B P F I   A m p l i t u d e B S F   A m p l i t u d e
The combination of time- and frequency-domain features has yielded outstanding results, such as in the research by Sánchez, R.-V. et al. [23]. Therefore, the former condition indicators have been combined with time-domain features such as basic statistics (mean, standard deviation, RMS and shape factor), high-order statistics (kurtosis and skewness) and impulsive metrics (peak value, impulse factor, crest factor and clearance factor). These have been obtained by taking advantage of the MATLAB®-powered Predictive Maintenance Toolbox 2.2™, ultimately leading to the MATLAB® function shown in [24,25]. Both condition indicator sets have been ranked according to one-way ANOVA [26] & the Kruskal–Wallis test [27] and used to train four widely renowned ML models (DT, SVM, k-NN and NB).

2.3. Machine Learning-Based Classification Method & Statistical Analysis

The most important step when approaching a ML classification problem is selecting the best feature combination to achieve the most accurate results possible. This is of the utmost importance when the dataset’s size is limited. The most natural questions to be asked once the condition indicators have been calculated are: how are these features meaningful to the classification problem? Are any of them better than the rest? Two ranking techniques for datasets containing more than two classes will be applied to answer these questions: one-way ANOVA [26] and the Kruskal–Wallis test [27].
Comparing the means of three or more unrelated groups within a dependent variable (feature), one-way ANOVA determines whether there are statistically significant differences between them. A non-parametric alternative to the one-way ANOVA test is the Kruskal–Wallis test. Unlike one-way ANOVA, where means are compared, the Kruskal–Wallis test contrasts the samples’ distribution to determine whether they belong to the same population. Both tests make assumptions that the data of each group in every dependent variable must meet. The Kruskal–Wallis test is known to be used if one-way ANOVA’s assumptions are not met for each class within every dependent variable, since they are more potent than those of the Kruskal–Wallis test.
One-way ANOVA tests the null hypothesis that all group means are equal ( H 0 : μ 1 = μ 2 = = μ J ) against the alternative hypothesis that at least one group mean is different from the others ( H 1 : μ j μ j for at least one j and j) in a one-way layout, where i = 1 , , I is the observation number and j = 1 , , J is the group number. The one-way ANOVA result is the ratio of across-group variation to within-group variation F. If F is larger than the critical value of the F-distribution [28] with ( J 1 , J ( I 1 ) ) degrees of freedom and a significance level of α , the null hypothesis is rejected. Therefore, large values of F rank better if the degree of difference between the groups is considered to evaluate if the dependent variable is suitable for a classification problem.
In the Kruskal–Wallis test, the null hypothesis states that J groups from potentially different populations actually derive from a similar population, at least regarding their central tendencies or medians. As an alternative hypothesis, not all groups are derived from the same population. The Kruskal–Wallis test result is the test statistic H. If H is larger than the critical value of the chi-square distribution [29] with J 1 degrees of freedom and a significance level of α , the null hypothesis is rejected. Therefore, large values of H rank better if the degree of difference between the groups is to be considered.
For the one-way ANOVA results to be reliable, the residuals of the I observations belonging to the J groups need to meet the following assumptions:
  • They have to be normally distributed in each group being compared. In practice, the dependent variable is tested to be normally distributed in each group rather than the residuals since the results are the same.
  • There is homogeneity of variances (homoscedasticity) between each group of residuals. Again, the population variances in each group are tested to be equal rather than the residuals’ variances in each group.
  • Independence. The residuals (or rather the observations) need to be independent.
One-way ANOVA is known to be a robust test against the normality assumption, especially for large datasets. When the homoscedasticity condition has been violated, the Welsch-correlated ANOVA test has proved reliable. It has been stated that the most critical assumption to fail is the lack of independence between observations. For the results of the Kruskal–Wallis test to be reliable, the observations within each group need to meet the following assumptions:
  • They do not have to be normally distributed in each group. However, the observations within each group have to belong to the same continuous distribution.
  • There is homogeneity of variances (homoscedasticity) between each group of observations.
  • Independence of observations.
The use of the Kruskal–Wallis test is recommended when the populations to be compared are clearly asymmetric; it is fulfilled that they are all in the same direction and that the variance is homogeneous. To assess if the data within every group and feature are suitable for the results of one-way ANOVA or the Kruskal–Wallis test to be reliable, several hypothesis tests will be applied using the MATLAB®-powered Statistics and Machine Learning Toolbox 11.7™. These are the Anderson–Darling test [30] to test the normality assumption, the Levene test [31] to test the homoscedasticity condition and the Kolmogorov–Smirnov test [32] to test the equality of continuous distributions condition.
Once the features are ranked, the procedure to investigate which features will deliver the highest accuracy for every ML algorithm will be as follows: for every case, each ML algorithm will be computed with a features vector that will be of dimension 1, firstly, increasing its dimensionality until it reaches its maximum, following the order of importance established by each statistical test. Therefore, two tables will be produced, each with the features vector’s dimensionality being increased differently. This way, the best results can be achieved with the fewest features possible, understanding which features are valid and which are not for every algorithm.

2.4. Deep Learning Condition Indicators Overview

In addition, a CNN has been built to classify REBs signals from a DL point of view. CNNs are a type of feed-forward neural network initially designed for image processing. Therefore, the CWRU’s healthy and faulty REBs signals must be transformed into images, with every pixel within an image being a single feature. CNNs are known to deliver outstanding results when the training dataset size is considerably large [33]. This paper will test its performance using a 14.736 image database generated from the CWRU’s raw time signals using the code shown in [34,35].
To transform a 1D time-domain signal into a 2D image to serve as an input to the CNN, N data are split off from a signal x . Each datum is aligned sequentially, shaping m rows of n points. A matrix is then built by placing one of the m rows onto the following. To create square images of acceptable size while preserving as much information as possible about the defect, a 120 k data signal section has been transformed into 48 50 × 50 pixels images. The REBs signals sampled at 12 kHz have approximately 120 k data, given that each reading took about 10 seconds. Suppose the test stand’s shaft rotational speed is 1730 rpm. In that case, a single rolling element will impact approximately 29 times against the fault every second, giving a total of 290 impacts of each rolling element with the fault during these ten seconds (this pattern is not valid if the defect is in the rolling element). In mathematical terms, the transformation described above looks like this:
I m x n = x ( t ) x ( t + n 1 ) x ( t + ( m 1 ) n ) x ( t + m n 1 )
where I denotes the signal image and x ( t ) the vibration datum of time t. In this particular case, t 0 = 1 and m = n = 50 .
Some signal images are depicted in Figure 4:

2.5. CNN Architecture and Hyper-Parameters Overview

The architecture of the proposed CNN model, shown in [36,37], which comprises one input layer responsible for receiving external data, two hidden layers accountable for filtering the inputs, a fully connected layer responsible for the classification and an output layer. Each hidden layer comprises a convolutional layer, a batch normalisation layer, an activation layer and a max-pooling layer. The convolutional layer convolves the local input regions with filter kernels and then generates the output features by computing the activation function. The R e L U function is used as the activation layer to sort out the issue of vanishing gradients. The batch normalisation layer, placed between the convolutional layer and the activation layer, helps to reduce the CNN’s sensitivity to network initialisation. Finally, a max-pooling layer is placed between the first activation layer and the second convolutional layer. This is known to improve the CNN’s accuracy. The fully connected layer comprises as many neurons as there are labels—four, in this case. The s o f t m a x function is used as an activation layer for the latter.
Every ML algorithm aims to minimise a loss/cost function or maximise a likelihood function, depending on the existing model, to find its optimal values and achieve the most accurate prediction. In a CNN, every unit within a convolutional layer is a slight regression model trained by learning the filter kernel’s parameters. The loss function, in this case, is defined by the task. For image classification, the cross-entropy loss function is used to compute the difference between the s o f t m a x output probability distribution and the label probability distribution. The stochastic gradient descent with momentum is used to minimise the cross-entropy loss function. It is an upgraded version of the stochastic gradient descent that accelerates gradient vectors in the right direction. The hyper-parameters for the proposed CNN are shown in Table 3. Furthermore, the training options for the SGDM are set to 0.01 for the learning rate, 15 epochs, a mini-batch size of 35 and a frequency of network validation in 50 iterations. These parameters have been tuned by trial and error.

3. Results

The following section presents the outcomes of applying one-way ANOVA and the Kruskal–Wallis test to the time- and frequency-domain features and the validation results of the different ML algorithms and the CNN.

3.1. Machine Learning-Based Classification

The metric used to assess the different algorithms under the various cases is accuracy, defined as the proportion of correctly classified examples divided by the total number of examples within the dataset. As mentioned in the previous section, to learn which features work better for every ML algorithm under each case, the procedure will be to compute each algorithm with a features vector of dimension 1, increasing its dimensionality until it reaches its maximum for every case, following the order of importance established by each statistical test. Prior to ranking any feature, let us observe various scatter plots of both datasets. The aim is to find significant differences between the features’ distributions for each group.
Figure 5 and Figure 6 depict 3D scatter plots from the time- and frequency-domain datasets. As drawn from these figures, the frequency-domain features seem to be the ones that take the most segregated values for each group, especially the logarithmic ratios between the maximum amplitudes of the envelope spectrum around the pulses spaced at the faults’ characteristic frequency. The time-domain features seem to take less segregated values between the groups. Therefore, lower accuracy is expected to be obtained from these features. Moreover, data belonging to the faulty balls class (2) are more spread than the other groups’ data in the cases where a significant difference between the latter can be observed.
Considering all of the above, two case studies will be defined.
  • Case (A) Classification using the whole dataset.
  • Case (B) Classification removing the second class from the dataset.
The features’ ranking results are shown in Figure 7 and collected in Table 4, Table 5, Table 6 and Table 7.
As shown in Figure 7, for every case, the frequency-domain features are ranked better than the time-domain features, especially with regard to the logarithmic ratios between the maximum amplitudes of the envelope spectrum around the pulses spaced at the faults’ characteristic frequencies. This matches the conclusions drawn from the scatter plots in Figure 5 and Figure 6. The hypothesis tests results are summarised in Table A1, Appendix A. Every feature within the dataset violates the normality and equality of continuous distributions requirements. Only one feature (mean) fulfils the homoscedasticity condition. The independence of observations condition is known to be fulfilled for the whole dataset, since the features have been calculated from the raw time signals, which were independently obtained as described in Section 2.1, from different REBs placed in different parts of the setup, in addition to the accelerometers.
The classification results are depicted in Figure 8, Figure 9, Figure 10 and Figure 11 and collected in Table 8, Table 9, Table 10 and Table 11. Using the MATLAB®-powered Statistics and Machine Learning Toolbox 11.7™, models have been trained by applying five-fold cross-validation, using 80% of the dataset for training and 20% for validation during each iteration. Five-fold cross-validation also protects against overfitting by estimating accuracy on each fold.
Figure 10 and Figure 11 clearly show that signals from REBs with faulty balls are the most difficult to classify using the ML models. The algorithms computed with case A dataset give, in all scenarios, worse results than those calculated with case B dataset. Delving deeper into the development of each model, in Figure 8 and Figure 9, it is not always true that a features vector’s higher dimensionality leads to better results. Two models exemplify that: DT and NB, which in all cases have required fewer features to achieve their best results than SVM or k-NN. Regardless, all models needed fewer features to achieve their best results in case B rather than in case A. This is because features calculated on the REBs’ signals with faulty balls do not help find correlations that ultimately lead to a correct classification. The best-performing models are k-NN for case A [38,39] and SVM for case B [40,41].

3.2. Deep Learning-Based Classification

The previous section stated that REBs’ signals with faulty balls are expected to be the most difficult for the ML algorithms to classify. The question is whether the CNN will also have the same difficulties finding correlations between REBs’ signals with faulty balls. The two case studies defined in Section 3.1 will be applied to answer this question. The proposed CNN architecture and the chosen hyper-parameters have been defined in Section 2.5. For the ML case, the CNN has been trained with 80% of the images in the dataset and validated with the remaining 20% of them in each case. Both subsets have been randomly split. The classification results are shown in Table 12.
Worse results have been obtained from the CNN. Although a database of 14.736 images has been generated, taking full advantage of the length of the CWRU’s signals, this has not been sufficient to improve the results obtained from some ML models. The computational time to train the proposed CNN has been much higher than what was required for the ML models, which was a few seconds. To improve the CNN results, more images could be obtained from more signals, generating a more extensive dataset.

4. Discussion

This paper has trained various ML algorithms and a CNN to classify healthy and faulty REBs. Each model’s best results are summarised in Figure 12. The different models have been trained using the CWRU bearing data center signals. These models are prepared to receive features of never-before-seen REBs’ signals to diagnose their state of health, with an accuracy corresponding to the validation result obtained for each model. Considering the results obtained, the following can be stated:
  • Although the frequency-domain features are ranked better, according to the applied statistical methods, than the time-domain features, it can be observed in the accuracy development of some models that some frequency-domain features contribute to worse signal classification against some time-domain features.
  • The logarithmic ratio between the maximum amplitude of the impulse responses spaced at the BPFI and the maximum amplitude of the impulse responses spaced at the BPFO is the feature containing the most relevant information about REBs’ health status. That follows from the high accuracy obtained by computing the algorithms with this feature alone. According to the statistical methods, it is also the best-positioned feature in the rankings.
  • The combination of time- and frequency-domain features yields better results than by being computed separately. This can be deduced since the best results combine both types of features. To design a robust ML model that generalises correctly, the most important thing to find is the features’ combination providing the most relevant information about the REBs’ health status.
Table 13 compares the results of some relevant papers cited in the introduction with those obtained in this paper. The Root Mean Square Error (RMSE) of the model proposed by Dong et al. suggests an almost perfect REB degradation prediction compared to the data obtained in their experiment.
The results of ML models trained with features based on AE analysis with the HHT, proposed by Pandya et al., demonstrate the usefulness of these physical features. The observer-based non-linear models of Piltan et al. have helped improve the accuracy of the DT model. However, these authors divide the CWRU dataset into numerous groups, considering fault sizes and different load regimes. The latter work overlooks the fact that the load is practically meaningless in this case, as there is no mechanism in the CWRU experiment that converts the torque into a radial load supported by the REBs. Furthermore, the usefulness of these non-physical models with more heterogeneous groups is unclear. Pan et al. obtained excellent results in their novel neural network enhanced with the second-generation wavelet transform. Peng et al. show that the simple method of converting raw time signals into greyscale images to train a CNN provides outstanding results. Once again, the CWRU dataset is subdivided into numerous groups in this work, considering different fault sizes and load regimes. Finally, the model proposed by Duong et al. demonstrates the usefulness of DSWI obtained from faulty REBs’ signals to train CNNs.
This work evaluated the applicability of physical methods, such as the CPW-enhanced envelope spectrum, for the ML-based classification problem. Feature extraction from this method has yielded excellent results demonstrating its usefulness in combination with time-domain features. These results have been superior to those of the CNN trained with greyscale images of raw time signals under the same conditions. The results obtained in this paper are to be understood as a complement to the study of relevant features for diagnosing REBs based on ML.

5. Conclusions

In general terms, excellent accuracies have been obtained from the ML models. The statistical tests’ results and the different algorithms’ accuracy have shown that the frequency-domain features, particularly the logarithmic ratios between the maximum amplitudes of the envelope spectrum around the pulses spaced at the faults’ characteristic frequency, have been the most important ones. These, in addition to some of the most critical time-domain features, have been able to yield the highest accuracies in this paper. Among the models, the k-NN and SVM classifiers were the best performers. The accuracy of these algorithms is given in Table 8, Table 9, Table 10 and Table 11. It can also be seen from Figure 10 and Figure 11 that signals corresponding to REBs with faulty balls have been the most difficult for the algorithms to classify. The CNN’s accuracy can be improved by increasing the dataset’s size, as mentioned in Section 4. These models can be potentially helpful for automatically diagnosing the health status of REBs.
The main drawback of the proposed method is that the database on which the study has been conducted is imbalanced, as seen in Table 1. This means that the classification result may be biased, as the classifiers are more sensitive to detecting the majority class and less sensitive to the minority class. A common technique to counteract the imbalanced data problem is to oversample the dataset. Nevertheless, after applying cross-validation, it is known to lead to over-optimistic results, for which conventional model evaluation metrics such as accuracy are not good indicators of the algorithms’ performance. In the industry, faults in REBs are typically not evenly distributed, with malfunctions such as those affecting inner and outer races occurring more frequently than others. Therefore, dealing with imbalanced databases is a common challenge. The best practice to address this issue is to apply cross-validation by stratifying each k-fold to capture the unbalanced distribution of groups on each target feature. In this paper, as discussed, conventional five-fold cross-validation has been implemented. This is an essential point of improvement for future research.

Author Contributions

Conceptualization, E.Q.-C.; methodology, E.Q.-C., D.C.-F. and E.G.-M.; software, D.C.-F.; validation, E.Q.-C., D.C.-F. and E.G.-M.; formal analysis, D.C.-F., E.Q.-C. and E.G.-M.; investigation, D.C.-F., E.Q.-C. and E.G.-M.; resources, D.C.-F., E.Q.-C. and E.G.-M.; data curation, D.C.-F.; writing—original draft preparation, D.C.-F.; writing—review and editing, E.Q.-C.; visualisation, E.Q.-C.; supervision, E.Q.-C. and E.G.-M.; project administration, E.Q.-C. and E.G.-M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in FigShare and have been cited accordingly.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AEAcoustic Emission.
AFSMOAdvanced Fuzzy Sliding Mode Observer.
ANOVAANalysis Of VAriance.
BABase Accelerometer.
BPFIBall Pass Frequency of the Outer Race.
BPFOBall Pass Frequency of the Inner Race.
BSFBall Spin Frequency.
CNNConvolutional Neural Network.
CPWCepstrum Pre-Whitening.
CWRUCase Western Reserve University.
DEADrive End Accelerometer.
DLDeep Learning.
DSWIDefect Signature Wavelet Image.
DTDecision Tree.
FEAFan End Accelerometer.
HHTHilbert–Huang Transform.
k-NNk-Nearest Neighbors.
MLMachine Learning.
NBNaïve Bayes.
REBRolling Element Bearing.
RMSRoot Mean Square.
RMSERoot Mean Square Error.
SKSpectral Kurtosis.
SVMSupport Vector Machine.

Appendix A

Table A1. Hypothesis tests applied to every group within each feature.
Table A1. Hypothesis tests applied to every group within each feature.
FeaturesNormality Condition 1Homoscedasticity Condition 2Equality of Continuous Distributions Condition 3
Wp-Value123
Clearance_Factor_0010.77270111
Clearance_Factor_11N/A11
Clearance_Factor_21=N/A0
Clearance_Factor_31==N/A
Crest_Factor_0027.65350110
Crest_Factor_11N/A10
Crest_Factor_21=N/A1
Crest_Factor_31==N/A
Impulse_Factor_0015.18340111
Impulse_Factor_11N/A11
Impulse_Factor_21=N/A0
Impulse_Factor_31==N/A
Kurtosis_005.41750.0012111
Kurtosis_11N/A11
Kurtosis_21=N/A1
Kurtosis_31==N/A
Mean_011.30190.2739111
Mean_11N/A00
Mean_21=N/A0
Mean_31==N/A
Peak_Value_004.5550.0039111
Peak_Value_11N/A00
Peak_Value_21=N/A0
Peak_Value_31==N/A
RMS_004.33230.0052111
RMS_11N/A10
RMS_21=N/A0
RMS_31==N/A
Shape_Factor_0010.65470111
Shape_Factor_11N/A10
Shape_Factor_21=N/A1
Shape_Factor_31==N/A
Skewness_004.34220.0051000
Skewness_11N/A10
Skewness_21=N/A0
Skewness_31==N/A
Std_004.36110.005111
Std_11N/A10
Std_21=N/A0
Std_31==N/A
BPFO_Amplitude_00107.1380111
BPFO_Amplitude_10N/A01
BPFO_Amplitude_21=N/A1
BPFO_Amplitude_31==N/A
BPFI_Amplitude_01148.660111
BPFI_Amplitude_11N/A11
BPFI_Amplitude_21=N/A1
BPFI_Amplitude_31==N/A
BSF_Amplitude_015.17630.0017111
BSF_Amplitude_11N/A10
BSF_Amplitude_21=N/A1
BSF_Amplitude_31==N/A
LOG_BPFI_Amplitude_BPFO_Amplitude_0031.19920101
LOG_BPFI_Amplitude_BPFO_Amplitude_11N/A11
LOG_BPFI_Amplitude_BPFO_Amplitude_20=N/A1
LOG_BPFI_Amplitude_BPFO_Amplitude_31==N/A
LOG_BSF_Amplitude_BPFO_Amplitude_0034.66740011
LOG_BSF_Amplitude_BPFO_Amplitude_11N/A11
LOG_BSF_Amplitude_BPFO_Amplitude_20=N/A1
LOG_BSF_Amplitude_BPFO_Amplitude_31==N/A
LOG_BPFI_Amplitude_BSF_Amplitude_0031.15320110
LOG_BPFI_Amplitude_BSF_Amplitude_11N/A11
LOG_BPFI_Amplitude_BSF_Amplitude_20=N/A1
LOG_BPFI_Amplitude_BSF_Amplitude_31==N/A
1 The Anderson–Darling test returns a decision for the null hypothesis that data in vector x are from a population with a normal distribution. The alternative hypothesis is that x is not from a population with a normal distribution. The result is 1 if the test rejects the null hypothesis at the 5% significance level, or 0 otherwise. 2 The Levene test returns a decision for the null hypothesis that columns of data matrix X have the same variance. The alternative hypothesis is that not all columns of data have the same variance. The Levene test rejects the null hypothesis that the variances are equal if W > the upper critical value of the F distribution with J − 1 and J(I − 1) degrees of freedom at a significance level of α. The p-value represents the probability of observing a test statistic as extreme as or more extreme than the observed value under the null hypothesis. 3 The Kolmogorov–Smirnov test returns a decision for the null hypothesis that data in vectors x1 and x2 are from the same continuous distribution. The alternative hypothesis is that x1 and x2 are from different continuous distributions. The result is 1 if the test rejects the null hypothesis at the 5% significance level, or 0 otherwise.

References

  1. Global Wind Energy Council (GWEC). Global Wind Report 2021. Available online: https://gwec.net/global-wind-report-2021/#:~:text=Today%2C%20there%20is%20now%20743,carbon%20emissions%20of%20South%20America (accessed on 14 September 2022).
  2. Kiral, Z.; Karagülle, H. Vibration analysis of rolling element bearings with various defects under the action of an unbalanced force. Mech. Syst. Signal Process. 2006, 20, 1967–1991. [Google Scholar] [CrossRef]
  3. Sawalhi, N.; Randall, R.B. Vibration response of spalled rolling element bearings: Observations, simulations and signal processing techniques to track the spall size. Mech. Syst. Signal Process. 2011, 25, 846–870. [Google Scholar] [CrossRef]
  4. Smith, W.A.; Randall, R.B. Rolling element bearing diagnostics using the Case Western Reserve University data: A benchmark study. Mech. Syst. Signal Process. 2015, 64–65, 100–131. [Google Scholar] [CrossRef]
  5. Heidari, A.; Navimipour, N.J.; Unal, M. Applications of ML/DL in the management of smart cities and societies based on new trends in information technologies: A systematic literature review. Sustain. Cities Soc. 2022, 85, 104089. [Google Scholar] [CrossRef]
  6. Heidari, A.; Jabraeil Jamali, M.A.; Navimipour, N.J.; Akbarpour, S. Deep Q-Learning Technique for Offloading Offline/Online Computation in Blockchain-Enabled Green IoT-Edge Scenarios. Appl. Sci. 2022, 12, 8232. [Google Scholar] [CrossRef]
  7. Dong, S.; Yin, S.; Tang, B.; Chen, L.; Luo, T. Bearing Degradation Process Prediction Based on the Support Vector Machine and Markov Model. Shock Vib. 2014, 1–15. [Google Scholar] [CrossRef] [Green Version]
  8. Pandya, D.H.; Upadhyay, S.H.; Harsha, S.P. Fault diagnosis of rolling element bearing with intrinsic mode function of acoustic emission data using APF-KNN. Expert Syst. Appl. 2013, 40, 4137–4145. [Google Scholar] [CrossRef]
  9. Piltan, F.; Prosvirin, A.E.; Jeong, I.; Im, K.; Kim, J.-M. Rolling-Element Bearing Fault Diagnosis Using Advanced Machine Learning-Based Observer. Appl. Sci. 2019, 9, 5404. [Google Scholar] [CrossRef] [Green Version]
  10. Pan, J.; Zi, Y.; Chen, J.; Zhou, Z.; Wang, B. LiftingNet: A Novel Deep Learning Network With Layerwise Feature Learning from Noisy Mechanical Data for Fault Classification. IEEE Trans. Ind. Electron. 2018, 65, 4973–4982. [Google Scholar] [CrossRef]
  11. Peng, X.; Zhang, B.; Gao, D. Research on fault diagnosis method of rolling bearing based on 2DCNN. In Proceedings of the 2020 Chinese Control and Decision Conference (CCDC), Hefei, China, 22 August 2020; pp. 693–697. [Google Scholar] [CrossRef]
  12. Zhao, Z.; Li, T.; Wu, J.; Sun, C.; Wang, S.; Yan, R.; Chen, X. Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Trans. 2020, 107, 224–255. [Google Scholar] [CrossRef] [PubMed]
  13. Duong, B.P.; Kim, J.Y.; Jeong, I.; Im, K.; Kim, C.H.; Kim, J.M. A Deep-Learning-Based Bearing Fault Diagnosis Using Defect Signature Wavelet Image visualisation. Appl. Sci. 2020, 10, 8800. [Google Scholar] [CrossRef]
  14. Case Western Reserve University (CWRU). Bearing Data Center. Available online: https://engineering.case.edu/bearingdatacenter/welcome (accessed on 18 September 2022).
  15. Barszcz, T. Vibration-Based Condition Monitoring of Wind Turbines, 1st ed.; Springer: Cham, Switzerland, 2019; pp. 8–23. [Google Scholar] [CrossRef]
  16. Randall, R.B. Vibration-Based Condition Monitoring: Industrial, Automotive and Aerospace Applications, 1st ed.; John Wiley & Sons: Hoboken, NJ, USA, 2010; pp. 66–95. [Google Scholar] [CrossRef]
  17. Antoni, J.; Randall, R.B. The spectral kurtosis: Application to the vibratory surveillance and diagnostics of rotating machines. Mech. Syst. Signal Process. 2004, 20, 308–331. [Google Scholar] [CrossRef]
  18. Antoni, J. Fast computation of the kurtogram for the detection of transient faults. Mech. Syst. Signal Process. 2006, 21, 108–124. [Google Scholar] [CrossRef]
  19. Sawalhi, N.; Randall, R.B. Signal pre-whitening using cepstrum editing (liftering) to enhance fault detection in rolling element bearings. In Proceedings of the 24 International Congress on Condition Monitoring and Diagnostic Engineering Management (COMADEM2011), Stavanger, Norway, 30 May 2011; pp. 330–336. [Google Scholar]
  20. Borghesani, P.; Pennacchi, P.; Randall, R.B.; Sawalhi, N.; Ricci, R. Application of cepstrum pre-whitening for the diagnosis of bearing faults under variable speed conditions. Mech. Syst. Signal Process. 2012, 36, 370–384. [Google Scholar] [CrossRef]
  21. Cascales Fulgencio, D. Frequency_Domain_Features, version 4; FigShare. 2022. Available online: https://figshare.com/articles/software/Frequency_Domain_Features_m/21150967/4 (accessed on 17 September 2022).
  22. Cascales Fulgencio, D. all_12k_sets_table_I, version 1; FigShare. 2022. Available online: https://figshare.com/articles/dataset/all_12k_sets_table_I_mat/21151042/1 (accessed on 17 September 2022).
  23. Sánchez, R.-V.; Lucero, P.; Macancela, J.-C.; Rubio Alonso, H.; Cerrada, M.; Cabrera, D.; Castejón, C. Evaluation of Time and Frequency Condition Indicators from Vibration Signals for Crack Detection in Railway Axles. Appl. Sci. 2020, 10, 4367. [Google Scholar] [CrossRef]
  24. Cascales Fulgencio, D. Time_Domain_Features, version 4; FigShare. 2022. Available online: https://doi.org/10.6084/m9.figshare.21150976.v2. (accessed on 17 September 2022).
  25. Cascales Fulgencio, D. all_12k_sets_table_II, version 1; FigShare. 2022. Available online: https://figshare.com/articles/dataset/all_12k_sets_table_II_mat/21151033/1 (accessed on 17 September 2022).
  26. NIST/SEMATECH. e-Handbook of Statistical Methods. One-Way ANOVA. 2012. Available online: https://www.itl.nist.gov/div898/handbook/ppc/section2/ppc231.htm (accessed on 17 September 2022).
  27. NIST/SEMATECH. e-Handbook of Statistical Methods. Kruskal-Wallis Test. 2012. Available online: https://www.itl.nist.gov/div898/handbook/prc/section4/prc41.htm (accessed on 17 September 2022).
  28. Statistics Online Computational Resource (SOCR). F-Distribution Tables. 2002. Available online: http://www.socr.ucla.edu/Applets.dir/F_Table.html (accessed on 19 September 2022).
  29. Statistics Online Computational Resource (SOCR). Chi-Square Distribution Tables. 2002. Available online: http://www.socr.ucla.edu/Applets.dir/ChiSquareTable.html (accessed on 19 September 2022).
  30. NIST/SEMATECH. e-Handbook of Statistical Methods. Anderson-Darling Test. 2012. Available online: https://www.itl.nist.gov/div898/handbook/eda/section3/eda35e.htm (accessed on 19 September 2022).
  31. NIST/SEMATECH. e-Handbook of Statistical Methods. Levene Test for Equality of Variances. 2012. Available online: https://www.itl.nist.gov/div898/handbook/eda/section3/eda35a.htm (accessed on 19 September 2022).
  32. NIST/SEMATECH. e-Handbook of Statistical Methods. Kolmogorov-Smirnov Goodness-of-Fit Test. 2012. Available online: https://www.itl.nist.gov/div898/handbook/eda/section3/eda35g.htm (accessed on 19 September 2022).
  33. Zhang, W.; Peng, G.; Li, C. Bearings fault diagnosis based on convolutional neural networks with 2-D representation of vibration signals as input. MATEC Web Conf. 2017, 95, 13001. [Google Scholar] [CrossRef]
  34. Cascales Fulgencio, D. Images_Generator, version 4; FigShare. 2022. Available online: https://figshare.com/articles/software/Images_Generator_m/21151021/4 (accessed on 19 September 2022).
  35. Cascales Fulgencio, D. all_12k_sets_table_III, version 1; FigShare. 2022. Available online: https://figshare.com/articles/dataset/all_12k_sets_table_III_mat/21151036/1 (accessed on 19 September 2022).
  36. Cascales Fulgencio, D. TWODCNN, version 4; FigShare. 2022. Available online: https://figshare.com/articles/software/TWODCNN_m/21151018/4 (accessed on 19 September 2022).
  37. Cascales Fulgencio, D. all_images_datastore, version 1; FigShare. 2022. Available online: https://figshare.com/articles/dataset/all_images_datastore_mat/21151045/1 (accessed on 19 September 2022).
  38. Cascales Fulgencio, D. trainClassifier_KNN_A, version 4; FigShare. 2022. Available online: https://figshare.com/articles/software/trainClassifier_KNN_A_m/21150970/4 (accessed on 19 September 2022).
  39. Cascales Fulgencio, D. all_12k_sets_KNN_A_features, version 1; FigShare. 2022. Available online: https://figshare.com/articles/dataset/all_12k_sets_KNN_A_features_mat/21151030/1 (accessed on 19 September 2022).
  40. Cascales Fulgencio, D. trainClassifier_SVM_B, version 4; FigShare. 2022. Available online: https://figshare.com/articles/software/trainClassifier_SVM_B_m/21150973/4 (accessed on 19 September 2022).
  41. Cascales Fulgencio, D. all_12k_sets_SVM_B_features, version 1; FigShare. 2022. Available online: https://figshare.com/articles/dataset/all_12k_sets_SVM_B_features_mat/21151039/1 (accessed on 19 September 2022).
Figure 1. CWRU experimental set-up layout.
Figure 1. CWRU experimental set-up layout.
Applsci 12 10882 g001
Figure 2. Envelope spectrum of a raw time signal. The signal belongs to the CWRU’s bearing database, fan end REB, 0.014′′ depth, centred-to-the-load outer race fault, tested at a motor speed of 1797 rpm.
Figure 2. Envelope spectrum of a raw time signal. The signal belongs to the CWRU’s bearing database, fan end REB, 0.014′′ depth, centred-to-the-load outer race fault, tested at a motor speed of 1797 rpm.
Applsci 12 10882 g002
Figure 3. Envelope spectrum of the signal from Figure 2 after CPW.
Figure 3. Envelope spectrum of the signal from Figure 2 after CPW.
Applsci 12 10882 g003
Figure 4. Greyscale images of drive-end REB signals. Faults located in the rolling elements (a,b), inner race (c) and outer race orthogonal to the load zone (d).
Figure 4. Greyscale images of drive-end REB signals. Faults located in the rolling elements (a,b), inner race (c) and outer race orthogonal to the load zone (d).
Applsci 12 10882 g004aApplsci 12 10882 g004b
Figure 5. Scatter plots of time−domain features. (a) Mean, Standard deviation & Impulse factor. (b) Standard deviation, Peak value & Clearance factor. (c) RMS, Crest factor & Shape factor. (d) Skewness, Clearance factor & Kurtosis.
Figure 5. Scatter plots of time−domain features. (a) Mean, Standard deviation & Impulse factor. (b) Standard deviation, Peak value & Clearance factor. (c) RMS, Crest factor & Shape factor. (d) Skewness, Clearance factor & Kurtosis.
Applsci 12 10882 g005
Figure 6. Scatter plots of frequency−domain features. (a) BPFO Amplitude, BPFI Amplitude & BSF Amplitude. (b) Log(BPFI Amplitude/BPFO Amplitude), Log(BPFI Amplitude/BSF Amplitude) & BPFI Amplitude. (c) Log(BPFI Amplitude/BSF Amplitude), Log(BPFI Amplitude/BPFO Amplitude) & Log(BSF Amplitude/BPFO Amplitude). (d) Log(BPFI Amplitude/BSF Amplitude), Log(BSF Amplitude/BPFO Amplitude) & BSF Amplitude.
Figure 6. Scatter plots of frequency−domain features. (a) BPFO Amplitude, BPFI Amplitude & BSF Amplitude. (b) Log(BPFI Amplitude/BPFO Amplitude), Log(BPFI Amplitude/BSF Amplitude) & BPFI Amplitude. (c) Log(BPFI Amplitude/BSF Amplitude), Log(BPFI Amplitude/BPFO Amplitude) & Log(BSF Amplitude/BPFO Amplitude). (d) Log(BPFI Amplitude/BSF Amplitude), Log(BSF Amplitude/BPFO Amplitude) & BSF Amplitude.
Applsci 12 10882 g006
Figure 7. Bar graphs of cases A & B for both one-way ANOVA and the Kruskal-Wallis test. (a) Dataset of case A ranked according to the one-way ANOVA test. (b) Dataset of case B ranked according to the one-way ANOVA test. (c) Dataset of case A ranked according to the Kruskal-Wallis test. (c) Dataset of case B ranked according to the Kruskal-Wallis test.
Figure 7. Bar graphs of cases A & B for both one-way ANOVA and the Kruskal-Wallis test. (a) Dataset of case A ranked according to the one-way ANOVA test. (b) Dataset of case B ranked according to the one-way ANOVA test. (c) Dataset of case A ranked according to the Kruskal-Wallis test. (c) Dataset of case B ranked according to the Kruskal-Wallis test.
Applsci 12 10882 g007
Figure 8. ML algorithms performance. Features vectors’ dimensionality increased following one-way ANOVA order. Plots (a,c,e,g) for case A and plots (b,d,f,h) for case B.
Figure 8. ML algorithms performance. Features vectors’ dimensionality increased following one-way ANOVA order. Plots (a,c,e,g) for case A and plots (b,d,f,h) for case B.
Applsci 12 10882 g008
Figure 9. ML algorithms performance. Features vectors’ dimensionality increased following the Kruskal–Wallis test’s order. Plots (a,c,e,g) for case A and plots (b,d,f,h) for case B.
Figure 9. ML algorithms performance. Features vectors’ dimensionality increased following the Kruskal–Wallis test’s order. Plots (a,c,e,g) for case A and plots (b,d,f,h) for case B.
Applsci 12 10882 g009
Figure 10. All ML algorithms’ performance. Features vectors’ dimensionality increased following one-way ANOVA order.
Figure 10. All ML algorithms’ performance. Features vectors’ dimensionality increased following one-way ANOVA order.
Applsci 12 10882 g010
Figure 11. All ML algorithms performance. Features vectors’ dimensionality increased following the Kruskal–Wallis test’s order.
Figure 11. All ML algorithms performance. Features vectors’ dimensionality increased following the Kruskal–Wallis test’s order.
Applsci 12 10882 g011
Figure 12. Summary of results for the ML and DL models.
Figure 12. Summary of results for the ML and DL models.
Applsci 12 10882 g012
Table 1. 12k-sampled REBs signals. Health conditions and class labels.
Table 1. 12k-sampled REBs signals. Health conditions and class labels.
Health ConditionTotal DatasetClass Labels
Normal80
Inner race fault761
Ball fault762
Outer race fault1473
Table 2. Characteristic frequencies of REB faults.
Table 2. Characteristic frequencies of REB faults.
Fault DescriptionCharacteristic FrequencyFault Location
BPFO f r   ·   N r 2   ·   1 d B   ·   cos θ d P Outer race
BPFI f r   ·   N r 2   ·   1 + d B   ·   cos θ d P Inner race
BSF f r   ·   d P d B   ·   1 d B 2   ·   cos 2 θ d P 2 Rolling element
Where, fr is Rotational frequency of the shaft [Hz]. dB is Rolling element diameter [m]. dP is Pitch diameter [m]. Nr is Number of rolling elements (for a single row) [-]. θ is Load angle (contact angle from radial) [rad].
Table 3. Hyper-parameters of the proposed CNN.
Table 3. Hyper-parameters of the proposed CNN.
KernelStridePadding
HeightWidthVertical Step SizeHorizontal Step SizeTopBottomLeftRight
Convolutional Layers5511aut.aut.aut.aut.
Pooling layers22220000
Table 4. Features sorted by importance according to one-way ANOVA. Case A.
Table 4. Features sorted by importance according to one-way ANOVA. Case A.
FeaturesF
LOG_BPFI_Amplitude_BPFO_Amplitude258.3292
LOG_BSF_Amplitude_BPFO_Amplitude122.8081
LOG_BPFI_Amplitude_BSF_Amplitude116.2330
BPFI_Amplitude92.2322
BPFO_Amplitude77.7082
BSF_Amplitude9.5288
Crest factor6.0064
Shape factor5.8229
Impulse factor3.4917
Clearance factor2.7894
Kurtosis1.9256
Mean1.4055
Peak value1.3931
Standard deviation0.9941
RMS0.9637
Skewness0.6254
Table 5. Features sorted by importance according to one-way ANOVA. Case B.
Table 5. Features sorted by importance according to one-way ANOVA. Case B.
FeaturesF
LOG_BPFI_Amplitude_BPFO_Amplitude289.5471
LOG_BPFI_Amplitude_BSF_Amplitude129.0033
BPFI_Amplitude95.1479
LOG_BSF_Amplitude_BPFO_Amplitude90.1346
BPFO_Amplitude61.6727
BSF_Amplitude6.0888
Shape factor4.0389
Clearance factor2.9775
Kurtosis2.8668
Standard deviation2.7570
Impulse factor2.6555
RMS2.6403
Peak value2.5531
Crest factor1.8718
Mean1.2457
Skewness0.7408
Table 6. Features sorted by importance according to the Kruskal–Wallis test. Case A.
Table 6. Features sorted by importance according to the Kruskal–Wallis test. Case A.
FeaturesH
LOG_BPFI_Amplitude_BPFO_Amplitude217.3216
BPFI_Amplitude167.1512
LOG_BSF_Amplitude_BPFO_Amplitude155.3533
BPFO_Amplitude138.4140
LOG_BPFI_Amplitude_BSF_Amplitude120.2450
BSF_Amplitude43.3603
Standard deviation26.0577
RMS25.2882
Shape factor17.9515
Peak value15.7401
Mean12.3161
Kurtosis11.9904
Clearance factor5.3638
Impulse factor5.1157
Crest factor4.8157
Skewness3.0968
Table 7. Features sorted by importance according to the Kruskal–Wallis test. Case B.
Table 7. Features sorted by importance according to the Kruskal–Wallis test. Case B.
FeaturesH
LOG_BPFI_Amplitude_BPFO_Amplitude152.5596
BPFI_Amplitude145.4471
LOG_BPFI_Amplitude_BSF_Amplitude100.1278
BPFO_Amplitude97.8711
LOG_BSF_Amplitude_BPFO_Amplitude91.5079
BSF_Amplitude24.0144
Standard deviation18.0231
RMS17.5023
Peak value15.3398
Mean11.6661
Kurtosis10.9086
Shape factor10.1311
Clearance factor5.9409
Impulse factor5.4390
Crest factor3.8549
Skewness2.7573
Table 8. ML algorithms performance. Features vector’s dimensionality increased following one-way ANOVA order. Case A.
Table 8. ML algorithms performance. Features vector’s dimensionality increased following one-way ANOVA order. Case A.
Features Vector’s DimensionalityDTSVMk-NNNaïve Bayes
10.7880.8050.7850.805
20.8080.8240.8270.821
30.7650.8180.8010.808
40.8240.8400.8210.834
50.8210.8310.8180.831
60.830.8370.8400.837
70.8210.8240.8240.844
80.8080.8400.8270.840
90.8210.8080.8210.840
100.8210.8410.8470.840
110.8110.8310.8240.847
120.8270.8340.8240.847
130.8140.8240.8270.837
140.8310.8310.8500.847
150.8270.8470.8660.837
160.8210.8210.8760.840
The best result obtained for each model by applying the proposed method has been marked in bold.
Table 9. ML algorithms performance. Features vector’s dimensionality increased following one-way ANOVA order. Case B.
Table 9. ML algorithms performance. Features vector’s dimensionality increased following one-way ANOVA order. Case B.
Features Vector’s DimensionalityDTSVMk-NNNaïve Bayes
10.8960.9090.9220.905
20.9000.9090.9000.887
30.9390.9440.9260.931
40.9350.9390.9130.935
50.9440.9390.9200.944
60.9480.9480.9570.948
70.9520.9520.9570.935
80.9570.9520.9570.939
90.9440.9480.9520.935
100.9480.9650.9610.944
110.9480.9650.9650.935
120.9520.9740.9610.935
130.9390.9700.9650.948
140.9480.9610.9650.944
150.9440.9610.9610.939
160.9480.9480.9610.939
The best result obtained for each model by applying the proposed method has been marked in bold.
Table 10. ML algorithms performance. Features vector’s dimensionality increased following the Kruskal–Wallis test’s order. Case A.
Table 10. ML algorithms performance. Features vector’s dimensionality increased following the Kruskal–Wallis test’s order. Case A.
Features Vectors’ DimensionalityDTSVMk-NNNaïve Bayes
10.7880.8050.7850.805
20.8080.8180.7850.814
30.8180.8310.8180.831
40.8240.8340.8210.831
50.8340.8400.8140.834
60.8340.8370.8400.837
70.8210.8370.8210.837
80.8310.8370.8240.834
90.7980.8110.8270.834
100.8500.8270.8340.840
110.8240.8340.8240.840
120.7950.8180.8310.834
130.8370.8340.8700.847
140.8240.8370.8370.834
150.8270.8470.8660.837
160.8210.8210.8760.840
The best result obtained for each model by applying the proposed method has been marked in bold.
Table 11. ML algorithms performance. Features vectors’ dimensionality increased following the Kruskal–Wallis test’s order. Case B.
Table 11. ML algorithms performance. Features vectors’ dimensionality increased following the Kruskal–Wallis test’s order. Case B.
Features Vector’s DimensionalityDTSVMk-NNNaïve Bayes
10.8960.9090.9220.905
20.9480.9480.9390.935
30.9390.9440.9260.931
40.9480.9440.9260.948
50.9440.9390.9200.944
60.9480.9480.9570.948
70.9570.9700.9740.939
80.9480.9740.9740.957
90.9520.9740.9740.952
100.9520.9700.9700.944
110.9440.9700.9740.952
120.9520.9650.9650.935
130.9520.9780.9700.948
140.9520.9740.9700.939
150.9440.9610.9610.939
160.9480.9480.9610.939
The best result obtained for each model by applying the proposed method has been marked in bold.
Table 12. CNN results for cases A and B.
Table 12. CNN results for cases A and B.
Case ACase B
Validation accuracy0.77610.9067
Elapsed time5 min 14 s5 min 44 s
Epochs1515
Maximum iterations50403765
Iterations per epoch336253
Frequency50 iterations50 iterations
Learning rate0.010.01
Table 13. Results comparison.
Table 13. Results comparison.
Ref.MethodRMSERef.MethodAccuracy
[7]SVM model0.0521[11]CNN (Greyscale Images Training Set)0.987
SVM Markov-model0.0091
Ref.MethodAccuracyCNN (Greyscale Images Validation Set)0.997
[8]Bayesian80.55–91.67%
NB71.67–91.11%CNN (Greyscale Images Test Set)0.989
MLP87.22–96.11%
k-NN87.78–92.77%Ref.MethodAccuracy
Weighted k-NN91.67–96.11%[13]KNN + PCAVibration Image Method51.42%
J4889.44–90.55%Wavelet Spectrogram48.99%
Random Forest90.56–95.55%DSWI61.13%
Ref.MethodAccuracyMCSVM + PCAVibration Image Method80.97%
[9]Constant torque loadSMO80–100%Wavelet Spectrogram81.91%
ASMO81–100%DSWI87.76%
AFSMO84–100%LeNet-5Vibration Image Method46.77%
Variable torque loadSMO75–100%Wavelet Spectrogram89.11%
ASMO81–100%DSWI95.97%
AFSMO75–100%AlexNetVibration Image Method68.54%
Ref.MethodAccuracyWavelet Spectrogram90.72%
[10]FS1 + SVM75.00% & 73.57%DSWI97.98%
FS2 + SVM92.50% & 78.81%DCNNVibration Image Method83.06%
LiftingNet99.63% & 93.19%Wavelet Spectrogram93.15%
DSWI98.79%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Cascales-Fulgencio, D.; Quiles-Cucarella, E.; García-Moreno, E. Computation and Statistical Analysis of Bearings’ Time- and Frequency-Domain Features Enhanced Using Cepstrum Pre-Whitening: A ML- and DL-Based Classification. Appl. Sci. 2022, 12, 10882. https://doi.org/10.3390/app122110882

AMA Style

Cascales-Fulgencio D, Quiles-Cucarella E, García-Moreno E. Computation and Statistical Analysis of Bearings’ Time- and Frequency-Domain Features Enhanced Using Cepstrum Pre-Whitening: A ML- and DL-Based Classification. Applied Sciences. 2022; 12(21):10882. https://doi.org/10.3390/app122110882

Chicago/Turabian Style

Cascales-Fulgencio, David, Eduardo Quiles-Cucarella, and Emilio García-Moreno. 2022. "Computation and Statistical Analysis of Bearings’ Time- and Frequency-Domain Features Enhanced Using Cepstrum Pre-Whitening: A ML- and DL-Based Classification" Applied Sciences 12, no. 21: 10882. https://doi.org/10.3390/app122110882

APA Style

Cascales-Fulgencio, D., Quiles-Cucarella, E., & García-Moreno, E. (2022). Computation and Statistical Analysis of Bearings’ Time- and Frequency-Domain Features Enhanced Using Cepstrum Pre-Whitening: A ML- and DL-Based Classification. Applied Sciences, 12(21), 10882. https://doi.org/10.3390/app122110882

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop