Next Article in Journal
Code Integrity and Confidentiality: An Active Data Approach for Active and Healthy Ageing
Previous Article in Journal
Image Generation and Recognition for Railway Surface Defect Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Feature Selection and Multi-Label Driven Intelligent Fault Diagnosis Method for Gearbox

College of Intelligent Manufacturing and Industrial Modernization, Xinjiang University, Urumchi 830017, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(10), 4792; https://doi.org/10.3390/s23104792
Submission received: 25 April 2023 / Revised: 11 May 2023 / Accepted: 13 May 2023 / Published: 16 May 2023
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

:
Gearboxes are utilized in practically all complicated machinery equipment because they have great transmission accuracy and load capacities, so their failure frequently results in significant financial losses. The classification of high-dimensional data remains a difficult topic despite the fact that numerous data-driven intelligent diagnosis approaches have been suggested and employed for compound fault diagnosis in recent years with successful outcomes. In order to achieve the best diagnostic performance as the ultimate objective, a feature selection and fault decoupling framework is proposed in this paper. That is based on multi-label K-nearest neighbors (ML-kNN) as classifiers and can automatically determine the optimal subset from the original high-dimensional feature set. The proposed feature selection method is a hybrid framework that can be divided into three stages. The Fisher score, information gain, and Pearson’s correlation coefficient are three filter models that are used in the first stage to pre-rank candidate features. In the second stage, a weighting scheme based on the weighted average method is proposed to fuse the pre-ranking results obtained in the first stage and optimize the weights using a genetic algorithm to re-rank the features. The optimal subset is automatically and iteratively found in the third stage using three heuristic strategies, including binary search, sequential forward search, and sequential backward search. The method takes into account the consideration of feature irrelevance, redundancy and inter-feature interaction in the selection process, and the selected optimal subsets have better diagnostic performance. In two gearbox compound fault datasets, ML-kNN performs exceptionally well using the optimal subset with subset accuracy of 96.22% and 100%. The experimental findings demonstrate the effectiveness of the proposed method in predicting various labels for compound fault samples to identify and decouple compound faults. The proposed method performs better in terms of classification accuracy and optimal subset dimensionality when compared to other existing methods.

1. Introduction

Gearboxes are widely used as important power transmission components in industrial systems; however, due to the harsh working environment and the combined effect of multiple sources of excitation, gearbox failures are inevitable. In the actual environment, a failure often does not occur alone but in the same component or different components at the same time as multiple independent single failures, which constitutes a compound failure. For the whole mechanical system, the destructiveness of compound faults is much greater than that of single faults [1], so it is important to make an accurate diagnosis of compound faults in gearboxes. Moreover, when one component of a gearbox fails, it affects the vibration characteristics of another component [2], which makes it possible that considering only a single component may lead to erroneous results, so it is extremely important to monitor the operating conditions of each component in the gearbox simultaneously.
Compound faults have made fault diagnosis and predictive maintenance of mechanical equipment dramatically more difficult, and decoupling compound failures remains a major challenge today [3]. In recent years, data-driven approaches have provided a promising approach for condition monitoring and fault diagnosis, where various data collected from the system are used to train artificial intelligence-based predictive models [4]. Since vibration signals are most sensitive to changes in the implemented operating state of the equipment, and the signal features have a strong correspondence with the fault state of each component inside the equipment, they are often used as a carrier of fault information [5]. As the equipment structure becomes more complex and the phenomenon of coupling multiple fault information becomes more tricky, the requirements for diagnosing faults are getting higher and higher, and obtaining higher recognition accuracy and reducing the training cost of the model are constantly pursued goals. In data-driven fault diagnosis methods, the selection of the most representative features is crucial to the classification performance of the prediction model, so we intend to propose a method that automatically selects the optimal subset optimized for the classifier from the original high-dimensional feature set, with the ultimate goal of achieving decoupling of compound faults and obtaining the highest recognition accuracy.

2. Motivation and Literature Review

As mentioned in the previous section, data-driven intelligent fault diagnosis methods have considerable potential, especially the introduction of computer theoretical technical disciplines such as machine learning (ML) and deep learning (DL) has had a rather positive impact, and many studies have shown the effectiveness of using them as classifiers [6,7]. However, the lack of interpretability, high computational cost and high dependence on tuning parameters make DL less practical and reliable than ML in most scenarios [8]. Several ML-based fault classification models [9,10,11] have demonstrated the effectiveness of these techniques in fault diagnosis.
Data-driven fault diagnosis algorithms are usually composed of three parts: feature extraction, feature selection, and fault classification [12]. Among them, the set of fault features constructed during feature extraction is crucial to achieving an accurate diagnosis. Since the collected vibration signals are often a mixture of signals from different components, this leads to the fact that the fault type will become difficult to diagnose when different components fail at the same time, influenced by the coupling of fault features with each other [13]. Coupled with the complexity of the compound fault mechanism, it is difficult for a single feature or single-domain feature to accurately describe the changes in the operating state of the equipment, while the construction of a multi-domain feature set can comprehensively characterize the changing trend of the equipment fault state from multiple feature spaces. Some existing research works have shown that multi-domain feature sets have better performance in fault diagnosis compared to constructing single-domain feature sets [14]. Therefore, in order to effectively characterize the gearbox operating state, it has become a common practice to extract a large number of representative features, such as time domain and frequency domain features from the signal, and then construct a feature set that contains as much fault information as possible, which is widely used by researchers in the field of fault diagnosis.
Although constructing a multi-domain feature set that is as comprehensive as possible can effectively overcome these problems, it will also inevitably lead to a feature set that will contain some redundant features or irrelevant features [15], which may reduce computational efficiency and diagnostic accuracy. Additionally, a fault feature set with an excessive amount of dimensionality will also pose a severe test for the subsequent steps. Therefore, dimensionality reduction techniques are used to obtain an optimal subset from the original feature set with moderate dimensionality and eliminate irrelevant and redundant features. There are two major types of mainstream dimensionality reduction techniques that are often applied in fault feature selection. One of them is called feature reduction, which maps the feature set from a high-dimensional feature space to a low-dimensional feature space and reduces the dimensionality by constructing new features; the typical methods are principal component analysis (PCA) [16], linear discriminant analysis (LDA) [17], etc.
However, the physical meaning of new features generated by feature reduction methods is far from the original features, blurring the information of the original features or even the opposite. In contrast, feature selection can retain the original physical meaning of fault features by selecting the most characterized sensitive features from the original high-dimensional feature set without changing the original feature structure [18], which has good interpretability, and this physical meaning can be used to analyze the formation mechanism of faults and the design of suitable classifiers [19]. Feature selection algorithms can be classified as Filter, Wrapper, and Hybrid based on the degree of dependence on the classifier in the selection process [20,21,22]. The Filter algorithm, which is separate from the classifier and is typically used as a preprocessing step to provide the classifier with a reduced feature set, can quickly eliminate a large number of irrelevant features from the original set [23,24]. The Wrapper algorithm combines with the classifier in the selection process, uses the classification accuracy as the evaluation criterion to measure the quality of the feature subset, and the final retained feature subset has very good classification performance [9,25]. In fact, any search strategy can be used as a guide to the search direction of the Wrapper algorithm, so the search process has a strong generalization capability, but it also results in an extremely high time complexity.
As mentioned above, the benefits and drawbacks of the Filter and Wrapper algorithms appear to be complementary. Therefore it is an effective strategy to use the ranking information obtained by the Filter algorithm to guide the search direction of the Wrapper method [26] and balance the classification accuracy and computational cost; most of the existing hybrid feature selection algorithms are designed as a two-stage framework. Zhang et al. [5] and Liang et al. [27] first used the Filter method to obtain a pre-selected subset and then further searched for optimal features using a genetic algorithm and particle swarm optimization, respectively. Mochammad et al. [28] first selected five features from each of the original feature sets using three different Filter models and then searched for the optimal subset by exhausting the combinations among these features. Ganjie et al. [29] first ranked the features according to their relevance to class labels, then applied different clustering methods to divide them into multiple subsets, and finally obtained the optimal subset by traversing all subsets. Chen et al. [11] applied the evolutionary multitasking technique for feature selection by estimating weights for features by Relief-F in the first stage and dividing the feature set into two tasks based on the weights. In the second stage, the optimal subset is obtained by knowledge transfer between two tasks and using particle swarm optimization.
The overall objective of this paper is to develop an efficient feature selection and fault decoupling framework. In the original feature extraction stage of the proposed intelligent diagnosis method, the effect of noise is diminished through pre-processing, and the original high-dimensional feature set is constructed by extracting statistical features in the time domain and frequency domain. Then, the Filter model is used for feature pre-ranking, the weighting scheme is used for feature re-ranking, and the Wrapper model is used for correctly pinpointing the optimal subset automatically in the proposed three-stage feature selection framework. Finally, the decoupling of compound faults is accomplished using ML-kNN as a classifier. Specifically, this study contributes the following:
  • A feature set construction method that efficiently handles a low signal-to-noise ratio and nonlinear signals using statistical feature extraction techniques;
  • A feature selection framework that is applicable to multi-label data and can automatically search for the optimal subset from the original high-dimensional feature set;
  • A physical interpretative fault diagnosis framework that can maximize the preservation of fault features.
The remainder of this paper is organized as follows. The theoretical basis of the main components of the proposed method is briefly covered in Section 3. Section 4 provides a detailed description of the proposed method. In Section 5, the method is applied to real data examples of bearing-gear compound fault detection and compared with some classical methods. Finally, Section 6 draws conclusions and discusses possible directions for future work.

3. Theoretical Basis of the Study

This section introduces the theoretical background of the feature selection model and multi-label learning technology used in the research.

3.1. Feature Selection Models

3.1.1. Filter Models

In the Filter method, commonly used evaluation criteria include distance measure, information measure and correlation measure. For example, the Fischer score method emphasizes the distance between labels and is based on distance measurement. The information gain method assesses features from the viewpoint of information and is based on information measures. Based on correlation measures, the Pearson correlation coefficient method emphasizes the correlation between feature and label.
  • Fischer score:
In the Fisher score (FS), each feature is selected independently according to its Fisher criterion score [30], and the aim is to find features that satisfy the distance that maximizes the inter-class distance and minimizes the intra-class distance. Assume that there are n samples in the feature set belonging to Y labels, each of which contains nk samples. Define the inter-lass scatter Sb(fi) of the ith feature fi as:
S b ( f i ) = k = 1 Y n k ( m i k m k ) 2
where nk represents the number of samples in the kth category, mki represents the mean of the values taken by the samples in the kth category on the ith feature, and mk represents the mean of the values taken by the samples in all categories on the ith feature.
Define the intra-class scatter Skt(fi) of the kth sample at the ith feature fi as:
S t k ( f i ) = j = 1 n k ( x i , j k m i k ) 2
where xki,j is the value of the jth sample that belongs to the kth class of samples in the ith feature.
When the inter-class scatter Sb(fi) of the ith feature is larger, and the intra-class scatter Skt(fi) is smaller, the stronger the characterization ability of the feature is, and the calculation formula for FS can be obtained as follows:
F s ( f i ) = S b ( f i ) k = 1 C n k S t k ( f i ) = k = 1 C n k ( m i k m k ) 2 k = 1 C [ n k j = 1 n k ( x i , j k m i k ) 2 ]
Following FS evaluation, the smaller the obtained score, the less significant the features are, whereas the larger the obtained score, the higher the feature importance and the stronger the characterization ability. However, the Fisher score does not measure the correlation degree between features, so it cannot eliminate the redundant features in the original set;
  • Information gain:
In the 1940s, Shannon first introduced the concept of information entropy and provided a mathematical expression to calculate it [31], which allowed the value of information to be expressed quantitatively and provided a standard by which the value of information could be determined. For a label Y = {y1, y2, ⋯, yn} in a classification task, and assuming that the probability corresponding to the value taken in the label is P(yi), the information entropy that this label has is given by the following equation:
H ( Y ) = i n P ( y i ) log 2 ( P ( y i ) )
From Equation (4), it can be seen that the greater the uncertainty of the label Y, the greater the entropy, implying that more information is needed to determine the label as well. Assuming that a feature in the feature set can be represented as X = {x1, x2, ⋯, xn}, the conditional entropy that X has is given by the following equation:
H ( Y | X ) = j = 1 n P ( g j ) i = 1 l P ( y i | x i ) log 2 ( P ( y i | x i )
P(yi) in Equation (4) represents the prior probability of label Y, and P(yi|xi) in Equation (5) represents the conditional probability of label Y when feature X is determined. Information gain (IG) is a measure of the difference between the prior uncertainty of a feature and the expected posterior uncertainty, and for classification tasks, information gain is calculated by first calculating the feature’s occurrence in each class number of occurrences of the feature in each class, which is then used to describe the information gain of the feature for each class [32]. Given the feature x and its corresponding label y, the IG is calculated as follows:
I ( x ; y ) = H ( y ) H ( y | x )
From the definition, when H(y) − H(y|x) = 0, it means that the label y is completely irrelevant to the feature x, while when I(x; y) > 0, it means that the label y is related to the feature x. And the larger I(x; y) is, the closer the association between the label and the feature; that is, the greater the information gain of the feature, the better its characterization ability and the more critical for fault classification;
  • Pearson correlation coefficient:
Pearson correlation coefficient (PCC) is used to measure the degree of strength of linear correlation between variables [33] and can also be used to measure the correlation between features and labels. A feature with excellent representational power has a strong correlation with the label, while the features should be uncorrelated with each other. The PCC between two variables is defined as the quotient of their covariance and standard deviation:
ρ ( x , y ) = cov ( x , y ) σ x σ y = E [ ( x μ x ) ( y μ y ) ] σ x σ y
Equation (7) is commonly used to predict continuous outcomes, and when generalized to the multi-classification case, given a feature x and its corresponding label y, the PCC is calculated as follows:
ρ ( x ) = i = 1 m ( x i x ¯ ) ( y i y ¯ ) i = 1 m ( x i x ¯ ) 2 i = 1 m ( y i y ¯ ) 2
Obviously, the larger the value of ρ, the stronger the correlation between features and labels, and the greater the gain that can be provided for the fault classification task.

3.1.2. Wrapper Models

In general, there are three basic search strategies in the Wrapper method: global search strategy, sequential search strategy, and random search strategy [34,35]. The goal of the global search strategy is finding the optimal global subset of the original feature set, but this goal can only be guaranteed when almost all permutations have been exhausted, which is obviously extremely unrealistic for high-dimensional sets. Sequential search strategy and random search strategy are collectively referred to as heuristic search strategies, which add constraints to the search process to guide the search in the optimal direction. As a result, they have lower time complexity than the global search strategy, but the search results are usually inferior to global search. However, this sacrifice is worthwhile considering the amount of computation that can be saved by applying heuristic search strategies.
  • Binary search:
Binary search (BS), also known as half search, starts from the middle element when searching and stops searching if the middle element meets the search requirement; otherwise, it continues from the middle element and searches until it finds the solution. The search process is as follows: the original feature set is split into two roughly equal parts, the left subset and the right subset. If the classification error rate of the left subset is below a threshold, the left subset is retained for further search and the right subset is deleted, and vice versa. Obviously, each search can narrow the search by half. Until the next search, the iterative process stops when the error rates of both the left and right subsets are higher than the set threshold.
BS is a very effective search strategy, but the prerequisite for it to give full play to its advantages is that the search object must be a set that conforms to the sequential structure. That is, each candidate element in the original set should be ordered according to some rules; otherwise, the search process will be meaningless;
  • Sequential forward search:
Sequential forward search (SFS) is a bottom-up search strategy similar to the enumeration method. After selecting an optimal feature, which is the most beneficial one for the classification task, each feature selected thereafter is the feature that performs best when combined with the selected features. The specific search process of SFS is to add one feature at a time to the initial subset, and the iterative process stops when the classification accuracy does not improve with the increase of features;
  • Sequential backward search:
Sequential backward search (SBS) is a top-down search strategy, and its search strategy is the opposite of SFS, which starts from all features and eliminates one feature at a time and can make the remaining features perform optimally when combined. The specific search process is to reduce one feature at a time in the initial subset, and the iterative process stops when the classification accuracy does not improve with the reduction of features.

3.2. Multi-Label Classification

Unlike traditional machine learning methods, the samples processed in the multi-label learning task contain multiple labels at the same time, which is more in line with the scenarios faced during real-world applications. More importantly, treating compound faults as an independent fault type means that a completely new label needs to be assigned to each compound fault state. The number of combinations formed will increase exponentially as the number of fault categories or the degree of fault changes, which makes it extremely impractical to obtain sufficient samples for each fault state, and the difficulty of fault classification will increase significantly. The number of labels required for compound fault diagnosis will be drastically reduced in multi-label learning, on the other hand, because the compound fault states will be simultaneously classified into multiple labels that correspond to a single fault during diagnosis and will no longer need additional labels.
Existing multi-label learning methods can be classified into two main categories: problem transformation (PT) and algorithm adaption (AA). In brief, PT algorithms make multi-label data adaptive to existing algorithms, while AA algorithms are adaptive to multi-label data [36]. In other words, the PT method tends to convert multi-label data into single-label data, which in turn converts the multi-label classification problem into multiple simple single-label classification problems. The AA method, on the other hand, tends to improve existing single-label classification algorithms so that these algorithms can directly deal with multi-label data.
The effectiveness of converting multi-label data into single-label data by the PT method and then performing feature selection has been demonstrated by a large number of researchers, such as Oscar et al. [37] used the PT method to convert multi-label problems into single-label problems and executed Relief-F algorithm on the single-label dataset converted by PT method. Newton et al. [38] combined two PT methods and two Filter molders to obtain four multi-label feature selection methods. Through numerous sets of experiments, they showed how effectively the single-label feature selection algorithm worked with the transformed dataset. Therefore, we intend to first convert the original multi-label data into single-label data by the PT method, then execute the feature selection method, and finally achieve the diagnosis and decoupling of compound faults by the AA classifier.

3.2.1. Label Powerset

Label powerset (LP) [39] is a straightforward and effective PT method that converts an L-class multi-label problem into a multi-class single-label problem with 2|L|-class labels. Its processing can be briefly summarized as treating the different combinations of labels present in the training set as a new class label, thus converting the multi-label data into single-label data and taking the correlation between the labels into account during the conversion process [37,40].

3.2.2. Multi-Label k-Nearest Neighbor Classifier

ML-kNN is a classical AA method [41], which is derived from the kNN algorithm. The fundamental procedure can be succinctly stated as follows: The k-nearest neighbor in the training set is initially determined for each test sample. The label set of the test instance is then determined using the maximum a posteriori probability principle based on the statistics obtained from the label set of these adjacent instances, that is, the number of adjacent instances belonging to each possible category. Since ML-kNN determines the k-nearest neighbor of the test sample through the distance between the test sample and all the samples in the training set and then determines its category, without a complicated mapping process, it is explainable with clear physical meaning.

4. The Proposed Methodology

4.1. An Overview of the Proposed Feature Selection and Fault Decoupling Method

The flowchart of the proposed feature selection scheme and fault decoupling procedure is shown in Figure 1. The proposed feature selection and fault decoupling method includes data extraction and feature extraction, feature ranking and selection, and fault decoupling and diagnosis, as described below.
  • Data acquisition and feature extraction: A data acquisition system is used to obtain the gearbox’s vibration signal. In order to more effectively remove disturbing components from the signal and highlight fault characteristic information, the raw signal is first pre-processed and then time domain statistical features and frequency domain statistical features are extracted from the raw and pre-processed signals, as shown in Section 4.2;
  • Feature ranking and selection: In this paper, we will use the LP method to transform the multi-label data and perform feature selection before utilizing ML-kNN as a classifier to evaluate the effectiveness of feature selection. In other words, the PT method is merely utilized to achieve feature selection, and once feature selection for the original feature set has been finished, the original multi-label problem will be taken into account once more to achieve multi-label classification using the AA method. As a result, our proposed processing method consists of 3 primary steps, as shown in Figure 2. The LP method is used in step 1 to transform the multi-label set L into ingle-label set Ls. The proposed feature selection method, which is illustrated in Section 4.3, is used to search the most sensitive features from the origin feature set in step 2. In step 3, the quality of the optimal subset Lm is evaluated by ML-kNN.
  • Fault decoupling and diagnosis: Using the index of the optimal features obtained through the training set, the corresponding optimal features are selected from the original high-dimensional feature set of the testing set. Finally, the fault decoupling results of the testing set are obtained by the trained ML-kNN classifier, which is trained by the training set.

4.2. Original Feature Extraction of Vibration Signals

The obtained vibration signal is generally non-smooth and non-linear and has a low signal-to-noise ratio since it is restricted by the actual environment. This raises the difficulty of feature extraction and makes it more difficult to extract fault information from the vibration signal. It can often be difficult to adequately describe the fault state of the equipment by extracting only single or single domain features due to the complicated coupling behavior in the compound fault. Therefore, in this paper, the Teager energy operator and empirical modal decomposition (EMD) algorithm will be used to preprocess the signal and then extract the statistical features from the vibration signal, and the process is shown in Figure 3.
Specifically, to emphasize the transient components of the signal, the paper first demodulates the original signal using the Teager energy operator [42]. Then the vibration signal is divided into numerous intrinsic mode functions (IMFs) via the EMD algorithm. These IMFs contain local characteristic signals of the original signal at different time scales [43], and each component exhibits singularity and reflects a piece of unique frequency information in the signal. The first few orders of IMF that contain almost all the information are selected, and then time domain features and frequency domain features are extracted from them to construct multi-domain feature sets. These features are summarized in Table A1, denoted as T1–T11 and F1–F13, respectively, and more information on their physical significance can be obtained in the literature [44]. In order to obtain more fault information, features will be extracted from both the original and demodulated signals. As a result, for each signal sample, 24 × 2 × (M + 1) features will be extracted, where M is the number of chosen IMFs. In this paper, the first 5 IMFs will be selected for analysis, so the original feature vector of each signal will contain 288 features.

4.3. The Proposed 3-Stage Hybrid Feature Selection Framework

Filter models evaluate features from different perspectives, measure the possible characterization ability under a single measure, and can be effective for rapid feature ranking. However, due to different evaluation perspectives, the outcomes of various Filter models may be highly dissimilar. The same feature may be rejected as irrelevant in 1 model while accepted as sensitive in another. In order to provide pre-ranking results from various perspectives, this paper uses the FS, IG, and PCC methods in the first stage to evaluate features from the perspectives of distance, information, and relevance, respectively.
However, these evaluation results alone cannot eliminate redundant features. When feature a and feature b, for instance, both have high Fisher scores and a strong correlation, removing 1 of them won’t have an impact on the subsequent training process and the learning performance won’t be lost, but the FS model will keep both of them, leaving the feature set with redundant features. The same is true of PCC and IG models, but the evaluation results of PCC and IG methods can effectively exclude redundant features while being relative to select sensitive features, so these 3 sets of assessment findings can each provide complementary information to 1 another. Therefore, in order to reduce the number of redundant features and eliminate irrelevant features from the original feature set, as well as to minimize the dimensionality of the original set and maintain the capacity for classification, it is possible to combine the pre-ranking results from the 3 models. In addition, the sole use of the Filter model also ignores the potential gain in characterization ability that results from the combination of features. For instance, whereas features a and b individually have extremely low scores, combining them results in potentially very high scores, indicating that using features a and b together can yield significantly greater benefits for fault classification tasks. Filter models will, however, eliminate features a and b because they only calculate the scores of each feature individually without taking into account the combined impact between features, which plainly results in a waste of learning performance.
In order to address the aforementioned 2 shortcomings, a method is required to synthesize the assessment findings of the 3 Filter models. This is because the characterization ability of features can only be better compared when they are utilized in combination. In the weighting scheme proposed by ZHANG et al. [45], the accuracy of the classifier is used as a weight, and the pre-ranking results of the Filter model are combined to synthesize a new fusion evaluation result. The scheme first chooses the top k features in the pre-ranking result, then inputs a subset consisting of these k features into the classifier for training. The score for the corresponding feature is calculated as the product of the obtained classification accuracy and the feature’s ranking result, and the weighted score for the feature is calculated as the cumulative sum of multiple Filter models, as given in Equation (9).
{ S c o r e i = R a n k i n g i × A c c u r a c y i N e w S c o r e = i = 1 3 S c o r e i
where Scorei represents the score of the ith Filter method on the features, i denotes the FS, PCC and IG models, respectively, and NewScore represents the weighted score.
This scheme achieves impressive performance on single-label data but is not applicable to the case of multi-label data. This is because, in single-label learning, the labels are only predicted correctly or incorrectly. However, in the case of multi-label learning, it is possible that some of the predicted labels will match the real ones while others won’t. Therefore, it is also necessary to account for this partially correct scenario using a revised fitness function in this paper, as illustrated in Equation (10).
F i t = ( 1 + β 2 ) × S A × A P β 2 × S A + A P
where SA represents severe accuracy, and AP represents average accuracy. It is clear that the influence of SA on the fitness function dominates when β is less than 1, while AP dominates when β is greater than 1.
Additionally, the scheme depicted in Equation (9) simply merges the evaluation results of Filter models without realizing that the evaluation results of different Filter models do not contribute exactly to the classification task and overly idealizes the weight of Scorei as equal. As a result, this paper applies integration techniques to combine the 3 pre-ranking outcomes. The weighted average method, which obtains the fusion results by performing a weighted average for a set of data, is the most common and straightforward integration technique. Therefore, this paper will combine the evaluation results of the 3 Filter models by assigning appropriate weights to Scorei through the weighted average method on the basis of Equation (9), as indicated in Equation (11).
{ F u s i o n S c o r e = i = 1 3 ω i S c o r e i i = 1 3 ω i = 1
where Scorei represents the score of the ith Filter method on the features, ωi ≥ 0, and FusionScore represents the weighted fusion score.
Obviously, there are an unlimited number of combinations of weights ωi, which will be solved by genetic algorithm in this paper. The weights ωi solved by generation are transformed into individuals in the initial randomized population by binary coding, and the optimal solution of weights ωi is obtained after several iterations of selection, crossover and variation. The first k features with the highest score in each individual are chosen as the subset and fed into the classifier. The fitness of each individual is calculated through Equation (10), and the optimal individual in each iteration is retained. The complete process of the weighting scheme is shown in Figure 4.
The weighting scheme determines the ranking order of the sensitivity of the candidate features, but it cannot directly locate the optimal subset from it. The Wrapper model has an unmatched advantage over the Filter model in terms of pinpointing an optimal subset, but it also has a high temporal complexity, particularly in the case of blind search. However, if there is a clear search direction, the time complexity of using the Wrapper model to find the optimal subset can be greatly reduced. Because the fusion ranking results are already ranked according to the strength of the feature representation, the Wrapper model’s necessary search direction is also identified. Therefore, finding the optimal subset of the range from the weighted re-ranking results of features just requires a few quick and effective search algorithms. In this paper, a mixture of 3 search strategies will be employed to enhance search speed and guarantee search quality. Briefly, The BS strategy is used to quickly eliminate a large number of irrelevant features and determine the approximate range of the optimal subset. The results of the BS strategy are then used as the initial subset, followed by the SFS and SBS strategies are used to search 1 by 1, and finally, the results of the 2 strategies are compared, and the 1 with the highest classification accuracy is retained, to achieve the precise location of the optimal subset. Naturally, the BS, SFS, and SBS strategies all take into account how features work together during the classification process.
The features that are irrelevant and redundant to the classification objective will finally be removed from the original feature set through the 3-stage progression. In order to maintain or even increase the accuracy of fault classification while reducing the dimensionality of the feature set, the obtained feature subset will be the optimal 1 that considers both the characterization ability of the individual features and the characterization ability of the combined features. Figure 5 depicts the whole flowchart of the proposed three-stage hybrid feature selection framework, and the effectiveness of this framework will be tested in the following section.

5. Experiments and Analyses

5.1. Experimental Apparatus and Data Acquisition

The vibration signal samples used in this paper were obtained using a wind turbine drive system fault simulation test bench (WTDS) manufactured by Spectra Quest. According to Figure 6, the test bench is primarily made up of a variable speed drive, a planetary gearbox, a two-stage parallel shaft gearbox, and a programmed magnetic brake. Among them, the magnetic brake simulates the operating load by applying torque by modulating the input current. The motor will maintain a constant speed of 1000 r/min throughout the acquisition process for the test sample.
As illustrated in Figure 7, the faulty parts were all mounted on the parallel shaft gearbox’s low-speed shaft, where they were detected by the vibrating sensor on the bearing end cover. The sampling frequency is adjusted to 20.48 kHz during acquisition, and the single fault or compound fault is introduced by swapping out several faulty parts. The collected fault categories include normal state, five single faults and three compound faults formed by coupling the single fault of gear and single fault of bearing. Table 1 displays the nine fault categories that were acquired overall. Each class intercepts 200 samples with 5000 points of segment length, and for each class, samples are completely randomly divided into training and testing sets in the ratio of 1:1. In total, there are 1800 samples for nine classes, 900 of which are the training data and the remaining 900 are the testing data.

5.2. Experimental Results and Analysis

By preprocessing the raw vibration signals and extracting the statistical features, each sample has 288 features, and the gearbox case, which contains nine state types, obtains an original high-dimensional set of 1800 × 288-dimensional. If all the original features are used as inputs to the classifier, the computational costs will be significant, and the classifier might not produce optimal results. Therefore, it is necessary to use the feature selection method to screen out the most beneficial feature subset for the fault diagnosis task. The characterization ability of these features is currently unclear, so first, a rapid pre-ranking of these 288 features is performed by three Filter models as described above, and the pre-ranking results are shown in columns 2–4 in Table 2. For instance, the feature numbered 1 is ranked ninth in the FS model, 33rd in the IG model, and 71st in the PCC model, while the feature numbered 288 is ranked 210th in the FS model, 203rd in the IG model, and 134th in the PCC model. This shows that different models produce different results for the same feature, so the evaluation outcomes of fault features from various evaluation viewpoints are all distinct. Therefore, multiple measurements should be combined to get evaluation findings that are more accurate and comprehensive.
These pre-ranking results of three Filter models are combined using a weighting scheme. The training set is utilized to measure the classification accuracy after a subset of features comprised of the top 20 ranked features from each method is fed into the ML-kNN classifier. The average accuracy was calculated independently after this procedure had been carried out ten times, and the resulting values were FS: 0.8967, IG: 0.9167, and PCC: 0.7956. The genetic algorithm was then used to find the weights ωi in Equation (11), with the population size set to 20, the maximum number of generations to 50, and the crossover rate and mutation rate of the genetic algorithm set to 0.8 and 0.1, respectively. Equation (10) was used to calculate fitness from the first 20 features of each individual. In this equation, β was set to 0.7, and the weights ωi of the three sets of Score were ultimately determined to be 0.7261, 0.2325 and 0.0414, respectively. Finally, the FusionScore is synthesized by Equation (11):
For the feature numbered 1, its FusionScore is calculated as follows:
0.7261 × 0.8967 × 9 + 0.2325 × 0.9167 × 33 + 0.0414 × 0.7956 × 71 = 15.2318
For the feature numbered 288, its FusionScore is calculated as follows:
0.7261 × 0.8967 × 210 + 0.2325 × 0.9167 × 203 + 0.0414 × 0.7956 × 134 = 184.4093
According to their FusionScore, each feature in the original feature set is ranked in descending order, as seen in the final column of Table 2. At this time, the feature numbered 1 is ranked 10th in the weighting scheme ranking, whereas the feature numbered 288 is placed 203rd.
In this case, the higher the ranking of the features in the fusion ranking result, the better the performance in the fault classification task. The BS strategy was first used to expedite the search process by dividing the original feature set into a left subset and a right subset and measuring the recognition accuracy of the two subsets separately in order to precisely determine the features contained in the optimal subset. For the first iteration, the left subset’s subset accuracy is 94.11%, while the right subset’s subset accuracy is substantially lower at just 60.89%; hence, the left subset is chosen for future investigation.
This procedure was repeated until the subset accuracy of the left subset and right subset fell below the threshold (94%), as seen in Figure 8. After three iterations, the original feature set’s 288 features had been reduced to 36 features. Then, using the SBS and SFS strategies, respectively, these 36 features were used as the initial subset. While the SFS strategy stopped after 36 features had been searched and had an accuracy of 94.44%, the SBS method stopped after 31 features had been searched and had an accuracy of 95.67%. In contrast, the SBS strategy produced search results with higher accuracy; hence, its search results were regarded as the optimal subset, and the feature selection process was finished; the original 1800 × 288-dimensional feature set was reduced to 1800 × 31-dimensional.
The optimal subset consisting of these 31 features in the training set is input into the ML-kNN classifier, and the testing set validates the performance of feature selection; the predicted labels are shown in Figure 9. The horizontal axis indicates the number of labels predicted to correspond to a certain fault type in the current sample, and it can be found that both single and compound faults are effectively identified. The compound faults are successfully decoupled into two single faults, such as the labels corresponding to two single faults, S4 and S6, which are predicted according to the compound fault sample S9. The labels corresponding to S2 and S6 are predicted to the compound fault sample S8.
The preceding analysis revealed that partial correctness should also be taken into account in multi-label classification. Therefore, in addition to subset accuracy (SA) and average precision (AP), the following four indicators were chosen as assessment indicators to properly assess the performance of the proposed method: Hamming loss (HL), ranking loss (RL), one-error (OE), and coverage (Cov) [18], and these six indicators are calculated as follows:
S A ( h ) = 1 p i = 1 p h ( x i ) = Y i
A P ( h ) = 1 p i = 1 p 1 | Y i | y Y i | { y | r a n k h ( x i , y ) r a n k h ( x i , y ) , y Y i } | r a n k h ( x i , y )
H L ( h ) = 1 p i = 1 p 1 q | h ( x i ) Y i |
O E ( h ) = 1 p i = 1 p [ arg max y L h ( x i , y ) ] Y i
R L ( h ) = 1 p i = 1 p 1 | Y i | | Y ¯ i | y Y i { ( y , y ) | h ( x i , y ) h ( x i , y ) }
C o v ( h ) = 1 p i = 1 p max y Y i r a n k h ( x i , y ) 1
where ⟦ ⟧ represents that the condition returns 1 if true and 0 otherwise, h( ) represents the trained classifier, Yi represents the label vector corresponding to the test sample xi, △ represents the symmetric difference between two sets, and |A| represents the base of set A.
It is straightforward to see from the expressions that the larger the value of subset accuracy and average precision, the better the performance, while the smaller the value of Hamming loss, ranking loss, one-error, and coverage, the better the performance. We will measure the performance of the proposed method by comparing these six indicators in the testing set. First, comparing the diagnostic performance of the feature set before and after feature selection, the results are listed in Table 3. From another standpoint, the computational cost can also be used to measure the performance of the classification, so this paper also compares the average computational cost of the classifier for training and testing procedures (in seconds).
Table 3 shows that the optimal subset outperforms the original set in each of the six indicators, and the computational cost is also a bit lower. Since the definition of the subset accuracy is extremely strict, the original feature set performs much worse than the optimal subset in terms of subset accuracy, lagging by 10.00%, although its average accuracy only lags by 0.76%. This shows that the proposed feature selection and fault decoupling framework may be utilized to diagnose gearbox compound faults in an efficient manner, improving classification performance while requiring fewer features.

5.3. Comparison with Other Methods

To demonstrate the superiority of the proposed method, we used the weighting scheme proposed by ZHANG et al. [45] as a comparison to synthesize the ranking results of three Filter models by Equation (9) only, compared the classification performance of features in the two sets of weighted ranking results, and the variation curve of the subset Tab;ccuracy with the number of features was plotted, as shown in Figure 10. In Figure 10, the horizontal axis indicates the percentage of the top-ranked features in the fusion ranking results. Generally speaking, the classification performance caused by selecting the top k features will be higher than selecting the top k − 1 features and lower than selecting the top k + 1 features. However, the classification performance is not simply superimposed due to the combined effect among the features, so the curve exhibits an oscillation phenomenon. But the overall performance still gradually improves with the addition of more features, so there is no significant shift in the general trend.
Analyzing Figure 10, it is intuitively clear that as the number of features rises, both methods’ performance in terms of subset accuracy steadily improves and finally stabilizes. However, the proposed weighting scheme can obtain higher subset accuracy, and the oscillation phenomenon is weaker than the comparison method, as shown by the fact that the proposed method is more reasonable and can assign appropriate weights according to their contribution degree of various Filter models for the better synthesis of the re-ranking results.
The proposed method is compared with three representative classical methods: the PPT + MI [46], LP + CHI [47], and PPT + Relief-F [37], described as Method 1, Method 2, and Method 3, respectively. In Method 1, multi-label data was initially transformed to single-label data using the pruned problem transformation (PPT), after which the optimal features in the converted feature set were chosen based on mutual information. And in Method 3, which is similar to Method 1, the best features are selected by using Relief in the converted feature set. In Method 2, the multi-label data is transformed to single-label data by LP as the proposed method, but the optimal features are selected only by CHI in the converted feature set. The variation curves of subset accuracy with the number of features are plotted as shown in Figure 11 for a more understandable comparison since these three algorithms cannot accurately obtain the optimal subset automatically and only rely on methods like setting thresholds to end the selection process. As can be observed by the fact that Method 2 requires the fewest features to achieve its highest subset accuracy, but the difference between Method 2 and Method 3 is relatively slight, and the curves of both are almost always in the intersection. However, the highest subset accuracy achieved by both of them is much lower than that of the proposed method, and they are quickly caught up and surpassed by the proposed method, which consistently outperforms the other three methods after using 10% of the features. Method 1, on the other hand, has the best performance for a short time in the beginning, and then its performance is consistently lower than the other methods; nevertheless, when the number of features rises, Method 1’s performance eventually catches up to Method 2 and Method 3.
Table 4 lists the four methods’ top performances, with the best outcomes in each column denoted by bold type. Since Method 1 and Method 3 evaluate the features from the distance measure by CHI and Relief-F, respectively, the selected features are the most beneficial to the classification task, and the subset accuracy is improved the quickest. However, due to the existence of redundant features, the maximum accuracy they can achieve is lower than that of the proposed method. While Method 2 evaluates the features from the information measured by MI, so its improvement speed of subset accuracy is relatively slowest, and it is also unable to avoid redundant features. The proposed method takes into account all three measures and is able to eliminate the redundant and irrelevant features in the subset to the maximum extent. The other three methods, which required the usage of more features, only achieved subset accuracy of 93.11%, 95.33%, and 96.00%, whereas the proposed method produced a subset accuracy of 96.22%. In all six indicators, Method 1 and Method 2 fared worse than the other two methods and failed to win any of them. Method 3, on the other hand, achieved a Hamming loss that was extremely consistent with the proposed method and performed quite similarly in terms of diagnostics. Several other indicators also won and lost against each other, with the differences being minor. However, Method 3 chooses roughly twice as many optimal features as the proposed method, which means the classifier must be trained for a longer period. This indicates that the proposed method can obtain higher diagnostic performance with fewer features by removing as many irrelevant and redundant features from the original feature set as is practicable.

5.4. Added Experiments

To validate the generality of the proposed method, the proposed method was validated using the PHM09 dataset, which was obtained from a challenge organized by the Prognostics and Health Management Association in 2009 [48], using a general-purpose industrial gearbox measured as shown in Figure 12, including three shafts, two pairs of helical gears, and six bearings. The dataset was collected for failures at 30, 35, 40, 45, and 50 Hz speeds under high and low load scenarios, respectively, which did not include repeated trials.
Six fault types collected at 2400 r/min and low load were used, with a sampling frequency of 66.67 kHz. Table 5 displays the individual fault kinds, and the faults are highlighted in bold throughout the table. For each fault state, 80 samples were intercepted with 5000 points of segment length for each one and split into training and testing sets in a 1:1 ratio. We refer to the dataset obtained through WTDS as Case 1 and the dataset obtained using PHM09 as Case 2.
For Case 2, a 480 × 288-dimensional original high-dimensional feature set was generated, similar to how Case 1’s features were extracted. The ability of each feature in the feature set generated for a given dataset to describe the operating state of the gearbox is very different for different datasets, so the 420 features are still evaluated by three filter models first and pre-ranked in descending order according to the evaluation results. The pre-ranking results of the features in each method are displayed in columns two to four of Table 6. The feature subset consisting of the top 10 features in each method was selected, and input into the ML-kNN classifier, and the classification accuracy was estimated by the training set; the results obtained were FS: 0.9583, IG: 1.0000, and PCC: 0.7375. The genetic algorithm was then used to find the weights ωi in Equation (11), with the population size set to 10, the maximum number of generations 50, and the crossover rate and mutation rate of the genetic algorithm set to 0.8 and 0.1, respectively. Equation (10) was used to calculate fitness from the first 10 features of each individual. In this equation, β was set to 0.7, and the weights ωi of the three sets of Score were ultimately determined to be 0.6740, 0.3064, and 0.0196, respectively. Equation (11) was then used to synthesize the FusionScore. According to their FusionScore, each feature in the feature set is ordered in the last column of Table 6 in ascending order.
The search process was then sped up by the BS strategy at the iterative search stage in order to discover the precise range of the optimal subset, as shown in Figure 13. The number of features in the original feature set was reduced from 288 to 18 after 4 iterations. The SBS and SFS strategies were then used to iterate the search once more using these 18 features as the initial subset. The SBS strategy stops at 15 features, at which time the subset accuracy is 100.00%, while the SFS strategy stops at 18 features, at which time the subset accuracy is also 100.00%. In contrast, the SBS strategy’s search results include fewer features; therefore, this search result is considered to be the optimal subset, and the feature selection process is completed; the original 480 × 288-dimensional feature set has been reduced to 480 × 15-dimensional.
The training set’s optimal subset is fed into ML-kNN for training, and the trained ML-kNN is used to diagnose the testing set with the predicted labels as shown in Figure 14, the horizontal axis indicating the number of labels corresponding to the current sample that is predicted to be a certain fault type. The samples in the PHM09 dataset are coupled to more fault states than in Case 1—four parts of S3 generate faults simultaneously, and three parts of S4 generate faults simultaneously—but the proposed method can still fully decouple them into multiple single faults which are coupled to form that compound fault state.
The original high-dimensional feature set achieved a subset accuracy of 98.75%, attributable to the PHM09 dataset’s exceptionally ideal acquisition conditions, while the optimal subset achieved a subset accuracy of 100%. Figure 15 depicts the variation curves of the subset accuracy of the four methods with the number of features to make visual observation easier. The proposed method still performs the best, despite the fact that the margin between the four methods is significantly narrower than it was in Case 1. Only the proposed method reaches 100% subset accuracy when only the first 6% of features are employed, despite the fact that the subset accuracy of all four methods can rise as the number of features increases. Method 1 and Method 2 perform similarly and slightly worse than the proposed method, while Method 3 performs the worst, with a subset accuracy of about 96%.
In Case 2, the proposed method automatically determines an optimal subset with 15 features. The diagnostic performance of three different methods is compared when employing the top 15 features from their ranking results in order to more clearly illustrate the diagnostic benefits of the proposed feature selection and fault decoupling method. Table 7 summarizes the subset accuracy, average precision and Hamming loss of each method. It is clear that in Case 2, the proposed method still has the best performance, as it outperforms the other models in all evaluation metrics. Method 3 performs worse than Method 1, which is the opposite of Case 1, indicating that their performance is less stable than the proposed method in different source datasets.
Both Table 4 and Table 7 demonstrate that the proposed method has stable performance and is effective in feature selection and fault decoupling in both Case 1 and Case 2. The method can guarantee better accuracy while using the fewest possible feature dimensions. Therefore, we can conclude that the method performs better in terms of test accuracy than the three methods that were compared.

6. Conclusions and Future Works

This paper proposes an effective feature selection and fault decoupling framework in the field of intelligent fault diagnosis, which can achieve the highest diagnostic accuracy. The proposed method’s experimental findings on two gearbox datasets demonstrate that it can diagnose more accurately while using fewer features than previous methods. It also has the advantages of simple computation and high accuracy, which increases the usefulness and efficacy of intelligent diagnosis models. From this research, the following findings can be made:
  • In order to describe the operating condition of the gearbox as comprehensively as possible, we propose to construct multi-domain feature sets. The Teager energy operator and the EMD algorithm are utilized in the proposed feature extraction stage to filter out disturbing components from the signal and highlight the fault feature information. Each vibration signal is thus given a high-dimensional feature vector that contains irrelevant and redundant features, and it is important to adaptively screen the most sensitive features using feature selection techniques;
  • The original multi-label problem will be taken again into account to achieve multi-label classification by the AA method once the feature selection for the original feature set has been completed by employing the PT method. Consequently, the feature selection and fault decoupling framework proposed in this paper is more adaptable;
  • The original high-dimensional feature set contained redundant and irrelevant features, which may have increased training time and decreased recognition accuracy. The ML-kNN is utilized as the classifier in the proposed feature selection method, and three Filter models are initially employed to pre-rank the features. To achieve a reranking result that takes into account feature irrelevance, redundancy, and inter-feature interaction factors, the pre-ranking results are fused using a weighting scheme. In order to optimize the subset for the classifier and attain the highest diagnostic accuracy, a search stage that could adaptively discover the optimal subset is constructed based on three heuristic search strategies;
  • The proposed method does not entail complicated mapping, which is an intuitive and straightforward process. Therefore, this method has a good physical interpretation and helps to reveal the connection between faults and their related features, offering a new solution for diagnosing compound defects in gearboxes.
The most significant advantage of the proposed method over other methods is that better fault diagnosis performance can be achieved, but at the same time, the computational overhead in the feature selection stage is relatively high. Therefore it should also be considered in subsequent studies how to reduce computational costs while maintaining diagnostic performance. In addition, in this work, we transformed the multi-label data by label powerset, which may perform less well in datasets with imbalanced samples or a lot of label space. Therefore, more suitable processing methods need to be tried, which is the direction that the authors would make further efforts.

Author Contributions

Conceptualization, D.L. and X.Z.; Data curation, X.Z.; Formal analysis, X.Z.; Funding acquisition, H.J.; Investigation, H.J.; Methodology, D.L.; Project administration, X.Z.; Validation, X.Z.; Writing—original draft, D.L.; Writing—review & editing, D.L. and Z.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (52265016; 51865054); the Natural Science Foundation of Xinjiang Uygur Autonomous Region (2022D01C36).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author. The data are not publicly available due to laboratory regulations.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

The equation of statistical features adopted in this work is shown in Table A1.
Table A1. Statistical feature of the time and frequency domains.
Table A1. Statistical feature of the time and frequency domains.
Time-Domain FeatureFrequency-Domain Feature
T 1 = n = 1 N x ( n ) N T 7 = n = 1 N ( x ( n ) T 1 ) 4 ( N 1 ) T 2 4 F 1 = k = 1 K s ( k ) K F 8 = k = 1 K f k 4 s ( k ) k = 1 K f k 2 s ( k )
T 2 = n = 1 N ( x ( n ) T 1 ) 2 N 1 T 8 = T 5 T 4 F 2 = k = 1 K ( s ( k ) F 1 ) 2 K 1 F 9 = k = 1 K f k 2 s ( k ) k = 1 K s ( k ) k = 1 K f k 2 s ( k )
T 3 = ( n = 1 N | x ( n ) | N ) 2 T 9 = T 5 T 3 F 3 = k = 1 K ( s ( k ) F 1 ) 3 K ( F 2 ) 3 F 10 = F 6 F 5
T 4 = n = 1 N ( x ( n ) ) 2 N T 10 = T 4 1 N n = 1 N | x ( n ) | F 4 = k = 1 K ( s ( k ) F 1 ) 4 K F 2 2 F 11 = k = 1 K ( f k F 5 ) 3 s ( k ) K F 6 3
T 5 = max | x ( n ) | T 11 = T 5 1 N n = 1 N | x ( n ) | F 5 = k = 1 K f k s ( k ) k = 1 K s ( k ) F 12 = k = 1 K ( f k F 5 ) 4 s ( k ) K F 6 4
T 6 = n = 1 N ( x ( n ) T 1 ) 3 ( N 1 ) T 2 3 F 6 = k = 1 K ( f k F 5 ) 2 s ( k ) K F 13 = k = 1 K ( f k F 5 ) s ( k ) K F 6
F 7 = k = 1 K f k 2 s ( k ) k = 1 K s ( k )
From the table, x(n) is a signal series for n= 1, 2, …, N, N is the number of data points, s(k) is a spectrum for k= 1, 2, …, K, K is the number of spectrum lines, and fk is the frequency value of the kth spectrum line, fˆ(λl) is an eigenvector for λl, λl is the eigenvalue of the lth degree, l = 1, 2, …, N.

References

  1. Yu, F.; Liu, Y.; Zhao, Q. Compound Fault Diagnosis of Gearbox Based on Wavelet Packet Transform and Sparse Representation Classification. In Proceedings of the 2020 Chinese Control and Decision Conference (CCDC), Hefei, China, 22–24 August 2020; pp. 1105–1109. [Google Scholar]
  2. Dibaj, A.; Ettefagh, M.M.; Hassannejad, R.; Ehghaghi, M.B. A hybrid fine-tuned VMD and CNN scheme for untrained compound fault diagnosis of rotating machinery with unequal-severity faults. Expert Syst. Appl. 2021, 167, 114094. [Google Scholar] [CrossRef]
  3. Huang, R.; Liao, Y.; Zhang, S.; Li, W. Deep Decoupling Convolutional Neural Network for Intelligent Compound Fault Diagnosis. IEEE Access 2019, 7, 1848–1858. [Google Scholar] [CrossRef]
  4. Okwuosa, C.N.; Hur, J.-W. A Filter-Based Feature-Engineering-Assisted SVC Fault Classification for SCIM at Minor-Load Conditions. Energies 2022, 15, 7597. [Google Scholar] [CrossRef]
  5. Zhang, X.; Zhang, Q.; Chen, M.; Sun, Y.; Qin, X.; Li, H. A two-stage feature selection and intelligent fault diagnosis method for rotating machinery using hybrid filter and wrapper method. Neurocomputing 2018, 275, 2426–2439. [Google Scholar] [CrossRef]
  6. Pan, L.; Zhao, L.; Song, A.; She, S.; Wang, S. Research on gear fault diagnosis based on feature fusion optimization and improved two hidden layer extreme learning machine. Measurement 2021, 177, 109317. [Google Scholar] [CrossRef]
  7. Wen, L.; Li, X.; Gao, L. A New Two-Level Hierarchical Diagnosis Network Based on Convolutional Neural Network. IEEE Trans. Instrum. Meas. 2020, 69, 330–338. [Google Scholar] [CrossRef]
  8. Akpudo, U.E.; Hur, J.-W. A Cost-Efficient MFCC-Based Fault Detection and Isolation Technology for Electromagnetic Pumps. Electronics 2021, 10, 439. [Google Scholar] [CrossRef]
  9. Lee, C.; Le, T.; Lin, Y. A Feature Selection Approach Hybrid Grey Wolf and Heap-Based Optimizer Applied in Bearing Fault Diagnosis. IEEE Access 2022, 10, 56691–56705. [Google Scholar] [CrossRef]
  10. Liu, H.; Zhou, M.; Liu, Q. An Embedded Feature Selection Method for Imbalanced Data Classification. IEEE/CAA J. Autom. Sin. 2019, 6, 703–715. [Google Scholar] [CrossRef]
  11. Chen, K.; Xue, B.; Zhang, M.; Zhou, F. An Evolutionary Multitasking-Based Feature Selection Method for High-Dimensional Classification. IEEE Trans. Cybern. 2022, 52, 7172–7186. [Google Scholar] [CrossRef]
  12. Lei, Y.; Jia, F.; Lin, J.; Xing, S.; Ding, S.X. An Intelligent Fault Diagnosis Method Using Unsupervised Feature Learning Towards Mechanical Big Data. IEEE Trans. Ind. Electron. 2016, 63, 3137–3147. [Google Scholar] [CrossRef]
  13. Dhamande, L.S.; Chaudhari, M.B. Compound gear-bearing fault feature extraction using statistical features based on time-frequency method. Measurement 2018, 125, 63–77. [Google Scholar] [CrossRef]
  14. Yan, X.; Jia, M. A novel optimized SVM classification algorithm with multi-domain feature and its application to fault diagnosis of rolling bearing. Neurocomputing 2018, 313, 47–64. [Google Scholar] [CrossRef]
  15. Zhao, X.; Jia, M. Fault diagnosis of rolling bearing based on feature reduction with global-local margin Fisher analysis. Neurocomputing 2018, 315, 447–464. [Google Scholar] [CrossRef]
  16. Wen, X.; Xu, Z. Wind turbine fault diagnosis based on ReliefF-PCA and DNN. Expert Syst. Appl. 2021, 178, 115016. [Google Scholar] [CrossRef]
  17. Zhou, Y.; Yan, S.; Ren, Y.; Liu, S. Rolling bearing fault diagnosis using transient-extracting transform and linear discriminant analysis. Measurement 2021, 178, 109298. [Google Scholar] [CrossRef]
  18. Qian, W.; Long, X.; Wang, Y.; Xie, Y. Multi-label feature selection based on label distribution and feature complementarity. Appl. Soft Comput. 2020, 90, 106167. [Google Scholar] [CrossRef]
  19. Malhi, A.; Gao, R.X. PCA-based feature selection scheme for machine defect classification. IEEE Trans. Instrum. Meas. 2004, 53, 1517–1525. [Google Scholar] [CrossRef]
  20. Cadenas, J.M.; Carmen Garrido, M.; Martinez, R. Feature subset selection Filter-Wrapper based on low quality data. Expert Syst. Appl. 2013, 40, 6241–6252. [Google Scholar] [CrossRef]
  21. Dokeroglu, T.; Deniz, A.; Kiziloz, H.E. A comprehensive survey on recent metaheuristics for feature selection. Neurocomputing 2022, 494, 269–296. [Google Scholar] [CrossRef]
  22. Li, X.; Ren, J. MICQ-IPSO: An effective two-stage hybrid feature selection algorithm for high-dimensional data. Neurocomputing 2022, 501, 328–342. [Google Scholar] [CrossRef]
  23. Patel, S.P.; Upadhyay, S.H. Euclidean distance based feature ranking and subset selection for bearing fault diagnosis. Expert Syst. Appl. 2020, 154, 113400. [Google Scholar] [CrossRef]
  24. Shi, J.; Yi, J.; Ren, Y.; Li, Y.; Zhong, Q.; Tang, H.; Chen, L. Fault diagnosis in a hydraulic directional valve using a two-stage multi-sensor information fusion. Measurement 2021, 179, 109460. [Google Scholar] [CrossRef]
  25. Ding, Y.; Fan, L.; Liu, X. Analysis of feature matrix in machine learning algorithms to predict energy consumption of public buildings. Energy Build. 2021, 249, 111208. [Google Scholar] [CrossRef]
  26. Wang, Y.; Feng, L. Hybrid feature selection using component co-occurrence based feature relevance measurement. Expert Syst. Appl. 2018, 102, 83–99. [Google Scholar] [CrossRef]
  27. Liang, K.; Dai, W.; Du, R. A Feature Selection Method Based on Improved Genetic Algorithm. In Proceedings of the 2020 Global Reliability and Prognostics and Health Management (PHM-Shanghai), Shanghai, China, 16–18 October 2020; pp. 1–5. [Google Scholar]
  28. Mochammad, S.; Kang, Y.; Noh, Y.; Park, S.; Ahn, B. Stable Hybrid Feature Selection Method for Compressor Fault Diagnosis. IEEE Access 2021, 9, 97415–97429. [Google Scholar] [CrossRef]
  29. Ganjei, M.A.; Boostani, R. A hybrid feature selection scheme for high-dimensional data. Eng. Appl. Artif. Intell. 2022, 113, 104894. [Google Scholar] [CrossRef]
  30. Ma, Y.; Cheng, J.; Wang, P.; Wang, J.; Yang, Y. Rotating machinery fault diagnosis based on multivariate multiscale fuzzy distribution entropy and Fisher score. Measurement 2021, 179, 109495. [Google Scholar] [CrossRef]
  31. Shannon, C.E. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 623–656. [Google Scholar] [CrossRef]
  32. Omuya, E.O.; Okeyo, G.O.; Kimwele, M.W. Feature Selection for Classification using Principal Component Analysis and Information Gain. Expert Syst. Appl. 2021, 174, 114765. [Google Scholar] [CrossRef]
  33. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar]
  34. Mladenic, D. Feature selection for dimensionality reduction. In Subspace, Latent Structure and Feature Selection; Saunders, G., Grobelnik, M., Gunn, S., ShaweTaylor, J., Eds.; Lecture Notes in Computer Science; Springer Science+Business Media: Berlin/Heidelberg, Germany, 2006; Volume 3940, pp. 84–102. [Google Scholar]
  35. Liu, H.; Yu, L. Toward integrating feature selection algorithms for classification and clustering. IEEE Trans. Knowl. Data Eng. 2005, 17, 491–502. [Google Scholar]
  36. Moyano, J.M.; Gibaja, E.L.; Cios, K.J.; Ventura, S. Review of ensembles of multi-label classifiers: Models, experimental study and prospects. Inf. Fusion 2018, 44, 33–45. [Google Scholar] [CrossRef]
  37. Reyes, O.; Morell, C.; Ventura, S. Scalable extensions of the ReliefF algorithm for weighting and selecting features on the multi-label learning context. Neurocomputing 2015, 161, 168–182. [Google Scholar] [CrossRef]
  38. Newton, S.; Everton, A.C.; Maria, C.M.; Huei, D.L. A Comparison of Multi-label Feature Selection Methods using the Problem Transformation Approach. Electron. Notes Theor. Comput. Sci. 2013, 292, 135–151. [Google Scholar]
  39. Tsoumakas, G.; Katakis, I.; Vlahavas, I. Random k-Labelsets for Multilabel Classification. IEEE Trans. Knowl. Data Eng. 2011, 23, 1079–1089. [Google Scholar] [CrossRef]
  40. Maltoudoglou, L.; Paisios, A.; Lenc, L.; Martinek, J.; Kral, P.; Papadopoulos, H. Well-calibrated confidence measures for multi-label text classification with a large number of labels. Pattern Recognit. 2022, 122, 108271. [Google Scholar] [CrossRef]
  41. Zhang, M.; Zhou, Z. ML-KNN: A lazy learning approach to multi-label leaming. Pattern Recognit. 2007, 40, 2038–2048. [Google Scholar] [CrossRef]
  42. Van Tung, T.; AlThobiani, F.; Ball, A. An approach to fault diagnosis of reciprocating compressor valves using Teager-Kaiser energy operator and deep belief networks. Expert Syst. Appl. 2014, 41, 4113–4122. [Google Scholar]
  43. Robert, B.; Randall, J.A. Why EMD and similar decompositions are of little benefit for bearing diagnostics. Mech. Syst. Signal Process. 2023, 192, 110207. [Google Scholar]
  44. Lei, Y.; He, Z.; Zi, Y.; Hu, Q. Fault diagnosis of rotating machinery based on multiple ANFIS combination with GAS. Mech. Syst. Signal Process. 2007, 21, 2280–2294. [Google Scholar] [CrossRef]
  45. Zhang, K.; Li, Y.; Scarf, P.; Ball, A. Feature selection for high-dimensional machinery fault diagnosis data using multiple models and Radial Basis Function networks. Neurocomputing 2011, 74, 2941–2952. [Google Scholar] [CrossRef]
  46. Doquire, G.V.M. Feature Selection for Multi-label Classification Problems. In Proceedings of the International Work-Conference on Artificial Neural Networks, Torremolinos, Spain, 8–10 June 2011; pp. 9–16. [Google Scholar]
  47. Trochidis, K.; Tsoumakas, G.; Kalliris, G.; Vlahavas, I. Multi-label classification of music into emotions. In Proceedings of the 9th International Conference on Music Information Retrieval (ISMIR 2008), Philadelphia, PA, USA, 14–18 September 2008; pp. 325–330. [Google Scholar]
  48. PHM, Phm Data Challenge 2009 [DB]. 2009. Available online: https://www.phmsociety.org/competition/PHM/09 (accessed on 10 April 2009).
Figure 1. The flowchart of the proposed feature and fault decoupling method.
Figure 1. The flowchart of the proposed feature and fault decoupling method.
Sensors 23 04792 g001
Figure 2. The feature selection process for multi-label data.
Figure 2. The feature selection process for multi-label data.
Sensors 23 04792 g002
Figure 3. The flowchart of original feature extraction.
Figure 3. The flowchart of original feature extraction.
Sensors 23 04792 g003
Figure 4. The flowchart of the proposed weighting scheme.
Figure 4. The flowchart of the proposed weighting scheme.
Sensors 23 04792 g004
Figure 5. The flowchart of the proposed three-stage hybrid feature selection framework.
Figure 5. The flowchart of the proposed three-stage hybrid feature selection framework.
Sensors 23 04792 g005
Figure 6. Wind turbine drive system fault simulation test bench.
Figure 6. Wind turbine drive system fault simulation test bench.
Sensors 23 04792 g006
Figure 7. The installation position of the vibrating sensor.
Figure 7. The installation position of the vibrating sensor.
Sensors 23 04792 g007
Figure 8. The visual presentation of the iterative search process for Case 1.
Figure 8. The visual presentation of the iterative search process for Case 1.
Sensors 23 04792 g008
Figure 9. Gearbox compound fault dataset classification confusion matrix for Case 1.
Figure 9. Gearbox compound fault dataset classification confusion matrix for Case 1.
Sensors 23 04792 g009
Figure 10. The comparison result between the proposed method and the previous method.
Figure 10. The comparison result between the proposed method and the previous method.
Sensors 23 04792 g010
Figure 11. The comparison result between the proposed method and other methods for Case 1 in terms of subset accuracy.
Figure 11. The comparison result between the proposed method and other methods for Case 1 in terms of subset accuracy.
Sensors 23 04792 g011
Figure 12. A pictorial illustration of the experimental gearbox. (a) Gearbox interior. (b) Schematic diagram of gearbox structure.
Figure 12. A pictorial illustration of the experimental gearbox. (a) Gearbox interior. (b) Schematic diagram of gearbox structure.
Sensors 23 04792 g012
Figure 13. The visual presentation of the iterative search process for Case 2.
Figure 13. The visual presentation of the iterative search process for Case 2.
Sensors 23 04792 g013
Figure 14. Gearbox compound fault dataset classification confusion matrix for Case 2.
Figure 14. Gearbox compound fault dataset classification confusion matrix for Case 2.
Sensors 23 04792 g014
Figure 15. The comparison result between the proposed method and other methods for Case 2 in terms of subset accuracy.
Figure 15. The comparison result between the proposed method and other methods for Case 2 in terms of subset accuracy.
Sensors 23 04792 g015
Table 1. Testing data description.
Table 1. Testing data description.
Fault TypesS1S2S3S4S5S6S7S8S9
Normal
Gear surface wear
Gear root crack
Gear missing tooth
Bearing inner race fault
Bearing outer race fault
From the table, √ indicates the fault type in the sample corresponding to each column.
Table 2. Ranking results for features of different methods for Case 1.
Table 2. Ranking results for features of different methods for Case 1.
OrderFSIGPCCWeighting Scheme
1223146240223
2225159229225
3236172251236
4234185262234
5247236273247
6106223172245
7245234146235
8107198185226
91211159107
102262252281
2853414126142
28699143139194
287194121239132
28888128106187
Table 3. Performance comparison with the origin feature set and the optimal subset.
Table 3. Performance comparison with the origin feature set and the optimal subset.
Computational Cost (s)SAAPHLRLCovOE
Origin feature set0.9710.86220.97650.02780.01140.38780.0444
Optimal subset0.4690.96220.98410.01130.00840.36890.0300
Table 4. Performance comparison with other frameworks for Case 1.
Table 4. Performance comparison with other frameworks for Case 1.
MethodNumber of Optimal FeaturesComputational Cost (s)SAAPHLRLCovOE
The proposed method310.4690.96220.98410.01130.00840.36890.0300
Method 1450.4890.93110.97300.01870.01240.39560.0500
Method 2440.4890.95330.97650.01500.00960.38110.0467
Method 3570.5170.96000.98430.01130.00880.36780.0278
From the table, the bold indicates the best performance in each column.
Table 5. Testing data description.
Table 5. Testing data description.
CaseGearBearingShaft
16T48T24T40T123456InputOutput
S1GoodGoodGoodGoodGoodGoodGoodGoodGoodGoodGoodGood
S2GoodGoodChippedGoodGoodGoodGoodGoodGoodGoodGoodGood
S3GoodGoodBrokenGoodGoodGoodGoodCombinationInnerGoodBent shaftGood
S4GoodGoodGoodGoodGoodGoodGoodCombinationBallGoodImbalanceGood
S5GoodGoodBrokenGoodGoodGoodGoodGoodInnerGoodGoodGood
S6GoodGoodGoodGoodGoodGoodGoodGoodGoodGoodBent shaftGood
From the table, the bold indicates the specific fault type.
Table 6. Ranking results for features of different methods for Case 2.
Table 6. Ranking results for features of different methods for Case 2.
OrderFSIGPCCWeighting Scheme
122580229225
2236223251236
322382240223
4988228234
523452549
6116964
71726211
88814513
95912531
10801023281
285336328372
286341423195
287591445759
2887414313974
Table 7. Performance comparison with other frameworks for Case 2.
Table 7. Performance comparison with other frameworks for Case 2.
MethodSAAPHL
The proposed method1.00001.00000.0000
Method 10.99171.00000.0021
Method 20.98751.00000.0031
Method 30.96670.98610.0078
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, D.; Zhang, X.; Zhang, Z.; Jiang, H. A Hybrid Feature Selection and Multi-Label Driven Intelligent Fault Diagnosis Method for Gearbox. Sensors 2023, 23, 4792. https://doi.org/10.3390/s23104792

AMA Style

Liu D, Zhang X, Zhang Z, Jiang H. A Hybrid Feature Selection and Multi-Label Driven Intelligent Fault Diagnosis Method for Gearbox. Sensors. 2023; 23(10):4792. https://doi.org/10.3390/s23104792

Chicago/Turabian Style

Liu, Di, Xiangfeng Zhang, Zhiyu Zhang, and Hong Jiang. 2023. "A Hybrid Feature Selection and Multi-Label Driven Intelligent Fault Diagnosis Method for Gearbox" Sensors 23, no. 10: 4792. https://doi.org/10.3390/s23104792

APA Style

Liu, D., Zhang, X., Zhang, Z., & Jiang, H. (2023). A Hybrid Feature Selection and Multi-Label Driven Intelligent Fault Diagnosis Method for Gearbox. Sensors, 23(10), 4792. https://doi.org/10.3390/s23104792

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop