Next Article in Journal
Rényi Entropies of Multidimensional Oscillator and Hydrogenic Systems with Applications to Highly Excited Rydberg States
Previous Article in Journal
Research on Insulator Defect Detection Based on an Improved MobilenetV1-YOLOv4
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bearing Fault Diagnosis Method Based on Convolutional Neural Network and Knowledge Graph

School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(11), 1589; https://doi.org/10.3390/e24111589
Submission received: 6 September 2022 / Revised: 27 October 2022 / Accepted: 28 October 2022 / Published: 2 November 2022

Abstract

:
An effective fault diagnosis method of bearing is the key to predictive maintenance of modern industrial equipment. With the single use of equipment failure mechanism or operation of data, it is hard to resolve multiple complex variable working conditions, multiple types of fault and equipment malfunctions and failures related to knowledge and data. In order to solve these problems, a fault diagnosis method based on the fusion of deep learning with a knowledge graph is proposed in this paper. Firstly, the knowledge rules of bearing data is used for entity extraction. Next, the multiscale optimized convolutional neural network (MOCNN) proposed in this paper is used for fault classification to achieve relationship extraction. Finally, the fault diagnosis graph of the bearing is constructed for fault-assisted decision-making as well as the detailed display of fault information. According to experiment analysis, the fault diagnosis model based on MOCNN proposed in this paper, which integrates the end-to-end convolutional neural network and the attention mechanism, still achieves an accuracy of 97.86% under the data set of 160 types of faults. Compared with the deep learning models such as Resnet and Inception in the noise environment of multiple working conditions and variable working conditions, the model proposed in this paper not only shows a faster convergence speed and stable performance, but also a higher accuracy in evaluation indicators, which is beneficial to practical use.

1. Introduction

Rolling bearings are widely used in industrial production equipment, and their health status has a huge impact on the performance, stability and service life of the entire mechanical equipment [1,2]. Therefore, the fault diagnosis and condition assessment of rolling bearings are of great significance. Normally, the fault diagnosis methods of bearings are mainly divided into two types, one based on mechanism experience [3] and the other based on data driving [4]. Affected by both external and internal factors, the failure degradation of rolling bearings often presents nonlinearity and instability. It is difficult to achieve ideal fault diagnosis results by merely relying on failure mechanism knowledge or data-operating for state identification and fault diagnosis. Therefore, the bearing fault diagnosis method based on the fusion of a knowledge graph and deep learning indicates a new research direction.
The concept of a knowledge graph was first proposed by Google in 2012 to improve the search engine [5]. Compared with traditional knowledge bases, knowledge graphs represents richer semantic relationships, higher knowledge quality, and better visualization effects, providing great convenience for human-computer interaction. The knowledge graph is composed of entities, attributes and relationships. Knowledge graphs are widely used. For example, Oramas et al. [6] used knowledge graphs to recommend sounds and music; Li et al. [7] combined the knowledge graph with SQL databases, proposed the framework design of industrial software design and development process, and verified it in specific industrial scenarios; Liu et al. [8] used the integration method of deep learning and knowledge graph to prove the effectiveness of market prediction and made stock investments. However, due to the complexity of knowledge in the field of fault diagnosis, it is extremely difficult to construct and use knowledge graphs rationally. If we can construct the knowledge graph in an accurate way, it will play a key role in maintaining the health of mechanical equipment. Therefore, this paper will attempt to construct the knowledge graph of rolling bearing fault diagnosis in order to contribute to the health maintenance of equipment.
The construction of a bearing fault graph requires fusion with data. As a subset of machine learning, deep learning has made significant progress in computer vision [9], natural language processing [10], biology [11], and other fields, and it has also been widely used in the field of fault diagnosis, such as by Liang et al. [12], who proposed a gearbox fault diagnosis model that combines wavelet transform (WT) and CNN. It was verified on two gearbox datasets and achieved good results, but as the network got deeper, the model increased more, which caused performance degradation. Thus, it is difficult to achieve better results for more complex data. The proposal of the Inception network has changed people’s perception of relying on increasing the depth and width of the network to improve its performance. It changes the structure of the network so that the parameters reduce while the depth and width of the network increases [13], and the size of the convolution kernel will directly affect the performance of the network, and a single-sized convolution kernel can lead to incomplete extraction of information [14].
In response to the problems above, this paper proposes a fault diagnosis method based on the fusion of convolutional neural networks and knowledge graphs. The main contributions are as follows:
(1)
Extracting entities based on the equipment mechanism and knowledge rules of bearing data, then applying the MOCNN to categorize faults through data labeling to complete extraction of data relationship, and eventually, the knowledge graph of bearing faults is established. Furthermore, by forming the graph to assist decision-making as well as to display detailed fault information, it realizes a complete fault diagnosis, which moves away from a single reliance on mechanism knowledge or a data-driven diagnosis technique and toward the union of data and knowledge.
(2)
This paper offers an end-to-end one-dimensional multiscale convolutional neural network model (MOCNN) that combines advantages of one-dimensional convolutional networks in processing one-dimensional input. The model first implements sensitive feature extraction of the input 1D signals from different angles using the modified Inception module, and then adds a residual link to the Inception block to learn more abundant features. Following that, the retrieved features are fine-tuned using the channel attention and spatial attention modules. To improve the performance of the model and the effectiveness of defect diagnosis, the L2 regularization is introduced to the attention module and the classification layer.
(3)
The model is combined with the knowledge graph, marked by the CWRU data set, and adopted a new experimental division method. Compared with Resnet and Inception, this model has a good diagnosis impact. It is more suitable for large data and multi-classification problems and also features fast and stable convergence on small data sets and multi-classification problems. Moreover, it performs well in noise environment tests in various noise environments, working situations, and variable working conditions.
The rest of this paper is organized as follows: Related work is described in Section 2. Section 3 introduces the fault diagnosis method, fault diagnosis model and data processing proposed in this paper. In Section 4, entities and relationships are extracted and a knowledge graph is constructed. At the same time, the fault diagnosis model is compared and discussed from different angles. A conclusion is provided in Section 5.

2. Related Work

2.1. Knowledge Graph

The construction methods of the knowledge graph are mainly divided into bottom-up and bottom-down [15], and the bottom-up construction method is to perform entity extraction, relationship extraction and attribute extraction on the data first, and then perform knowledge fusion and entity alignment. That is, different representations of unified entities are denoised to obtain a unified representation, then entity disambiguation and attribute matching are performed, and finally ontology construction and knowledge inference are performed to form a knowledge graph. The bottom-down construction approach is to first create the top-level concept ontology and then extract entities and relationships from data sources to match and update them into the initially created top-level ontology. Since domain knowledge graphs require significant knowledge expertise, this paper uses a combination of both for construction.

2.2. CNN

A convolutional neural network is a multilayer feedforward neural network that updates the parameters of the network by a backpropagation algorithm and consists of convolutional layers, pooling layers, and fully connected layers [16]. The convolutional layer is mainly for extracting features automatically and the one-dimensional convolutional neural network has a strong ability to extract sequential features [17]; the pooling layer is utilized to further highlight the extracted features while reducing the dimensionality of the data, which can both speed up and prevent overfitting, and after multiple rounds of convolution and pooling, the extracted features are inputted to the fully connected layer to realize the classification task [18].

2.3. Multiscale Convolution

Rolling bearing vibration signals often produce non-periodic vibration signals with the appearance of damage points. A single-size convolution kernel lacks adaptability when extracting features and cannot be changed according to changes in input signals. Therefore, this paper uses a multiscale convolution layer to replace the single-scale convolutional layer in the conventional network, allowing the model to learn the nonlinear relationship between input and output through richer features, thereby improving the classification accuracy of each category. Figure 1 shows the common Inception network module.

2.4. Residual Learning

In response to the gradient disappearance and explosion problems in deep neural network training [19], He et al. first proposed residual neural networks [20] (Resnet) in 2016, designed residual blocks with shortcut connections, and used multiple residual blocks connected to build residual neural networks. The structure of residual blocks is shown in Figure 1, which is defined as follows:
H x = F x + x
where x and H x represent the input and output functions respectively and F represents the residual function. The + in the figure is the summation of corresponding elements of the feature graph. The residual neural network is passed to the previous layer through a shortcut connection, thus effectively solving the problem of gradient disappearance along with the increasing depth during network training [21].

2.5. Attentional Mechanisms

CBAM is a combined spatial and channel attention mechanism network that learns along two independent dimensions of channel and space and obtains the corresponding weights [22]. Supposing that the input feature F is of size H × W × C , where H ,   W ,   C represent the length, width and number of channels of the feature, respectively. The channel attention is shown in Figure 2. The feature F is changed to 1 × 1 × C after maximum pooling and average pooling, and then inputted to the multilayer perceptron (MLP) with shared weights. After that, by using the sigmoid activation function, the MLP processed and summed features are mapped and the weight coefficients M C   are obtained, which are then multiplied with the input features to get the new features after the channel attention.
The spatial attention is shown in Figure 3. The features F are spliced on the channel after maximum pooling and average pooling, respectively, obtaining the H × W × 2 features, which are inputted to the convolution kernel of size 1 × 7 and the convolution layer with sigmoid activation function to obtain the weight coefficients M s , which are then multiplied with the input features to obtain the new features after spatial attention.

2.6. Evaluation Metrics

In this study, accuracy, macro-precision, macro-recall and macro- F 1 score are used to comprehensively evaluate the classification performance of the model [23,24]. They are defined as follows:
A c c u r a c y = T P + T N T P + F P + F N + T N × 100 %
P r e c i s i o n = T P T P + F P × 100 % ;   P r e c i s i o n m a c r o = i = 1 L P r e c i s i o n L
R e c a l l = T P T P + F N × 100 % ;   R e c a l l m a c r o = i = 1 L R e c a l l L
F 1   s c o r e m a c r o = 2 × P r e c i s i o n m a c r o × R e c a l l m a c r o   P r e c i s i o n m a c r o + R e c a l l m a c r o
where T P , F P , F N , T N represent true positive, false positive, false negative, and true negative, separately.

3. Proposed Method

This paper proposes a fault diagnosis method based on a knowledge graph and deep learning, which establishes a fault knowledge graph by extracting mechanistic knowledge and empirical rules of the equipment. The MOCNN model is capable of realizing data classification and can effectively combine mechanical knowledge with operating data of the equipment, and eventually forms a fault diagnosis model with the fusion of knowledge and data. Moreover, this model produces the knowledge graph to facilitate decision-making as well as the detailed display of fault information.

3.1. The Proposed Fault Diagnosis Model

The convolution layer is connected to the batch normalization (BN) layer after the convolution layer; the BN layer performs a batchwise mean reduction and divides the variance operation on the input data, which can speed up the training of the network [25] and increase the robustness of the model, as shown in Equations (6)–(9). After the BN operation on the dataset, the datasets are processed using the linear activation unit ReLU, and the definition of the ReLU activation function is shown in Equation (10): the batch normalization layer normalizes each batch of data to alleviate the problem of gradient disappearance in deep neural networks and improve the nonlinear expression of model pairs. The ReLU activation function can effectively improve the dispersion problem of the gradient and enhance the convergence of the network model efficiency [26].
μ = 1 N i = 1 N x i
σ 2 = 1 N i = 1 N ( x i μ i ) 2
k i = x i μ σ 2 + ε
y i = γ k i + β
where μ is the mean value of the samples, x i represents the characteristics of the inputs, N denotes the number of samples, σ is the variance, k i represents the normalized sample values, y i is the output of k i scaled, ε is a constant close to 0, and γ and β are scaling factors.
Z t   l = f x t   l = m a x 0 , x t   l
where x t   l represents the input and Z t   l represents the output.
In order to solve the problem of large computational complexity of larger convolutional kernels, this model first uses a 1 × 1 convolutional kernel to reduce the number of channels of the features and then performs the corresponding convolutional operations on them. 1 × 3 pooling can remove the redundant information extracted from the previous layer of convolution, and a larger convolutional kernel of 1 × 7 is set because the correlation of features may vary, and to learn this feature, a large convolutional kernel is required. The number of convolutional kernels in each branch of the module Basic Block A is 16. The improved Inception-Resnet module in this paper can not only overcome the gradient disappearance problem of deep network, but also solve the problem of optimal combination of convolution kernel size. At the same time, splitting the convolution kernel can increase the diversity of features and nonlinear expression. The number of convolutional kernels in the four branches of module Basic Block B are 16, 16, 16, 16, 32, and 32. The structures of modules Basic Block A and Basic Block B are shown in Figure 4.
Basic Block C combines the advantages of Resnet and CBAM to mitigate the loss and waste of information during network propagation as the network deepens and to focus attention on features that are more important for the signal classification task. This model incorporates L2 regularity for both spatial attention and channel attention mechanisms, and the structure of this module is illustrated in Figure 5.
By improving different modules, the structure of the MOCNN model constructed in this paper is shown in Figure 6.
The Figure 6 shows the structure of the MOCNN network proposed in this paper, which consists of two groups of convolutional layers, batch normalization layers and activation layers, among which the number of the first group of convolutional kernels is 32, the second group is 64 and both are of size 1 × 3. The number of convolutional kernels in the convolutional layer behind the Basic Block B module is 128, and its size is also 1 × 3. Basic Block C connects with the GlobalAveragePooling layer, then sets parameters of the Dropout layer to 0.5 to prevent the model from overfitting. In all convolutional layers, the step size is set to 1. Categorical Crossentropy is used as the loss function, Adam is chosen as the optimizer, the initial learning rate is set to 0.001, and the L2 regular term is added to the dense layer of the model to prevent overfitting.

3.2. The Proposed Fault Diagnosis Method

The fault diagnosis scheme proposed in this paper is mainly divided into three parts. The process is shown in Figure 7. First of all, the rules are extracted using the knowledge of equipment mechanism, and the rule nodes are matched with the node attributes to complete the entity extraction, then the rules are used as a classification basis and combined with the equipment operation data for labeling; the MOCNN model is trained through the training set and the test set is used for testing to complete the task of fault classification and to realize relation extraction. Finally, the triplet is constructed according to the entity and relationship, and the construction of the fault knowledge graph is completed through the relationship triplet.

3.3. Data Preparation

A large number of training samples can improve the generalization performance of the model, so this paper uses the data enhancement manner to expand the number of training samples. Taking the dataset in Section 4.1 as an example, the sliding segmentation method is used to achieve overlapping interception of bearing signals, with each training sample having a length of 1024 and a sliding interval of 238, resulting in a sample expansion of 500 samples in each category. The method maintains the continuity of the signal, and the principle is shown in Equation (11).
N = L L 0 S + 1
where N is the number of samples, L is the length of the original vibration signal, L 0 is the length of each training sample, and S is the sliding interval of the sliding window.

4. Experimental Study and Analysis

This experiment includes the construction of KG and the classification experiments of fault data by MOCNN. The knowledge graph is constructed by using the Neo4j graph database for storage, and py2neo is used to implement the python program to operate Neo4j. Fault classification is performed using the MOCNN model to classify the data, and the performance of the model is also discussed.

4.1. Construction of Knowledge Graphs

4.1.1. Entity Extraction

The dataset used for the experiments is the bearing dataset from Case Western Reserve University (CWRU), which is widely used in the field of fault diagnosis. The mechanistic knowledge of this dataset is mainly an introduction of the dataset, including motor load, motor speed, bearing fault location and fault size. The rules are summarized according to the dataset, which mainly include:
(1) The fault type is divided into the drive end with a sensor sampling frequency of 12 kHz and 48 kHz, the fan end with a sampling frequency of 12 kHz, and the normal type of data; (2) The fault size of the drive end and the fan end is divided into 0.007 inches, 0.014 inches, 0.021 inches, and some are 0.028 inches; (3) The faults on the drive side and fan side were generated at 0 HP,1 HP,2 HP, and3 HP motor loads, and the normal operation data were also collected under four loads corresponding to rotational speeds of 1797 r/min, 1772 r/min, 1750 r/min, 1730 r/min, respectively; (4) Bearing failures at the drive end and fan end include rolling element, inner ring and outer ring failures, respectively, and the outer ring includes failures at 3, 6 and 12 o’clock. Symbolizing the name according to the rules above, all together there are six levels of fault nodes: the zero fault node can be set as a test bed as the central node of the knowledge graph; the first level fault node (Level_1): Type, a total of 4 categories; the second level fault node (Level_2): Size, 10 types in total; Level 3 failure node (Level_3): Load, 44 types in total; Level 4 failure node (Level_4): Location, 119 types in total; Level 5 failure node (Level_5): Clock, 160 in total. The attributes of the nodes entered into the knowledge graph include Node ID, Node Label, Node Type, and Node Description. Some fault nodes are shown in Table 1:

4.1.2. Relation Extraction

The MOCNN model constructed in this paper is used to classify the fault data, and the data set is divided into four classes, 10 classes, 44 classes, 119 classes and 160 classes, respectively, according to the type of classification nodes for data labeling. 1024 data points are taken as a time series segment, and a sliding window is used for data enhancement, and then the training set and test set are randomly divided in the ratio of 7:3 and the results of model testing are shown in Figure 8 and Table 2:
As shown in Table 3, DE12 has the largest amount of data (30,000, 1024), while NO has the least amount of data (2000, 1024), which represents a big difference in the amount of data. However, it converges quickly and achieves 100% accuracy, which proves the advantage of the model. Next is the second level fault node classification, and no normal data is added to this level of classification as normal data does not have fault characteristics. As shown in Table 4, there are 10 categories of data. Although the amount of data is unbalanced, the model quickly reached 99% accuracy, with slight fluctuations compared to the level 1 fault classification curve; the third level fault classification curve converges more slowly compared to the second one, the accuracy reaches 97.63%; the accuracy of the fourth level fault classification shows a slight decrease due to a sharp increase in classification categories in the face of unbalanced data volumes. Nevertheless, the accuracy still reaches 93.35%, which is a very impressive result in practical engineering applications; the data volume of the fifth level fault classification is relatively more balanced since each category is (500, 1024), the accuracy reaches 97.86% and the other evaluation indexes are basically the same, achieving a better fault diagnosis result.
Confusion matrix plot of Level_1(Type) and Level_2(Size) classification results are shown in Figure 9.
The GRU network is a model that maintains the LSTM effect with a simpler structure, fewer parameters and better convergence performance. The algorithm proposed in this paper is compared with different models in Level_5(Clock) in the fifth-level classification fault experiment. The results are shown in Table 5.
From the results, it can be obtained that the model has a great improvement in various evaluation metrics compared to the traditional machine learning algorithms. Although the DRNN, GRU and Inception network also show great improvement, they still have a large gap compared with the model proposed in this paper.
It can be seen that the MOCNN model has the comprehensive advantages of high accuracy, fast convergence speed and stable classification in multi-classification problems and achieves a better fault diagnosis effect. To analyze in a specific way, the beginning convolution layer, BN layer and Basic Block A play the role of feature extraction, thereby improving the convergence speed; Basic Block B avoids gradient disappearance while extracting features, making the model more stable; Basic Block C performs feature extraction by assigning weights, thus improving the training speed and convergence speed of the model. The final Dropout layer and L2 regularity effectively mitigate the overfitting problem and improve the generalization ability of the model.

4.1.3. Graph Construction

The extracted nodes and relationships can be expressed through triples as shown in Table 6. The entity nodes and relationship nodes are imported into the Neo4j database using a Python program and the knowledge graph constructed can be manipulated utilizing py2neo and Cypher statements. The graph can show details of fault nodes and other nodes that are related to them can be clearly seen.
As shown in Figure 10, neo4j can set the number of levels and nodes displayed. The figure shows a total of 30 nodes with a total of four levels. The red point in the center is Level_0: Equipment, the four yellow points are Level_1, which are DE12, DE48, FE12, Normal, the orange is Level_2, and the blue is Level_3 and the connections at each level can be clearly observed. The Figure 11 shows some nodes intercepted when all nodes are displayed. When a node is selected, such as the outermost green node, the information of the node, such as id, label and name, will be displayed on the interface. At the same time, the Cypher language can be used to query the nodes.
For example, to query all nodes connected to the DE12_size7_load0 node, we can enter the command in neo4j.
Match p=(n:level_3{name: “DE12_size7_load0”})-:rel{relation: “belongs to”}]->() return p; then all the nodes connected to that node will be displayed. The graph can also perform a knowledge quiz to return information about faults to aid decision-making and improve the efficiency of fault diagnosis.

4.2. Model Performance Discussion

The model MOCNN achieves better fault diagnosis results with a large amount of data. This paper also tests the model in the conventional fault diagnosis experimental method, mainly for exploring the dataset under DE12 working conditions, and the data are processed and divided in the same way as described above.

4.2.1. Performance of MOCNN on Small Datasets

First, the data sets under 0 HP(1797), 1 HP(1772), 2 HP(1750), and 3 HP(1730) conditions are enhanced and divided to obtain the datasets Dataset A, Dataset B, Dataset C, and Dataset D, respectively. The four datasets are combined to obtain Dataset E, as shown in Table 7, and then the model is used to conduct 10 types of experiments. The results are shown in Table 8.
From Table 8, it can be seen that the accuracy rate of this model on the four small datasets has reached 100%, the diagnostic effect of 99.98% is also achieved on dataset E and the accuracy rate and loss curve can be obtained. The model converges quickly and then stabilizes after 50 epochs after a small fluctuation from the Figure 12.
Using T-SNE to visualize the changes in the classification results of MOCNN in different layers (under 3 HP conditions) as shown in Figure 13, the data in Figure 13a passes through the Input layer, and the cluttered distribution of the data can be observed. After passing through Basic Block A and Basic Block B, samples of the same type slowly converge together, and after passing through Basic Block C, a good classification effect has been achieved.

4.2.2. Anti-Noise Performance of MOCNN

Since the rolling bearing working environment is often accompanied by noise variations, a variable noise working environment is set up to verify that the method proposed in this paper has noise immunity. The definition formula of signal-to-noise ratio is shown in Equation (12). Next, we add different levels of noise to the above datasets A, B, C, and D for testing. It can be seen that the model has achieved 100% accuracy at 2 dB, 0 dB, and −2 dB, indicating that the model has good filtering and anti-noise functions. Meanwhile, the model also achieved 98.2% of the fault diagnosis results under the operating condition of −10 dB. The results are shown in Table 9 and Figure 14.
S N R d b = 10 log 10 P s i g n a l P n o i s e
P s i g n a l and P n o i s e are the power of the signal and the power of the noise, respectively.

4.2.3. Comparison with Other Algorithms

In order to fully illustrate the advantages of the model’s noise immunity function, it is compared with other algorithms adding different noises in Dataset D and Dataset E, adding −10 dB noise, and the results are shown in Table 10. It can be seen that the model has the highest fault diagnosis accuracy in all four noise environments and the best test results in Dataset E, which fully illustrates the superiority of the model’s noise immunity performance.
In order to fully illustrate the advantages of the model’s anti-noise function, the results of adding different noise to Dataset D compared with other algorithms are shown in Table 10 and Figure 15. Adding −10 dB noise to Dataset E for comparison, and the results are shown in Figure 16. It can be seen that the model has the highest fault diagnosis accuracy in the four noise environments, and the test results in Dataset E are also the best. Fully explain the superiority of the model’s anti-noise performance

4.2.4. Performance of the Model under Variable Load Conditions

With regard to the analysis of fault diagnosis results under variable load conditions, Datasets A, B and C, were trained at −10 dB signal-to-noise ratio, and then the fault diagnosis capability of the model under different conditions was verified by using test set data under three other load conditions. The results are shown in Figure 17.
The complexity of the model is an important index to measure the bearing fault diagnosis method. The time complexity determines the testing time of the model, and the space complexity determines the number of parameters of the model. Table 11 summarizes the number of parameters and testing time of each model, from which we can notice that MOCNN has a very light structure, uses a small number of parameters with better performance than Resnet, and has basically the same testing time as the Inception10 network, proving that MOCNN has a high parameter utilization rate, so this model is well suited for application in engineering practices.

5. Conclusions

In this paper, we propose a deep learning and knowledge graph fusion method for bearing fault diagnosis which combines the mechanistic knowledge of rolling bearings with vibration signal data to improve the efficiency of fault diagnosis. By using this method, the bearing data is firstly extracted by rules, entities are extracted, then the data is labeled and the optimized convolutional neural network model MOCNN is used to classify faults, realize relation extraction and obtain triad for knowledge graph construction to assist fault diagnosis decision and visualize the results. In addition, this paper also compares and discusses the models. Using the CWRU dataset, a new classification method can still achieve an accuracy of 97.86% when the faults are up to 160 classes, and it is fully compared with other models in environments such as multiple operating conditions and noise environments. The experimental results show that the one-dimensional fault diagnosis model proposed in this paper, regardless of the size of the dataset, not only has a fast convergence rate and stable operation, but also has the best noise immunity and migration performance, which proves the practicality of the fault diagnosis model proposed.
This study provides a new solution for fault diagnosis of mechanical equipment and also illustrates the difficulty of diagnosing compound faults in industrial equipment. The model in this paper is currently only for historical data, and in the future, we will also try real-time diagnosis and add other kinds of data to build a richer knowledge graph using deep learning algorithms.

Author Contributions

Conceptualization, Z.L. and Y.L.; methodology, Z.L.; software, Z.L.; validation, Z.L., Q.S. and Y.L.; formal analysis, Z.L.; investigation, Z.L.; resources, Y.L.; data curation, Z.L.; writing—original draft preparation, Z.L.; writing—review and editing, Y.L., Q.S., B.Q. All authors have read and agreed to the published version of the manuscript.

Funding

The work was supported by the National Key R&D Program of China (Grant No. 2020AAA0109300).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The experimental data are available from the Bearing Data Center of Case Western Reserve University deposited in https://engineering.case.edu/bearingdatacenter/welcome (accessed on 1 May 2022).

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  1. Cao, X.C.; Chen, B.Q.; Yao, B.; He, W.P. Combining translation-invariant wavelet frames and convolutional neural network for intelligent tool wear state identification. Comput. Ind. 2019, 106, 71–84. [Google Scholar] [CrossRef]
  2. Qian, Y.; Yan, R.; Gao, R.X. A multi-time scale approach to remaining useful life prediction in rolling bearing. Mech. Syst. Signal Process. 2017, 83, 549–567. [Google Scholar] [CrossRef] [Green Version]
  3. Lei, Y.; Yang, B.; Jiang, X.; Jia, F.; Li, N.; Nandi, A.K. Applications of machine learning to machine fault diagnosis: A review and roadmap. Mech. Syst. Signal Process. 2020, 138, 106587. [Google Scholar] [CrossRef]
  4. Li, X.; Zhang, W.; Ding, Q.; Li, X. Diagnosing Rotating Machines With Weakly Supervised Data Using Deep Transfer Learning. IEEE Trans. Ind. Inform. 2020, 16, 1688–1697. [Google Scholar] [CrossRef]
  5. Singhal, A. Official Google Blog: Introducing the Knowledge Graph: Things, not strings [EB/OL]. Google Inc (2012-05-02)[2020-02-28]. Available online: https://blog.google/products/search/introducing-knowledge-graph-things-not/ (accessed on 27 October 2022).
  6. Oramas, S.; Ostuni, V.C.; Noia, T.D.; Serra, X.; Di Sciascio, E. Sound and Music Recommendation with Knowledge Graphs. ACM Trans. Intell. Syst. Technol. 2016, 8, 21. [Google Scholar] [CrossRef] [Green Version]
  7. Li, R.; Dai, W.; He, S.; Chen, X.; Yang, G. A Knowledge Graph Framework for Software-Defined Industrial Cyber-Physical Systems. In Proceedings of the IECON 2019—45th Annual Conference of the IEEE Industrial Electronics Society, Lisbon, Portugal, 14–17 October 2019. [Google Scholar]
  8. Yang, H.; Liu, Y.; Meré, J.O.; Yang, H. Anticipating Stock Market of the Renowned Companies: A Knowledge Graph Approach. Complexity 2019, 2019, 9202457. [Google Scholar]
  9. Liu, Y.; Chen, W.; Qu, H.; Mahmud, S.M.H.; Miao, K. Weakly Supervised Image Classification and Pointwise Localization with Graph Convolutional Networks. Pattern Recognit. 2020, 109, 107596. [Google Scholar] [CrossRef]
  10. Young, T.; Hazarika, D.; Poria, S.; Cambria, E. Recent Trends in Deep Learning Based Natural Language Processing. arxiv 2017, arXiv:1708.02709. [Google Scholar]
  11. Baldi, P. Deep learning in biomedical data science. Annu. Rev. Biomed. Data Sci. 2018, 1, 181–205. [Google Scholar] [CrossRef]
  12. Liang, P.; Deng, C.; Wu, J.; Yang, Z.; Zhu, J.; Zhang, Z. Compound Fault Diagnosis of Gearboxes via Multi-label Convolutional Neural Network and Wavelet Transform. Comput. Ind. 2019, 113, 103132. [Google Scholar] [CrossRef]
  13. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  14. Zhu, J.; Nan, C.; Peng, W. Estimation of Bearing Remaining Useful Life based on Multiscale Convolutional Neural Network. IEEE Trans. Ind. Electron. 2018, 66, 1. [Google Scholar] [CrossRef]
  15. Mao, S.; Zhao, Y.; Chen, J.; Wang, B.; Tang, Y. Development of process safety knowledge graph: A Case study on delayed coking process. Comput. Chem. Eng. 2020, 143, 107094. [Google Scholar] [CrossRef]
  16. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  17. Wang, Z.; Wang, J.; Wang, Y. An intelligent diagnosis scheme based on generative adversarial learning deep neural networks and its application to planetary gearbox fault pattern recognition. Neurocomputing 2018, 310, 213–222. [Google Scholar] [CrossRef]
  18. Zou, F.; Zhang, H.; Sang, S.; Li, X.; He, W.; Liu, X.; Chen, Y. An anti-noise one-dimension convolutional neural network learning model applying on bearing fault diagnosis. Measurement 2021, 186, 110236. [Google Scholar] [CrossRef]
  19. Zhao, M.; Kang, M.; Tang, B.; Pecht, M. Deep Residual Networks With Dynamically Weighted Wavelet Coefficients for Fault Diagnosis of Planetary Gearboxes. IEEE Trans. Ind. Electron. 2018, 65, 4290–4300. [Google Scholar] [CrossRef]
  20. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  21. Zhao, M.; Tang, B.; Deng, L.; Pechtb, M. Multiple Wavelet Regularized Deep Residual Networks for Fault Diagnosis. Measurement 2019, 152, 107331. [Google Scholar] [CrossRef]
  22. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  23. Lei, Y.; Dogan, U.; Zhou, D.X.; Kloft, M. Data-dependent Generalization Bounds for Multi-class Classification. arxiv 2017, arXiv:1706.09814. [Google Scholar] [CrossRef] [Green Version]
  24. Patil, T.R.; Sherekar, S.S. Performance Analysis of Naive Bayes and J48 Classification Algorithm for Data Classification. Comput. Sci. 2013, 6, 256–261. [Google Scholar]
  25. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. arxiv 2015, arXiv:1502.03167. [Google Scholar]
  26. Yang, D.; Karimi, H.R.; Sun, K. Residual wide-kernel deep convolutional auto-encoder for intelligent rotating machinery fault diagnosis with limited samples. Neural Netw. 2021, 141, 133–144. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The common Inception network module and the structure of residual blocks: (a) Diagram of the Inception structure; (b) Diagram of the residual structure.
Figure 1. The common Inception network module and the structure of residual blocks: (a) Diagram of the Inception structure; (b) Diagram of the residual structure.
Entropy 24 01589 g001
Figure 2. Channel Attention Module.
Figure 2. Channel Attention Module.
Entropy 24 01589 g002
Figure 3. Spatial Attention Module.
Figure 3. Spatial Attention Module.
Entropy 24 01589 g003
Figure 4. The structures of modules Basic Block A and Basic Block B: (a) Basic Block A; (b) Basic Block B.
Figure 4. The structures of modules Basic Block A and Basic Block B: (a) Basic Block A; (b) Basic Block B.
Entropy 24 01589 g004
Figure 5. Basic Block C.
Figure 5. Basic Block C.
Entropy 24 01589 g005
Figure 6. The diagram of the proposed MOCNN structure.
Figure 6. The diagram of the proposed MOCNN structure.
Entropy 24 01589 g006
Figure 7. The process diagram of the proposed method for the bearing fault diagnosis.
Figure 7. The process diagram of the proposed method for the bearing fault diagnosis.
Entropy 24 01589 g007
Figure 8. Fault classification results at different levels of MOCNN.
Figure 8. Fault classification results at different levels of MOCNN.
Entropy 24 01589 g008
Figure 9. The confusion matrix of the MOCNN model for the Level_1 and Level_2 samples: (a) Level_1(Type); (b) Level_2(Size).
Figure 9. The confusion matrix of the MOCNN model for the Level_1 and Level_2 samples: (a) Level_1(Type); (b) Level_2(Size).
Entropy 24 01589 g009
Figure 10. Results of the fault type.
Figure 10. Results of the fault type.
Entropy 24 01589 g010
Figure 11. Node information of knowledge graph.
Figure 11. Node information of knowledge graph.
Entropy 24 01589 g011
Figure 12. Training accuracy and loss of the MOCNN model for the dataset E: (a) The classification accuracy of MOCNN on dataset E; (b) The loss of MOCNN on dataset E.
Figure 12. Training accuracy and loss of the MOCNN model for the dataset E: (a) The classification accuracy of MOCNN on dataset E; (b) The loss of MOCNN on dataset E.
Entropy 24 01589 g012
Figure 13. 3 HP two-dimension distribution images.
Figure 13. 3 HP two-dimension distribution images.
Entropy 24 01589 g013
Figure 14. Test results of MOCNN model under different working conditions and different noise environments.
Figure 14. Test results of MOCNN model under different working conditions and different noise environments.
Entropy 24 01589 g014
Figure 15. Test results of different algorithms in different noise environments.
Figure 15. Test results of different algorithms in different noise environments.
Entropy 24 01589 g015
Figure 16. Test results of different algorithms under −10 dB dataset E.
Figure 16. Test results of different algorithms under −10 dB dataset E.
Entropy 24 01589 g016
Figure 17. Comparison of results of different algorithms in off design experiments.
Figure 17. Comparison of results of different algorithms in off design experiments.
Entropy 24 01589 g017
Table 1. Some fault nodes of knowledge graph.
Table 1. Some fault nodes of knowledge graph.
Node IDNode TypeNode LabelNode Description
0Level_0EquipmentExperimental Equipment
1Level_1DE1212DriveEndFault
8Level_2DE12_size712DriveEndFault_FaultDiameter0.007
27Level_3DE12_size7_load012DriveEndFault_FaultDiameter0.007_Load0HP
93Level_4DE12_size7_load0_OR12DriveEndFault_FaultDiameter0.007_Load0HP_OuterRace
222Level_5DE12_size7_load0_OR_clock612DriveEndFault_FaultDiameter0.007_Load0HP_OuterRace_@6:00
Table 2. Fault classification results at different levels of MOCNN.
Table 2. Fault classification results at different levels of MOCNN.
Accuracy (%)Macro-Precision (%) Macro - F 1   Score   ( % ) Macro-Recall (%)Fault Type
Level_1(Type)1001001001004
Level_2(Size)99.9299.9399.2399.2310
Level_3(Load)97.6397.7897.6497.6544
Level_4(Location)93.3593.3593.3593.35119
Level_5(Clock)97.8697.8697.8697.86160
Table 3. The number of Level_1(Type) samples.
Table 3. The number of Level_1(Type) samples.
TypeDE12DE48FE12NO
Number30,00025,50022,5002000
Table 4. The number of Level_2(Size) samples.
Table 4. The number of Level_2(Size) samples.
TypeDE12_Size7DE12_Size14DE12_Size21DE12_Size28DE48_Size7DE48_Size14DE48_Size21FE12_Size7FE12_Size14FE12_Size21
Number10,000600010,000400010,000550010,00010,00065006000
Table 5. Results of the different algorithm models for the Level_5 samples.
Table 5. Results of the different algorithm models for the Level_5 samples.
MetricAccuracy (%)Macro-Precision (%) Macro - F 1   Score   ( % ) Macro-Recall (%)
Method
XGBoost19.3118.6317.3719.28
RF54.2054.6653.2354.38
DRNN70.2671.4768.4170.27
GRU83.5883.7783.6582.74
Inception1086.2386.9286.2785.67
MOCNN97.8697.8697.8697.86
Table 6. Partial fault relation nodes.
Table 6. Partial fault relation nodes.
Start EntityTail EntityRelation
DE12Equipmentbelongs to
DE12_size7DE12belongs to
DE12_size7_load0DE12_size7belongs to
DE12_size7_load0_ORDE12_size7_load0belongs to
DE12_size7_load0_OR_clock6DE12_size7_load0_ORbelongs to
Table 7. Division of CWRU dataset.
Table 7. Division of CWRU dataset.
Fault TypeClass LabelDataset A
Training/Test
Dataset B
Training/Test
Dataset C
Training/Test
Dataset D
Training/Test
Dataset E
Training/Test
DE12_size7_load0_BAB0.0073500/15003500/15003500/15003500/150014,000/6000
DE12_size14_load0_BAB0.0143500/15003500/15003500/15003500/150014,000/6000
DE12_size21_load0_BAB0.0213500/15003500/15003500/15003500/150014,000/6000
DE12_size7_load0_IRIR0.0073500/15003500/15003500/15003500/150014,000/6000
DE12_size14_load0_IRIR0.0143500/15003500/15003500/15003500/150014,000/6000
DE12_size21_load0_IRIR0.0213500/15003500/15003500/15003500/150014,000/6000
DE12_size7_load0_OROR0.0073500/15003500/15003500/15003500/150014,000/6000
DE12_size14_load0_OROR0.0093500/15003500/15003500/15003500/150014,000/6000
DE12_size21_load0_OROR0.0213500/15003500/15003500/15003500/150014,000/6000
NO_load0NO3500/15003500/15003500/15003500/150014,000/6000
Table 8. Test results of model MOCNN on the above five datasets.
Table 8. Test results of model MOCNN on the above five datasets.
Accuracy (%)Macro-Precision (%) Macro - F 1   Score   ( % ) Macro-Recall (%)
3 HP100100100100
2 HP100100100100
1 HP100100100100
0 HP100100100100
Mix Load99.9899.9899.9899.98
Table 9. Test results of MOCNN model in different working conditions and different noise environments.
Table 9. Test results of MOCNN model in different working conditions and different noise environments.
Load SNR2 dB0 dB−2 dB−4 dB−6 dB−8 dB−10 dB
Metric
0 HPAccuracy (%)10010010099.6699.4698.4798.2
Macro-precision (%)10010010099.6799.4898.4998.2
Macro-recall (%)10010010099.6699.4698.4998.2
Macro- F 1 score (%)10010010099.6799.4798.4898.2
1 HPAccuracy (%)10010010099.7399.7399.6198.66
Macro-precision (%)10010010099.7399.7499.6198.68
Macro-recall (%)10010010099.7399.7399.6098.68
Macro- F 1 score (%)10010010099.7399.7399.6098.68
2 HPAccuracy (%)10010010010010099.8799.8
Macro-precision (%)10010010010010099.8799.8
Macro-recall (%)10010010010010099.8799.8
Macro- F 1 score (%)10010010010010099.8799.8
3 HPAccuracy (%)10010010010010010099.46
Macro-precision (%)10010010010010010099.46
Macro-recall (%)10010010010010010099.46
Macro- F 1 score (%)10010010010010010099.46
Table 10. Test results of different algorithms in different noise environments.
Table 10. Test results of different algorithms in different noise environments.
SNRMetricMOCNNResnet-18Inception10VGG16Lenet5GRULSTM
10 dBAccuracy (%)10099.8998.6699.2610.0598.3320.33
Macro-precision (%)10099.8998.6699.300.9898.3115.13
Macro-recall (%)10099.8998.6099.2610.0598.3020.06
Macro- F 1 score (%)10099.8998.6299.271.7998.288.06
5 dBAccuracy (%)10099.8994.6799.269.899.9328.66
Macro-precision (%)10099.8995.3299.260.9899.9327.25
Macro-recall (%)10099.8994.5399.2610.0199.9328.48
Macro- F 1 score (%)10099.8994.3899.261.7899.930.1983
−5 dBAccuracy (%)99.9399.9395.298.9310.1399.9321.8
Macro-precision (%)99.9399.9395.5998.9510.1399.9313.83
Macro−recall (%)99.9399.9295.1598.9310.0399.9321.72
Macro- F 1 score (%)99.9399.9294.8398.921.8499.9313.57
−10 dBAccuracy (%)99.4696.4397.7398.0110.0394.1340.53
Macro-precision (%)99.4696.4297.7998.021.0295.2937.49
Macro-recall (%)99.4695.8897.6998.0110.0393.9939.98
Macro- F 1 score (%)99.4695.8397.7297.981.8194.1830.71
Table 11. Number of parameters and test time for the different algorithms.
Table 11. Number of parameters and test time for the different algorithms.
MOCNNResnet-18Inception10VGG16
Number of Parameter 2.6 × 10 5 3.9 × 10 6 5.4 × 10 4 1.3 × 10 7
Test Time/s0.6132.5980.5040.533
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, Z.; Li, Y.; Sun, Q.; Qi, B. Bearing Fault Diagnosis Method Based on Convolutional Neural Network and Knowledge Graph. Entropy 2022, 24, 1589. https://doi.org/10.3390/e24111589

AMA Style

Li Z, Li Y, Sun Q, Qi B. Bearing Fault Diagnosis Method Based on Convolutional Neural Network and Knowledge Graph. Entropy. 2022; 24(11):1589. https://doi.org/10.3390/e24111589

Chicago/Turabian Style

Li, Zhibo, Yuanyuan Li, Qichun Sun, and Bowei Qi. 2022. "Bearing Fault Diagnosis Method Based on Convolutional Neural Network and Knowledge Graph" Entropy 24, no. 11: 1589. https://doi.org/10.3390/e24111589

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop