Next Article in Journal
Parasitic Capacitance Effects on Active Clamp Flyback Output Characteristics: Application to IPOP Connection
Previous Article in Journal
Flyback Converter Using a D-Mode GaN HEMT Synchronous Rectifier
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fault Diagnosis of Tennessee Eastman Process with XGB-AVSSA-KELM Algorithm

1
Qianhu College, Nanchang University, Nanchang 330031, China
2
Information Engineering College, Nanchang University, Nanchang 330031, China
3
Chongqing Research Institute, Nanchang University, Chongqing 402660, China
*
Author to whom correspondence should be addressed.
Energies 2022, 15(9), 3198; https://doi.org/10.3390/en15093198
Submission received: 15 March 2022 / Revised: 21 April 2022 / Accepted: 25 April 2022 / Published: 27 April 2022

Abstract

:
In fault detection and the diagnosis of large industrial systems, whose chemical processes usually exhibit complex, high-dimensional, time-varying and non-Gaussian characteristics, the classification accuracy of traditional methods is low. In this paper, a kernel limit learning machine (KELM) based on an adaptive variation sparrow search algorithm (AVSSA) is proposed. Firstly, the dataset is optimized by removing redundant features using the eXtreme Gradient Boosting (XGBOOST) model. Secondly, a new optimization algorithm, AVSSA, is proposed to automatically adjust the network hyperparameters of KELM to improve the performance of the fault classifier. Finally, the optimized feature sequences are fed into the proposed classifier to obtain the final diagnosis results. The Tennessee Eastman (TE) chemical process is used to verify the effectiveness of the proposed method through multidimensional diagnostic metrics. The results show that our proposed diagnosis method can significantly improve the accuracy of TE process fault diagnosis compared with traditional optimization algorithms. The average diagnosis rate for 21 faults was 91.00%.

1. Introduction

Chemical process equipment has evolved over time to become more automated, complicated, and intelligent. Chemical processes frequently produce hazardous circumstances, such as toxic corrosion, as well as complicated equipment, making safety management difficult [1,2,3]. Furthermore, because the systems in the chemical process are intimately linked, a failure of one portion of the system can result in the entire system failing. Chemical process accidents result in numerous incidents of significant property and social losses each year [4,5,6,7]. As a result, there is a pressing need in the chemical industry to develop an intelligent fault detection and diagnosis system. However, the chemical process encompasses both the chemical industry and other process industries, a scope that is too broad to be discussed conveniently. Therefore, this paper chooses to focus on the TE process and to propose a fault diagnosis method applied to TE process.
There are three types of fault detection and diagnosis systems: quantitative, qualitative, and historical process data models [8,9,10]. Fault diagnostic methods based on historical process data, for example, process data or signals directly without the need for a complicated and exact mathematical model, avoiding a huge amount of prior information and time-consuming and difficult mathematical processes [11]. The amount of data available to researchers has expanded significantly as a result of the information and big data era, making it challenging for quantitative and qualitative methodologies to meet the requirements in terms of diagnosis accuracy and speed. As a result, data-driven fault diagnosis approaches are common in the chemical sector.
Data-based diagnostic methods and machine-learning-based diagnostic methods are the two broad categories of data-based diagnostic approaches [12]. Principal component analysis [13], partial least squares [14], independent component analysis [15], and Fisher discriminant analysis [16] are examples of statistics-based approaches. Some dimension-reduction methods, such as multi-scale principal component analysis [17] and kernel Fisher discriminant analysis [18], have made advancements in terms of the highly nonlinear correlation between variables and very complicated processes in chemical systems. However, as the number of data dimensions grows, the complexity of these statistics-based methodologies grows exponentially, resulting in dimension disaster [19,20].
With the development of clever algorithms in recent years, machine learning methods have become increasingly popular in defect diagnosis systems, supporting artificial neural networks [21], support vector machines [22] and other machine learning methods.
In 1943, Mcculloch and Pitts demonstrated that a neuron is a simple gate-valued device that performs a logic function [23]. In 1958, Rosenblatt proposed the perceptron, the first feed-forward neural network, which is a binary linear classifier based on an artificial neural network [24]. In 1982, Hopfield, a physicist at Caltech, proposed the HNN model [25]. Backward propagation, the method of making vector sweeps of all neurons in a perceptron at the same time, was independently proposed by several scientists in 1986 [26]. Rumelhart and McClelland devised a theory based on parallel distributed processing and developed a multilayer network-based back-propagation learning technique [27].
Extreme learning machine was proposed by Huang in 2006 as a new feedforward neural network training approach [28]. Extreme learning machines can train hundreds of times faster than typical BP networks and support vector machines. Huang et al. created KELM in 2011, which uses kernel functions instead of feature mapping in the hidden layer of ELM to speed up learning and improve the learning model’s stability and generalization [29].
Although progress has been achieved in the use of ELM in the detection of chemical process faults, there are still certain issues to be resolved. Some variables have a negligible effect on the results of the chemical process, while too many variables require computing resources. The correctness of the categorization model must be assessed after the process monitoring data have been obtained. According to long-term studies, many features are unrelated to the classification goals. Data features were sorted into three groups by John [30]: strongly correlated, weakly correlated, and irrelevant features. Feature selection involves finding a subset of features that can optimize and specify the evaluation criteria. Feature selection is applied in many fault diagnosis methods.
Among the commonly used feature selection algorithms, it is worth mentioning LDA, PCA, SVM, RFTB, GBDT, and so on. Ronald A. Fisher [31] in 1936 introduced the Linear Discriminant method, which is sometimes used to solve classification problems. The original linear discriminant was applied to binary classification problems and was later generalized in 1948 by C.R. Rao [32] to “multiclass linear discriminant analysis” or “multiple discriminant analysis”.
Lau et al. [33] used multi-scale principal component analysis (MSPCA) and Deng et al. [34] used deep principal component analysis (DePCA) to improve the adverse effects of traditional PCA on feature selection.
Onel et al. [35,36] proposed a Nonlinear Support Vector Machine-based feature selection algorithm, and the nonlinear support vector machine model outperforms existing support vector machine models in the literature in terms of detection accuracy and latency. It also uses feature selection techniques with minimal information loss compared to feature extraction techniques such as PCA.
The random forest was introduced by Breiman [37] in 2001. Random forest is a classification tree-based algorithm that can be used for multivariate classification and regression. RF methods have several advantages over other more commonly used multiple regression or classification methods. For example, it can account for interactions and non-linearities between variables. In addition, it can provide information about the statistical weights of each variable in the overall results [38,39,40].
The GBDT algorithm was of interest to researchers for its powerful predictive performance when it was proposed with the gradient boosting principle by Friedman [41], but the GBDT algorithm is an extremely memory-consuming algorithm and speed became the most criticized aspect. After years of development, XGBoost was proposed by Chen [42], which is an efficient implementation of the GBDT algorithm with many algorithmic and engineering improvements. xGBoost has been widely used in major contexts. For example, in 2015, 17 of the 29 winning solutions posted on Kaggle’s Blog 3 used XGBoost. Eight of these solutions used XGBoost alone to train their models, while most others combined XGBoost with neural networks. Further, the second most popular method, deep neural networks, was used in 11 solutions in the same year.
We felt it necessary to investigate the guiding implications of optimizing the KELM fault diagnosis feature selection sequence using the XGBoost algorithm, and the results of XGBoost, random forest tree bagger (RFTB) and Nonlinear Support Vector Machine (NSVM) feature extraction are later compared.
The Tennessee Eastman (TE) simulation platform is an open access, widely used platform to test control and troubleshooting models for complex industrial processes. Inspired by previous work, we propose a new method for chemical process fault diagnosis based on TE data. The method consists of three steps: feature extraction based on XGBOOST, the optimization-seeking algorithm AVSSA, and AVSSA-KELM fault detection based on fused depth features. The method flow is as follows:
(1)
Facing the time-varying, strongly coupled and nonlinear chemical dataset, we use XGBOOST to remove redundant information while retaining most of the intrinsic and discriminative information to prevent features of different classes from overlapping in some regions of the feature space;
(2)
A new algorithm AVSSA is proposed by introducing the adaptive variation process based on t distribution into SSA, which has a strong global search capability in the early iteration and a strong local search capability in the late iteration. We combine the AVSSA and KELM classifiers to automatically adjust the optimal network hyper-parameters of KELM;
(3)
Using the advantages of integrated learning and multi-core learning, AVSSA-KELM is proposed as the top classifier for fault diagnosis based on fused multi-domain features. The method has high generalization performance with good diagnostic capability and diagnostic speed.
This paper is structured as follows. Section 2 introduces the use of XGBoost for feature selection. Section 3 presents the methodology that introduces the principle of the AVSSA. Section 4 presents the methodology that introduces the principle of the KELM. Section 5 details the xgboost-based AVSSA-KELM fault diagnosis method. Section 6 describes the application of the proposed method to the TE process for verification. Section 7 compares the proposed method with other methods in terms of multidimensional indicators to demonstrate the superiority of the method. Section 8 provides a summary of this study.

2. eXtreme Gradient Boosting

Xgboost [43] (eXtreme Gradient Boosting) is a gradient-enhanced decision system whose base learner can be either a linear classifier or a tree, and in this paper we use the features of its tree model as a basis for quantifying the importance of each feature for feature selection.

2.1. Tree Model

For a given dataset, in the process of tree model construction, each layer greedily selects a feature segmentation point as a leaf node to maximize the gain of the whole tree after segmentation, which means that the more times a feature is segmented, the greater the benefit of the feature to the whole tree and the more important the feature is. Similarly, the greater the average gain of the feature each time it is segmented, the more important the feature is. The weight of each leaf node in the segmentation process can be expressed as w p i , q i , p i and q i are:
p i = z ^ ( t 1 ) l z i , z ^ ( t 1 ) q i = z ˜ ( t 1 ) 2 l z i , z ^ ( t 1 ) ,
where the training error l z i , z ^ i represents the difference between the target value z i and the predicted value z ^ i . In order to minimise the cost of the segmented tree, based on the weights of all leaf nodes and considering the gain of each feature as a segmentation point, there is:
Gain = left w + right w nosplit w .
Equation (2) illustrates that, for each split point, its ground benefit can be expressed as the total weight after the split (the sum of the total weight of the left subtree of the leaf node and the total weight of the right subtree) minus the total weight of the leaf node before the split.

2.2. Combination of Tree Models

As a non-parametric supervised learning model, the decision tree model is commonly used for classification, regression and feature extraction. The model does not require any a priori assumptions about the data and can quickly find decision rules based on the characteristics of the data. XGBoost uses an integration strategy based on the decision tree, using a gradient lift-off algorithm to continuously reduce the loss of the previously generated decision tree. XGBoost adds one tree at each iteration, and constructs a linear combination of K trees as:
z ^ i ( t ) = k = 1 K f k x i = z ^ i ( t 1 ) + f t x i , f k F ,
where F denotes the function space containing all trees, and f k x i denotes the weight of the ith sample in the kth tree that is assigned to the leaf node where it is located.

2.3. Importance Metrics

Importance metrics are a way to assess the importance of each feature in the feature set to which it belongs, and XGBoost constructs importance metrics based on the number of feature splits, FScore, the average gain value, AverageGain, and the average coverage of features, AverageCover. For the above three importance measures, there are:
Weight = | X | ,
Gain = Gain X ,
Cover = Cover X ,
where X is the set of the requested features classified to the leaf nodes: Weight is the node gain value at segmentation for each leaf node in X obtained from Equation (2), and cover is the number of samples falling at each node in X.

2.4. XGBoost Feature Extraction Algorithm

2.4.1. Algorithm Construction

We classify each dataset according to XGBoost, which is constructed as follows.
(1)
Starting from the root node. Traverse all features for each node;
(2)
For feature a i A , first sort by sample value a i ( 1 ) , , a i ( n ) , linearly scan a to determine the segmentation point with the best gain a i ( k ) , gain i ;
(3)
Select the feature with the highest gain from all the selected feature segmentation points to segment max Gain a 1 k i , , a m k n and update the mapping relationship from the samples to the tree nodes;
(4)
Segment all the way to the maximum depth, and calculate the residuals for constructing the next tree;
(5)
Integrate all the generated trees to complete the construction of the final tree model;
(6)
Derive the importance of the feature variables based on the number of decisions.
Pseudo-code for the basic steps, as shown in Algorithm 1.
Algorithm 1: XGBoost.
Energies 15 03198 i001

2.4.2. The XGBoost Parameter and the Segmentation Criterion

The feature selection hyperparameters mainly include the XGBoost parameter and the segmentation criterion, which are specified as follows:
XGBoost parameter (default value if not specified):
max_depth: 7, this parameter is controlled as the maximum depth of the tree. This value is used to control overfitting. the larger the max_depth, the more specific the model learns.
n_estimators: 80, this parameter controls the maximum number of trees generated, and also the maximum number of iterations. This value has an effect on the number of features selected in feature selection, but this is largely avoided by the selection criteria of this algorithm.
learning_rate: 0.1, this parameter controls the step size of each iteration, too large for accuracy and too small for slow operation.
subsample: 1.0, this parameter controls the proportion of random samples for each tree. By decreasing the value of this parameter, the algorithm will be more conservative and avoid overfitting. However, if this value is set too small, it may lead to underfitting. In this paper, it is very easy to underfit with a smaller value, so we set it to 1.
colsample_bytree: 0.5, which controls the proportion of columns (each column is a feature) that are randomly sampled per tree.
Partitioning method:
The partitioning method uses weight arrays of features (A), gain arrays of features (B), and cover arrays of features (C) as the data source, with weight as the core and gain and cover as the auxiliary. The features are sorted by weight from highest to lowest, and the top N features with 80 percent of Weight are taken as set A 1 . Then, gain and cover are sorted from smallest to largest, and the features with the lowest five percent of each are taken as sets B 1 and C 1 .
Feature sequences after feature selection = ( A 1 B 1 ¯ ) V ( A 1 C 1 ¯ ) .

3. Adaptive Variation Sparrow Search Algorithm

The Sparrow Search Algorithm (SSA) is an intelligent bionic optimization algorithm proposed by Xue J and Shen B. In their study [44], they concluded that SSA has global search ability and strong adaptability by comparing SSA with algorithms such as GWO, PSO, GSA, etc. They also pointed out that most of the sparrows in most of the experiments on the tested functions clustered towards the global optimum and did not fall into local extremes. However, for example, with the Damavandi function, most of the sparrows converge on local minima. In order to increase the population diversity and avoid the algorithm easily falling into the local optimum, we can make the sparrow jump out of the local optimum position and improve the global search ability of the algorithm by introducing the mutation process of adaptive t-distribution. Thus, we propose our new algorithm called the Adaptive Variation Sparrow Search Algorithm (AVSSA).

3.1. Algorithmic Bionics Principles

SSA is a new group intelligence optimization algorithm inspired by the foraging and anti-predatory behaviors of sparrows, and its bionic principle is as follows:
The sparrow population during foraging is divided into two parts: discoverers and followers, while scouts are also set up as an early warning mechanism. Discoverers are highly adapted and has a wide search range, guiding the population to search and forage. Followers follow the discoverers to forage for better adaptability. Meanwhile, sparrows also have anti-predatory behavior. When a sparrow searching on the periphery encounters a natural predator or other organism threatening the entire population, it will provide an early warning to the population to avoid the area and adjust its search location.

3.2. Sparrow Search Algorithm

Discoverers are responsible for determining the direction and area of search for food, while foraging, and have a larger search area than followers and a larger proportion of the population, typically 60–70% of the population. Whether all sparrows can find food preferentially depends on the value of the fitness function of each individual. At the same time, discoverers are mostly at the periphery of the population and are responsible for early warnings about the surrounding environment, namely, when a predator is discovered, the discoverers need to lead the followers to a safe location to continue the search.
Discoverers’ position updated:
x i d t + 1 = x i d t · e x p ( i α · T ) , R 2 < S T x i d t + Q · L , R 2 S T ,
where t is the current number of iterations; T is the maximum number of iterations; α is a uniform random number of (0, 1]; Q is the random number that obeys the standard normal distribution number of randomizers; L is a matrix with 1 × d elements of 1; R 2 [ 0 , 1 ] and S T [ 0.5 , 1 ] denote the warning value and safety value.
All remaining sparrows in the population acted as followers. If a discoverer has less food, namely a low fitness value, its current location is a poor foraging location, indicating that the current search direction is not suitable and it needs to change direction to another location. At the same time, as the followers follow the discoverer, when the discoverer finds food, followers can obtain some of the food or forage around the location.
Followers’ position updated:
x i d t + 1 = Q · e x p ( x w d t x i d t i 2 ) , i > n 2 x b d t + 1 + x i d t x b d t + 1 · A + · L , i n 2 ,
where x i d t is the position of the sparrow in the i th dimension at the t th iteration; x w d t is the worst position of the sparrow in the d th dimension at the t th iteration; x b d t + 1 is the optimal position of the sparrow in the d th dimension at the t + 1 th generation of the population; A is a 1 × d matrix, where each element is randomly assigned to −1 or 1; A + = A T ( A A T ) 1 ; L is a matrix with 1 × d elements of 1.
Some of the sparrows at the edge of the population are responsible for scouting, and they generally account for 10% to 20% of the population. In the face of the arrival of a threat, scouts can foresee the danger in time and send an alert to the population, at this time either the discoverers or followers need to adjust the search position away from the danger, into the safe area to search. At the same time, sparrows in the center of the population will move randomly, and when the alarm is issued, they will move closer to other sparrow positions to reduce the possibility of predation.
Scouts’ position updated:
x i d t + 1 = x b d t + β ( x i d t x b d t , f i f g ) x i d t + K x i d t x w d t f i f w + ϵ , f i = f g ,
where β is the control step parameter, obeying N(0, 1) random numbers; K is a random number and K [ 1 , 1 ] , which indicates the orientation of the sparrow’s movement, and is also a step control parameter; ϵ is a very small number to prevent the denominator from being 0; f i denotes the fitness value of the i th sparrow; f g and f w are the optimal and worst fitness values of the current sparrow population.

3.3. Adaptive Variation Process Based on t-Distribution

The probability density function of the t-distribution is as follows:
p t ( y ) = Γ n + 1 2 n π Γ n 2 1 + y 2 n n + 1 2 , < y < ,
where n is the degree of freedom.
In evolutionary planning, the Cauchy operator and Gaussian operator are the most commonly used variance operators, and the Cauchy and Gaussian distributions are special distributions at two boundary positions of the t-distribution. The standard Gaussian distribution density N (0, 1) has an expectation of 0 and a variance of 1. The standard Cauchy distribution probability density C (0, 1) has no expectation and the variance is infinite. According to the probability function of t distribution, t(n)→N(0, 1) when n , namely, when n , t-distribution is standard Gaussian distribution; t(n)→C(0, 1) when n = 1 , namely, when n = 1 , t-distribution is the standard Cauchy distribution. So the advantage of t-distribution is that the t-operator φ can be constructed, and the advantages of both types of operators can be exploited by bridging the Gaussian operator and the Cauchy operator through the t-operator.
In the variation operation, the original parameter values are replaced by a random number that fits a t-distribution with degrees of freedom equal to the number of iterations k. The variation formula is:
x i = x i + x i · φ ( k ) ,
where x i and x i are the positions of the i th sparrow before and after the t-distribution variation, respectively; k is the number of iterations; φ ( k ) is the t-operator with the number of iterations k as the degrees of freedom.
The adaptive t-distribution mutation-based sparrow search algorithm takes the number of iterations k of the sparrow algorithm is used as the degree of freedom of the t-distribution.
(1)
When k is small in the early iteration, the t-distribution is similar to the Cauchy mutation. The offspring produced by the Cauchy mutation have greater variability compared to the parent and are more perturbed than the Gaussian mutation, with a wider range of mutations, thanks to the Cauchy density function having a longer step length and a longer distribution at both ends giving it a higher probability of allowing the individual to escape the local optimum [45,46,47];
(2)
When k is large in the later iteration, the t-distribution resembles a Gaussian mutation. Since the peak of the Gaussian distribution curve is located where the mean value is, the search focuses on the local area near the original individual to enhance the local search capability [48,49,50].
This makes the t-distribution adaptive due to the change of iterations. Thus, we named this method the Adaptive Variation Sparrow Search Algorithm (AVSSA).
Based on the idealization and feasibility of the above model, the basic steps of the AVSSA can be summarized as the pseudo code shown in Algorithm 2.
Algorithm 2: The framework of AVSSA.
Energies 15 03198 i002

4. The Kernel Based Extreme Learning Machine

4.1. Extreme Learning Machine Overview

A typical single implicit layer feedforward neural network structure consists of an input layer, an implicit layer and an output layer, with the input layer fully connected to the implicit layer and the implicit layer to the output layer neurons [51]. The input layer has n neurons corresponding to n input variables, the hidden layer has l neurons, and the output layer has m neurons corresponding to m output variables. For the sake of generality, let the connection weights ω between the input and hidden layers be:
ω = ω 11 ω 12 ω 1 n ω 21 ω 22 ω 2 n ω l 1 ω l 2 ω l n ,
where ω denotes the connection weight between the ith neuron in the input layer and the jth neuron in the hidden layer.
Let the connection weight between the implicit layer and the output layer be β :
β = β 11 β 12 β 1 n β 21 β 22 β 2 n β l 1 β l 2 β l n ,
where self β j k denotes the connection weights between the jth neuron in the hidden layer and the kth neuron in the output layer.
Let the threshold value b of the neuron in the hidden layer be:
b = b 1 b 2 b l .
Let the input matrix X and output matrix Y of the training set with Q samples be:
X = x 11 x 12 x 1 n x 21 x 22 x 2 n x n 1 x n 2 x n Q
Y = y 11 y 12 y 1 n y 21 y 22 y 2 n y m 1 y m 2 y m Q .
Let the activation function of the neurons in the hidden layer be g ( x ) , the output T of the network is:
T = t 1 t 2 t Q n × Q
t j = t 1 j t 2 j t m j = i = 1 t β i 1 g ( ω i x j + b i ) i = 1 t β i 2 g ( ω i x j + b i ) i = 1 t β i m g ( ω i x j + b i ) m × 1 .

4.2. Kernel Based Extreme Learning Machine

The Kernel Based Extreme Learning Machine (KELM) [52,53] is an improved algorithm based on the Extreme Learning Machine (ELM) combined with a kernel function.
ELM is a single implicit layer feedforward neural network whose learning objective function F ( x ) can be represented by the matrix:
F ( x ) = h ( x ) × β = H × β = L ,
where x is the input vector, h ( x ) , H is the output of the hidden layer nodes, β is the output weight and L is the desired output.
Turning the network training into a problem solved by a linear system, β is determined according to β = H * · L , where H * is the generalised inverse matrix of H. To enhance the stability of the neural network, the regularisation factor c and the unit matrix I are introduced, so that the least squares solution for the output weights is:
β = H T ( H H T + I c ) 1 L .
Introducing the kernel function into the ELM, the kernel matrix is:
Ω E L M = H H T = h ( x i ) h ( x j ) = K ( x i , x j ) ,
where x i , x j is the test input vector, then Equation (19) can be expressed as:
F ( x ) = [ K ( x , x 1 ) ; ; K ( x , x n ) ] ( I c + Ω E L M ) 1 L ,
where ( x 1 , x 2 , , x n ) is the given training sample, n is the number of samples. K ( ) is the kernel function.

5. The Proposed XGB-AVSSA-KELM Method

We begin by utilizing the XGBoost to collect individual significance values to improve the model’s efficiency and accuracy. Then, based on the results, features are selected with redundant attributes deleted to reduce the number of network element nodes. Due to the network structure, the configuration of the regularization coefficients c and kernel function parameters S will have an impact on the classification performance of KELM. AVSSA has a unique algorithmic structure compared to previous metaheuristics, which provides a new method for balancing exploration and exploitation in the optimization process, and it can retain a high population diversity while boosting convergence efficiency. As a result, the AVSSA technique can be used to discover the best c and S to improve the network’s performance.
The process of the model is presented in Figure 1, and is described as follows:
Step 1: For training and prediction, input simulation data from the TE process into XGBoost;
Step 2: The importance values of XGBoost’s features are ranked;
Step 3: Select features and retrieve the dataset for the input network based on the ranking results;
Step 4: Set up the Kernel-Based Extreme Learning Machine and input the random regularization factor c as well as the kernel function parameter S;
Step 5: The optimal network hyper-parameters are obtained using the AVSSA algorithm;
Step 6: The optimal KELM diagnostic model is obtained by substituting the optimized regularization coefficients c kernel function parameters S into the KELM for training;
Step 7: The test samples are fed into the trained network to obtain the predicted output.

6. Experiment

6.1. Model Establishment

To detect and diagnose faults in the TE process database, the XGB-AVSSA-KELM model is used. Table 1 shows the errors found in the TE process database. Step changes in process variables, increased variability in process variables, and actuator faults are all linked to these faults (e.g., viscous valves). As a result, using a model, the data from the TE process simulation are utilized to diagnose and discover defects in the samples. In order to reduce redundant data and improve time series, the following significant variables of the TE process were chosen as assessment indicators for the diagnostic network in Table 4. The AVSSA algorithm was then integrated with KELM to produce the diagnostic model, which was trained using the optimized input data. We fed a test dataset into the trained diagnostic model in order to obtain classification results and confirm the model’s accuracy.

6.2. Tennessee Eastman Process

The Tennessee Eastman model is a simulation model built by Tennessee Eastman Chemical Company based on the actual chemical production process of the company [54]. Because it covers many typical phenomena in the actual complex chemical production process and contains common production failures in the chemical process, it is often used for troubleshooting studies in the industrial process field.
Figure 2 depicts the Tennessee Eastman process’s approximate schematic, which includes five primary units: reactor, condenser, compressor, stripper, and separator. The TE process uses four reactants, A, C, D, and E, as well as two products, G and H. Inertia component B and by-product F are also present. The specific meaning of manipulated variables and process measurements in the legend of the picture can be found in Table 2 and Table 3.
In the TE process, there are 12 manipulated variables and 41 process measurements. The dataset used for this experiment contains all 52 variables except for agitator speed, as this was not manipulated during the actual chemical process. Table 2 records the remaining 11 manipulated variables except for the agitator speed. Table 3 records the 41 process measurements. There are 22 training sets in all, including normal settings, for each fault. During a 24-h fault simulation, the fault training datasets were obtained. A 48-h running simulation was used to create the test datasets, with the problem being introduced at the 8-h mark. The sample time was set at three minutes. As a result, there are 480 failure training sets, and, consequently, fault training sets have a total of 480, while normal training sets have a total of 500. There are 960 samples in each fault test set, and the 161st sample is when they all start to fail.

6.3. Feature Selection

Feature selection is a key topic in defect detection and classification, and it has an impact on the model’s performance. The value of features in a model can be automatically evaluated using a gradient-enhanced decision system developed by Xgboost in order to find non-linear feature interactions and consistently extract significant features. For feature selection, the relevance score of an individual decision tree is generated using the tree model’s attributes as the basis for evaluating the value of each feature.
The feature selection result in fault 15 and fault 17, which have a normalized scale of 1, is shown in Figure 3 and Figure 4.
For each fault, XGBoost feature selection is performed, and we can obtain the feature selection sequence corresponding to each fault.
The feature selection sequence corresponding to each fault is shown in Table 4. The numbers 1 to 41 in the figure correspond to process measurements 1 to 41 in this Table 3; the numbers 42 to 52 in the figure correspond to manipulated variables 1 to 11 in Table 2.
In the figure, we can see that, in general, the number of features in the feature selection sequence is generally in the range of five to fifteen. Comparing the fifty-two feature values in the original feature sequence, the redundant and irrelevant features are effectively deleted and strongly correlated features are extracted, which reduces the computational resources required for fault diagnosis.
For faults 3, 5 and 9, we can see in the figure that the number of features in their feature selection sequence is much greater than for the other faults. It is well known that the difficulties in the TE process fault diagnosis include faults 3, 5 and 9, which correspond to the inability to effectively delete redundant and irrelevant features and extract strongly relevant features, resulting in too many features being selected for faults 3, 5 and 9.

7. Experimental Results

7.1. The Performance of Purposed Method

7.1.1. Effect of Feature Selection

To emphasize the significance of feature selection, using the t-SNE dimensionality reduction approach, the fault features were reduced to two dimensions and were visualized, with the results presented in Figure 5 and Figure 6.
Figure 5 shows fault 1, 2 and 3 selected by XGBOOST features. Faults 1 and 2 are barely mixed with other faults and have high identifiability, corresponding to faults 1 and 2, while faults 3 and 4 are completely mixed together, demonstrating that fault 3 without feature selection has a high identification rate with other faults with a high degree of similarity, corresponding to the ultra-low fault diagnostic rate of fault 3.
Figure 6 shows faults 1, 2, 3, and 4 selected by XGBOOST features, in which the faults 1, 2, 3 and 4 are not mixed together, indicating that fault identifiability has improved after feature extraction, and that faults 1, 2, 3 and 4 are more closely aggregated, indicating that data similarity within the faults has been improved by feature extraction, and that the commonality of each fault data point has been well identified, which corresponds to the increased fault diagnosis rate after feature extraction.

7.1.2. Fault Diagnostic Rate (FDR)

The values of true positive T P and true negative T N represent the number of observations representing the correct classification, while the false positive F P and false negative F N represent the number of misclassifications. The specific meanings they represent are shown in Table 5.
F D R = T P + T N T P + T N + F N + F P
As shown in Figure 7, the average FDR for the training dataset is 0.9992. This diagnostic model has exceptionally high diagnostic rates (above 0.99) for all faults in the training set, demonstrating the algorithm’s ability to spot the training set’s characteristics. The average FDR in the test dataset was 0.9100, demonstrating that the model works not just with the training dataset, but also with the test dataset. Faults 3, 9, 15, and 16 have long been recognized as long-standing issues in process diagnosis. Except for faults 3, 9, 15, and 16, the diagnostic rate for all faults in this paper is above 85%, with an average diagnosis rate of 94.61%. For fault 16, the algorithm used in this research improves the diagnosis rate to 0.9177, which is a significant improvement to fault 16. This study additionally improves the diagnostic rate for faults 3, 9, and 15 to 0.7677, 0.7296, and 0.6167, respectively, while retaining the total diagnostic rate. One of the superiorities of this method is the increase in diagnosis rate.

7.2. Performance Comparison

7.2.1. Convergence Speed Comparison

Figure 8 shows a comparison of the iteration curves for AVSSA (black line), Runge–Kutta [55] (red line), Hunger Games [56] (green line), African Vulture [57] (blue line) and Chameleon [58] (yellow line) for 21 faults. The data for all iteration curves are normalized using MAX-MIN and are presented in the figure as a line graph. Since the initial iteration curve was a straight line, it could not be normalized and is therefore assigned to a straight line with y = 0 in this work.
In Figure 8, it is clear that Chameleon (yellow line) and African Vulture (blue line) both use more iterations to find the optimal solution; Runge–Kutta (red line), AVSSA (black line), and Hunger games (green line) are all algorithms that find the optimal solution in fewer iterations, but they are all mixed together and difficult to see. Sparrow (Figure 9) and Hunger Games (Figure 10) are the fastest, Dragon Kuta (Figure 11) is in the center, while Chameleon (Figure 12) and African Vulture (Figure 13) are the slowest in terms of convergence speed.
The AVSSA algorithm has a smoother iteration curve and fewer lines than the others (see Figure 9, Figure 10 and Figure 11), which reflects the fact that the AVSSA algorithm has more cases of reaching the optimal value in the first iteration and demonstrates the AVSSA algorithm’s unique advantage in terms of convergence speed.

7.2.2. Feature Selection Algorithm Comparison

NSVM-AVSSA-KELM, RF-AVSSA-KELM and XGB-AVSSA-KELM were the three sets of experiments in which NSVM did not provide feature extraction for faults 3, 9 and 15, marked with *.
The results of the comparison of the three different feature selection algorithms are shown in Table 6, where the XGB-AVSSA-KELM algorithm is the best of the three, proving the superiority of XGB in the feature selection algorithm.

7.2.3. Ablation Experiments

In order to verify the relative importance of both XGBOOST and AVSSA relative to the XGB-AVSSA-KELM algorithm, and the superiority of XGB feature selection, six sets of experiments are set up in this paper. They are KELM, AVSSA-KELM, NSVM-AVSSA-KELM, RF-AVSSA-KELM XGB-KELM, and XGB-AVSSA-KELM. NSVM and RP are the two feature selection methods used by Xie [59] and Onel [35] respectively, where the results of feature selection are cited and combined with AVSSA-KELM for fault diagnosis in this paper.
Table 7 shows the experimental outcomes. Overall, the KELM algorithm has the lowest FDR of the six sets of tests, with the rest of the algorithms being higher, demonstrating that all of the methods performed have improved the FDR to some extent.
First, we explore the effect of the AVSSA algorithm on the combined model. According to Table 7, when KELM is compared with AVSSA-KELM, the FDR of AVSSA-KELM is improved by 9.63% compared with KELM; when XGB- KELM is compared with XGB-AVSSA-KELM, the FDR of XGB-AVSSA-KELM is improved by 15.61% compared with XGB- KELM. This indicates that the network hyperparameters of AVSSA-optimized KELM can effectively improve the correct rate of fault diagnosis.
Then, we explore the role of XGBOOST (XGB) feature selection on the combined model. According to Table 7, when XGB-KELM and KELM are compared, the FDR of XGB-KELM is improved by 4.21% compared with KELM; when XGB-AVSSA-KELM and AVSSA-KELM are compared, the FDR of XGB-AVSSA-KELM is improved by 10.19% compared with AVSSA-KELM. This indicates that XGBOOST for feature selection can effectively improve the correct rate of fault diagnosis.

7.2.4. Fault Diagnostic Rate Comparison

Table 8 and Figure 14 compare the performance of the proposed technique to that of other fault diagnostic methods in the TE process. Other methods are M-KFDA-SVM [60], EDBN-2 [61], GA-DCNN and RF-GA-CNN [62]. Because EDBN-2 does not include diagnostic rate data for faults 3, 9, and 15, and because faults 3, 9, and 15 are all difficult to diagnose in chemical processes, the FDR for all three is set to 60% in this work.
When compared to other fault diagnosis approaches, the model in this research performed the best, as shown in Table 8. AVSSA-KELM had the best overall classification performance based on XGBOOST feature selection, with a score of 0.91. The overall fault diagnosis rate in this paper was greater than 0.6.

8. Conclusions

In this paper, a deep learning-based fault diagnosis method for chemical processes is proposed. The method combines feature selection and dataset optimization strategies with a KELM classifier. First, the dataset is optimized by removing redundant features using the XGBOOST model. Second, a new optimization algorithm, AVSSA, is proposed to automatically adjust the network hyper-parameters of KELM to improve the performance of the fault classifier. Finally, the optimized feature sequences are input into the proposed classifier to obtain the final diagnosis results. The proposed diagnosis method is applied to the TE process, and the experimental results verify the applicability and effectiveness of the model. Compared with several other deep learning models, this method has higher accuracy and achieves the same accuracy with fewer iterations.
The method has achieved some results, but there are still some limitations that need to be improved in future work. Our optimization of a large number of network hyperparameters of KELM has led to a significant improvement in the diagnostic performance of KELM, but the upper limit of the accuracy of the KELM classifier is still limiting the diagnostic accuracy. Future work is needed to carry out a deeper structural optimization of the classifier. It is also necessary to detect faults in real plants, compare the “time” required to solve them with traditional fault diagnosis processes and explore comprehensive optimization for practical applications. In addition, for the extracted features, compared with the pure data-driven diagnosis method, we can combine the characteristics of the chemical process itself and consider the characteristics of the TE process to explore the chemical connection between the feature variables, which will be a new cross-optimization direction.

Author Contributions

Conceptualization, M.H. In addition, X.H.; methodology, M.H.; software, M.H.; validation, M.H. In addition, X.H.; formal analysis, M.H. In addition, X.H.; investigation, X.H. In addition, B.T.; resources, M.H., X.H., B.T. In addition, Z.D.; data curation, M.H., B.T. In addition, Z.D.; writing—original draft preparation, M.H. In addition, X.H.; writing—review and editing, M.H. In addition, X.H.; visualization, M.H. In addition, X.H.; supervision, M.H., B.T. In addition, B.T.; project administration, M.H., B.T. In addition, Z.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pagliaro, M. An industry in transition: The chemical industry and the megatrends driving its forthcoming transformation. Angew. Chem. Int. Ed. 2019, 58, 11154–11159. [Google Scholar] [CrossRef]
  2. Stoessel, F. Thermal Safety of Chemical Processes: Risk Assessment and Process Design; John Wiley & Sons: Hoboken, NJ, USA, 2021. [Google Scholar]
  3. Chen, C.; Reniers, G.; Khakzad, N. A dynamic multi-agent approach for modeling the evolution of multi-hazard accident scenarios in chemical plants. Reliab. Eng. Syst. Saf. 2021, 207, 107349. [Google Scholar] [CrossRef]
  4. Yang, Y.; Chen, G.; Reniers, G.; Goerlandt, F. A bibliometric analysis of process safety research in China: Understanding safety research progress as a basis for making China’s chemical industry more sustainable. J. Clean. Prod. 2020, 263, 121433. [Google Scholar] [CrossRef]
  5. Chen, C.; Khakzad, N.; Reniers, G. Dynamic vulnerability assessment of process plants with respect to vapor cloud explosions. Reliab. Eng. Syst. Saf. 2020, 200, 106934. [Google Scholar] [CrossRef]
  6. An, J.; Wang, Z.; Jiang, T.; Chen, P.; Liang, X.; Shao, J.; Nie, J.; Xu, M.; Wang, Z.L. Reliable mechatronic indicator for self-powered liquid sensing toward smart manufacture and safe transportation. Mater. Today 2020, 41, 10–20. [Google Scholar] [CrossRef]
  7. Amin, M.T.; Imtiaz, S.; Khan, F. Process system fault detection and diagnosis using a hybrid technique. Chem. Eng. Sci. 2018, 189, 191–211. [Google Scholar] [CrossRef]
  8. Venkatasubramanian, V.; Rengaswamy, R.; Yin, K.; Kavuri, S.N. A review of process fault detection and diagnosis: Part I: Quantitative model-based methods. Comput. Chem. Eng. 2003, 27, 293–311. [Google Scholar] [CrossRef]
  9. Venkatasubramanian, V.; Rengaswamy, R.; Kavuri, S.N. A review of process fault detection and diagnosis: Part II: Qualitative models and search strategies. Comput. Chem. Eng. 2003, 27, 313–326. [Google Scholar] [CrossRef]
  10. Venkatasubramanian, V.; Rengaswamy, R.; Kavuri, S.N.; Yin, K. A review of process fault detection and diagnosis: Part III: Process history based methods. Comput. Chem. Eng. 2003, 27, 327–346. [Google Scholar] [CrossRef]
  11. Xinyi, C.; Xuefeng, Y. Fault diagnosis in chemical process based on self-organizing map integrated with fisher discriminant analysis. Chin. J. Chem. Eng. 2013, 21, 382–387. [Google Scholar]
  12. Nor, N.M.; Hassan, C.R.C.; Hussain, M.A. A review of data-driven fault detection and diagnosis methods: Applications in chemical process systems. Rev. Chem. Eng. 2020, 36, 513–553. [Google Scholar] [CrossRef]
  13. Dunia, R.; Qin, S.J.; Edgar, T.F.; McAvoy, T.J. Identification of faulty sensors using principal component analysis. AIChE J. 1996, 42, 2797–2812. [Google Scholar] [CrossRef]
  14. Botre, C.; Mansouri, M.; Nounou, M.; Nounou, H.; Karim, M.N. Kernel PLS-based GLRT method for fault detection of chemical processes. J. Loss Prev. Process. Ind. 2016, 43, 212–224. [Google Scholar] [CrossRef]
  15. Xu, Y.; Deng, X. Fault detection of multimode non-Gaussian dynamic process using dynamic Bayesian independent component analysis. Neurocomputing 2016, 200, 70–79. [Google Scholar] [CrossRef]
  16. Yu, J. Localized Fisher discriminant analysis based complex chemical process monitoring. AIChE J. 2011, 57, 1817–1828. [Google Scholar] [CrossRef]
  17. Shi, X.; Lv, Y.; Fei, Z.; Liang, J. A multivariable statistical process monitoring method based on multiscale analysis and principal curves. Int. J. Innov. Comput. Inf. Control. 2013, 9, 1781–1800. [Google Scholar]
  18. Baudat, G.; Anouar, F. Generalized discriminant analysis using a kernel approach. Neural Comput. 2000, 12, 2385–2404. [Google Scholar] [CrossRef]
  19. Wang, C.; Li, H.; Zhang, K.; Hu, S.; Sun, B. Intelligent fault diagnosis of planetary gearbox based on adaptive normalized CNN under complex variable working conditions and data imbalance. Measurement 2021, 180, 109565. [Google Scholar] [CrossRef]
  20. Wang, N.; Li, H.; Wu, F.; Zhang, R.; Gao, F. Fault Diagnosis of Complex Chemical Processes Using Feature Fusion of a Convolutional Network. Ind. Eng. Chem. Res. 2021, 60, 2232–2248. [Google Scholar] [CrossRef]
  21. Pirdashti, M.; Curteanu, S.; Kamangar, M.H.; Hassim, M.H.; Khatami, M.A. Artificial neural networks: Applications in chemical engineering. Rev. Chem. Eng. 2013, 29, 205–239. [Google Scholar] [CrossRef]
  22. Mahadevan, S.; Shah, S.L. Fault detection and diagnosis in process data using one-class support vector machines. J. Process. Control 2009, 19, 1627–1639. [Google Scholar] [CrossRef]
  23. McCulloch, W.S.; Pitts, W. Bulletin of Mathematical Biophysics. Bull. Math. Biophys. 1943, 5, 115–133. [Google Scholar] [CrossRef]
  24. Rosenblatt, F. The perceptron: A probabilistic model for information storage and organization in the brain. Psychol. Rev. 1958, 65, 386. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Hopfield, J.J. Neural networks and physical systems with emergent collective computational abilities. Proc. Natl. Acad. Sci. USA 1982, 79, 2554–2558. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Hopfield, J.J.; Tank, D.W. Computing with neural circuits: A model. Science 1986, 233, 625–633. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  28. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  29. Huang, G.B.; Zhou, H.; Ding, X.; Zhang, R. Extreme learning machine for regression and multiclass classification. IEEE Trans. Syst. Man Cybern. Part B 2011, 42, 513–529. [Google Scholar] [CrossRef] [Green Version]
  30. John, G.H.; Kohavi, R.; Pfleger, K. Irrelevant features and the subset selection problem. In Machine Learning Proceedings 1994; Elsevier: Amsterdam, The Netherlands, 1994; pp. 121–129. [Google Scholar]
  31. Fisher, R.A. The use of multiple measurements in taxonomic problems. Ann. Eugen. 1936, 7, 179–188. [Google Scholar] [CrossRef]
  32. Rao, C.R. The utilization of multiple measurements in problems of biological classification. J. R. Stat. Soc. Ser. 1948, 10, 159–203. [Google Scholar] [CrossRef]
  33. Lau, C.; Ghosh, K.; Hussain, M.A.; Hassan, C.C. Fault diagnosis of Tennessee Eastman process with multi-scale PCA and ANFIS. Chemom. Intell. Lab. Syst. 2013, 120, 1–14. [Google Scholar] [CrossRef]
  34. Deng, X.; Tian, X.; Chen, S.; Harris, C.J. Deep principal component analysis based on layerwise feature extraction and its application to nonlinear process monitoring. IEEE Trans. Control Syst. Technol. 2018, 27, 2526–2540. [Google Scholar] [CrossRef]
  35. Onel, M.; Kieslich, C.A.; Pistikopoulos, E.N. A nonlinear support vector machine-based feature selection approach for fault detection and diagnosis: Application to the Tennessee Eastman process. AIChE J. 2019, 65, 992–1005. [Google Scholar] [CrossRef] [PubMed]
  36. Guzman, Y.A. Theoretical Advances in Robust Optimization, Feature Selection, and Biomarker Discovery. Ph.D. Thesis, Princeton University, Princeton, NJ, USA, 2016. [Google Scholar]
  37. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  38. Speiser, J.L.; Miller, M.E.; Tooze, J.; Ip, E. A comparison of random forest variable selection methods for classification prediction modeling. Expert Syst. Appl. 2019, 134, 93–101. [Google Scholar] [CrossRef] [PubMed]
  39. Vens, C.; Costa, F. Random forest based feature induction. In Proceedings of the 2011 IEEE 11th International Conference on Data Mining, Vancouver, BC, Canada, 11 December 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 744–753. [Google Scholar]
  40. Rodriguez-Galiano, V.F.; Ghimire, B.; Rogan, J.; Chica-Olmo, M.; Rigol-Sanchez, J.P. An assessment of the effectiveness of a random forest classifier for land-cover classification. ISPRS J. Photogramm. Remote Sens. 2012, 67, 93–104. [Google Scholar] [CrossRef]
  41. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  42. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd Acm Sigkdd International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  43. Parsa, A.B.; Movahedi, A.; Taghipour, H.; Derrible, S.; Mohammadian, A.K. Toward safer highways, application of XGBoost and SHAP for real-time accident detection and feature analysis. Accid. Anal. Prev. 2020, 136, 105405. [Google Scholar] [CrossRef]
  44. Xue, J.; Shen, B. A novel swarm intelligence optimization approach: Sparrow search algorithm. Syst. Sci. Control Eng. 2020, 8, 22–34. [Google Scholar] [CrossRef]
  45. Wang, W.C.; Xu, L.; Chau, K.W.; Xu, D.M. Yin-Yang firefly algorithm based on dimensionally Cauchy mutation. Expert Syst. Appl. 2020, 150, 113216. [Google Scholar] [CrossRef]
  46. Yang, Z.; Duan, H.; Fan, Y.; Deng, Y. Automatic carrier landing system multilayer parameter design based on Cauchy mutation pigeon-inspired optimization. Aerosp. Sci. Technol. 2018, 79, 518–530. [Google Scholar] [CrossRef]
  47. Sapre, S.; Mini, S. Opposition-based moth flame optimization with Cauchy mutation and evolutionary boundary constraint handling for global optimization. Soft Comput. 2019, 23, 6023–6041. [Google Scholar] [CrossRef]
  48. Song, S.; Wang, P.; Heidari, A.A.; Wang, M.; Zhao, X.; Chen, H.; He, W.; Xu, S. Dimension decided Harris hawks optimization with Gaussian mutation: Balance analysis and diversity patterns. Knowl.-Based Syst. 2021, 215, 106425. [Google Scholar] [CrossRef]
  49. Zhang, X.; Xu, Y.; Yu, C.; Heidari, A.A.; Li, S.; Chen, H.; Li, C. Gaussian mutational chaotic fruit fly-built optimization and feature selection. Expert Syst. Appl. 2020, 141, 112976. [Google Scholar] [CrossRef]
  50. Sarangi, A.; Samal, S.; Sarangi, S.K. Analysis of gaussian & cauchy mutations in modified particle swarm optimization algorithm. In Proceedings of the 2019 5th International Conference on Advanced Computing & Communication Systems (ICACCS), Coimbatore, India, 15–16 March 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 463–467. [Google Scholar]
  51. Huang, G.; Huang, G.B.; Song, S.; You, K. Trends in extreme learning machines: A review. Neural Netw. 2015, 61, 32–48. [Google Scholar] [CrossRef]
  52. Yang, Z.; Ce, L.; Lian, L. Electricity price forecasting by a hybrid model, combining wavelet transform, ARMA and kernel-based extreme learning machine methods. Appl. Energy 2017, 190, 291–305. [Google Scholar] [CrossRef]
  53. Yang, W.; Tian, Z.; Hao, Y. A novel ensemble model based on artificial intelligence and mixed-frequency techniques for wind speed forecasting. Energy Convers. Manag. 2022, 252, 115086. [Google Scholar] [CrossRef]
  54. Downs, J.J.; Vogel, E.F. A plant-wide industrial process control problem. Comput. Chem. Eng. 1993, 17, 245–255. [Google Scholar] [CrossRef]
  55. Ahmadianfar, I.; Heidari, A.A.; Gandomi, A.H.; Chu, X.; Chen, H. RUN beyond the metaphor: An efficient optimization algorithm based on Runge Kutta method. Expert Syst. Appl. 2021, 181, 115079. [Google Scholar] [CrossRef]
  56. Yang, Y.; Chen, H.; Heidari, A.A.; Gandomi, A.H. Hunger games search: Visions, conception, implementation, deep analysis, perspectives, and towards performance shifts. Expert Syst. Appl. 2021, 177, 114864. [Google Scholar] [CrossRef]
  57. Abdollahzadeh, B.; Gharehchopogh, F.S.; Mirjalili, S. African vultures optimization algorithm: A new nature-inspired metaheuristic algorithm for global optimization problems. Comput. Ind. Eng. 2021, 158, 107408. [Google Scholar] [CrossRef]
  58. Braik, M.S. Chameleon Swarm Algorithm: A bio-inspired optimizer for solving engineering design problems. Expert Syst. Appl. 2021, 174, 114685. [Google Scholar] [CrossRef]
  59. Xie, Z.; Yang, X.; Li, A.; Ji, Z. Fault diagnosis in industrial chemical processes using optimal probabilistic neural network. Can. J. Chem. Eng. 2019, 97, 2453–2464. [Google Scholar] [CrossRef]
  60. Nor, N.M.; Hussain, M.A.; Hassan, C.R.C. Fault diagnosis and classification framework using multi-scale classification based on kernel Fisher discriminant analysis for chemical process system. Appl. Soft Comput. 2017, 61, 959–972. [Google Scholar] [CrossRef]
  61. Wang, Y.; Pan, Z.; Yuan, X.; Yang, C.; Gui, W. A novel deep learning based fault diagnosis approach for chemical process with extended deep belief network. ISA Trans. 2020, 96, 457–467. [Google Scholar] [CrossRef] [PubMed]
  62. Deng, L.; Zhang, Y.; Dai, Y.; Ji, X.; Zhou, L.; Dang, Y. Integrating feature optimization using a dynamic convolutional neural network for chemical process supervised fault classification. Process. Saf. Environ. Prot. 2021, 155, 473–485. [Google Scholar] [CrossRef]
Figure 1. XGB-AVSSA-KELM algorithm flow chart.
Figure 1. XGB-AVSSA-KELM algorithm flow chart.
Energies 15 03198 g001
Figure 2. The TE process diagram.
Figure 2. The TE process diagram.
Energies 15 03198 g002
Figure 3. Feature selection for fault 15.
Figure 3. Feature selection for fault 15.
Energies 15 03198 g003
Figure 4. Feature selection for fault 17.
Figure 4. Feature selection for fault 17.
Energies 15 03198 g004
Figure 5. TSNE visual output of faults 1, 2, 3, 4.
Figure 5. TSNE visual output of faults 1, 2, 3, 4.
Energies 15 03198 g005
Figure 6. TSNE visual output of feature-selected faults 1, 2, 3, 4.
Figure 6. TSNE visual output of feature-selected faults 1, 2, 3, 4.
Energies 15 03198 g006
Figure 7. Fault diagnosis rate of XGB-AVSSA-KELM.
Figure 7. Fault diagnosis rate of XGB-AVSSA-KELM.
Energies 15 03198 g007
Figure 8. The iteration curves for the five methods.
Figure 8. The iteration curves for the five methods.
Energies 15 03198 g008
Figure 9. The iteration curves for AVSSA.
Figure 9. The iteration curves for AVSSA.
Energies 15 03198 g009
Figure 10. The iteration curves for the five methods.
Figure 10. The iteration curves for the five methods.
Energies 15 03198 g010
Figure 11. The iteration curves for AVSSA.
Figure 11. The iteration curves for AVSSA.
Energies 15 03198 g011
Figure 12. The iteration curves for the five methods.
Figure 12. The iteration curves for the five methods.
Energies 15 03198 g012
Figure 13. The iteration curves for AVSSA.
Figure 13. The iteration curves for AVSSA.
Energies 15 03198 g013
Figure 14. Comparison of FDR with other algorithms.
Figure 14. Comparison of FDR with other algorithms.
Energies 15 03198 g014
Table 1. The faults description of TE process.
Table 1. The faults description of TE process.
Variable NumberProcess VariableType
IDV(1)A/C feed ratio, B composition constant (stream 4)Step
IDV(2)B comoosition, A/C ratio constant (stream 4)Step
IDV(3)D feed temperature (stream 2)Step
IDV(4)Reactor cooling water inlet temperatureStep
IDV(5)Condenser c4mting water inIet temperatureStep
IDV(6)A feed loss (stream 1)Step
IDV(7)C header pressure loss-reduced availability (stream 4)Step
IDV(8)A,B,C feed composition (stream 4)Random variation
IDV(9)D feed temperature (stream 2)Random variation
IDV(10)C feed temperature (stream 4)Random variation
IDV(11)Reactor cooling water inlet temperatureRandom variation
IDV(12)Condenser cooling water inlet temperatureRandom variation
IDV(13)Reaction kineticsSlow drift
IDV(14)Reactor cooling water valveSticking
IDV(15)Condenser cooling water valveSticking
IDV(16)–(20)UnknownUnknown
IDV(21)Valve (stream 4)constant position
Table 2. Manipulated variables.
Table 2. Manipulated variables.
No.Manipulates VariablesNo.Manipulates Variables
1D feed flow valve7Separator pot liquid flow valve
2E feed flow valve8Stripper liquid product flow valve
3A feed flow valve9Stripper steam valve
4Total feed flow valve10Reactor cooling water flow
5Compressor recycle valve11Condenser cooling water flow
6Purge valve
Table 3. Process measurements.
Table 3. Process measurements.
No.Process MeasurementsNo.Process Measurements
1A feed22Separator cooling water outlet temperature
2D feed23A in reactor is feed
3E feed24B in reactor is feed
4Total feed25C in reactor is feed
5Recycle flow26D in reactor is feed
6Reactor feed rate27E in reactor feed
7Reactor pressure28F in reactor feed
8Reactor level29A in reactor feed
9Reactor temperature30B in reactor feed
10Purge rate31C in reactor feed
11Product separator temperature32D in reactor feed
12Product separator level33E in reactor feed
13Product separator pressure34F in reactor feed
14Product separator underflow35G in reactor feed
15Stripper level36H in reactor feed
16Stripper pressure37D in product flow
17Stripper underflow38E in product flow
18Stripper temperature39F in product flow
19Stripper steam flow40G in product flow
20Compressor work41H in product flow
21Reactor cooling water outlet temperature
Table 4. The result of feature selection.
Table 4. The result of feature selection.
FaultThe Result of Feature Selection
11, 16, 18, 39, 7, 13, 4, 6, 3, 20, 44, 9
210, 20, 47, 5, 7, 41, 15, 16, 9, 3, 49, 1, 25, 24, 30, 46
321, 16, 7, 6, 1, 40, 17, 18, 9, 47, 33, 14, 38, 4, 35, 52, 50, 29, 51, 34, 20, 13, 32, 46, 45, 12
418, 11, 7, 30, 51, 40, 38, 34, 20, 1
552, 16, 17, 20, 7, 11, 39, 18, 43, 14, 19, 28, 37, 13, 50, 41, 33, 1, 25, 35, 38, 8, 49, 3, 48, 9, 45, 12,
40, 4, 32, 36, 47, 24, 2
621, 11, 1, 20, 2, 13, 9, 14, 7, 28, 10, 19, 29, 44, 22, 46, 23
719, 38, 45, 46, 50
829, 1, 41, 16, 20, 44, 40, 23, 10, 47, 39, 7, 38, 14, 35, 19
97, 20, 16, 13, 3, 29, 11, 21, 1, 36, 37, 2, 33, 40, 10, 30, 46, 18, 26, 41, 25, 50, 34, 32, 19, 9, 6, 38,
39, 24, 35, 15, 23, 45, 12
1016, 19, 7, 18, 50, 6, 38, 3, 25
1151, 16, 20, 7, 6, 9, 22
1218, 3, 7, 19, 13, 16, 11, 20, 2
1319, 21, 50, 39, 7, 32, 51, 16, 1, 38
1421, 7, 51, 16, 6, 9, 34
1516, 20, 7, 29, 6, 30, 19, 21, 25, 35, 11, 18, 40
1616, 50, 20, 7, 19, 6, 32, 18
1716, 7, 21, 6
1816, 7, 18, 19, 41, 11, 3, 50, 22, 1, 25
1916, 7, 5, 20, 30, 46, 18, 6
2016, 7, 13, 21, 20, 46
2118, 16, 30, 6, 11, 7, 19
Table 5. Confusion Matrix.
Table 5. Confusion Matrix.
Confusion MatrixPrediction
Positive (p)Negative (N)
RealTrue(T)TPFN
Flase(F)FPTN
Table 6. The feature selection algorithm comparison table.
Table 6. The feature selection algorithm comparison table.
FaultNSVM-AVSSA-KELMRF-AVSSA-KELMXGB-AVSSA-KELM
10.99790.99691.0000
20.97290.97501.0000
3*0.54790.7677
41.00001.00001.0000
50.96870.98441.0000
61.00000.98441.0000
71.00001.00001.0000
80.90720.81980.9406
9*0.51980.7240
100.66250.73330.8938
110.81970.83540.9469
120.83750.89690.8969
130.73950.78960.9167
140.91140.99901.0000
15*0.53650.6167
160.70930.75940.9177
170.74370.84690.9479
180.90310.88230.9021
190.80000.86880.9063
200.79580.79690.8635
210.97080.67600.8698
Average0.8745 *0.83090.9100
Table 7. Results of ablation experiments with preserved KELM.
Table 7. Results of ablation experiments with preserved KELM.
FaultKELMAVSSA-KELMXGB-KELMXGB-AVSSA-KELM
10.99380.99900.99271.0000
20.97810.98130.98131.0000
30.59380.63540.48440.7677
41.00001.00001.00001.0000
50.48130.71460.97081.0000
60.99580.99580.99581.0000
71.00001.00001.00001.0000
80.64170.86350.61150.9406
90.72290.72400.61980.7240
100.60420.65520.56670.8938
110.55210.78650.61250.9469
120.47810.73750.44380.8969
130.60000.70310.61560.9167
140.81350.95520.99271.0000
150.43960.48650.64270.6167
160.61770.72910.66040.9177
170.68330.76250.69790.9479
180.90520.89900.87710.9021
190.60310.83330.75520.9063
200.73020.73850.74060.8635
210.51350.76980.56980.8698
Average0.71180.80810.75390.9100
Table 8. Comparison of FDR with other algorithms.
Table 8. Comparison of FDR with other algorithms.
FaultM-KFDA-SVMEDBN-2GA-DCNNRF-GA-CNNProposed Method
10.99801.00000.99310.99561.0000
20.99871.00000.98250.99061.0000
30.43330.60000.93560.95810.7677
40.95891.00000.99380.99811.0000
50.99501.00000.89620.62061.0000
60.96021.00000.97860.98931.0000
70.99441.00000.99881.00001.0000
80.95810.98290.93560.92690.9406
90.51560.60000.67880.77620.7240
100.93990.80810.96440.97060.8938
110.95380.99740.98310.98750.9469
120.95631.00000.97000.97880.8969
130.95240.91980.94620.94560.9167
140.98051.00000.99060.99501.0000
150.64740.60000.47750.43500.6167
160.97800.75560.43250.20690.9177
170.97321.00000.94310.94880.9479
180.95150.93430.93250.94560.9021
190.96870.95530.98560.99000.9063
200.93630.93170.92560.93560.8635
210.87640.83440.85000.89300.8698
Average0.90130.90090.89500.88040.9100
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hu, M.; Hu, X.; Deng, Z.; Tu, B. Fault Diagnosis of Tennessee Eastman Process with XGB-AVSSA-KELM Algorithm. Energies 2022, 15, 3198. https://doi.org/10.3390/en15093198

AMA Style

Hu M, Hu X, Deng Z, Tu B. Fault Diagnosis of Tennessee Eastman Process with XGB-AVSSA-KELM Algorithm. Energies. 2022; 15(9):3198. https://doi.org/10.3390/en15093198

Chicago/Turabian Style

Hu, Mingfei, Xinyi Hu, Zhenzhou Deng, and Bing Tu. 2022. "Fault Diagnosis of Tennessee Eastman Process with XGB-AVSSA-KELM Algorithm" Energies 15, no. 9: 3198. https://doi.org/10.3390/en15093198

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop