Next Article in Journal
A Computational Framework for Design and Optimization of Risk-Based Soil and Groundwater Remediation Strategies
Previous Article in Journal
Stability Analysis of a Chemostat Model for Phenol and Sodium Salicylate Mixture Biodegradation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fault Diagnosis Using Dynamic Principal Component Analysis and GA Feature Selection Modeling for Industrial Processes

Automation College, Hangzhou Dianzi University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Processes 2022, 10(12), 2570; https://doi.org/10.3390/pr10122570
Submission received: 27 October 2022 / Revised: 13 November 2022 / Accepted: 24 November 2022 / Published: 2 December 2022
(This article belongs to the Section Chemical Processes and Systems)

Abstract

:
With the continuous expansion of industrial production scale, most of the chemical process variables are nonlinear, multi-modal and dynamic. For some traditional multivariate statistical monitoring and fault diagnosis algorithms, such as principal component analysis (PCA), the premise of its application is that the process data is time-independent. To this end, a dynamic principal component analysis (DPCA) method is proposed. However, since the input matrix of DPCA fault diagnosis needs to add an augmented matrix to the original data matrix, the number of eigenvalues of the augmented matrix is too large and there are many redundant eigenvectors. Therefore, this paper proposes a fault diagnosis and monitoring algorithm combining feature selection and DPCA, which considers the dynamic characteristics of multivariate data and reduces the dimension of the input matrix. At present, the average modeling and diagnostic accuracy of PCA-based fault diagnosis on T2 statistic is 65.49%, and that on Q statistic is 76.78%. The average modeling and diagnostic accuracy of fault diagnosis based on DPCA on T2 statistic is 63.17%, and the average modeling and diagnostic accuracy on Q statistic is 83.65%. Finally, through a TE simulation process, this paper proves that the accuracy is greatly improved when using the method proposed in this paper compared with PCA and DPCA.

1. Introduction

Due to the rapid development of science and technology, the scale of modern industrial processes is also expanding. During the operation of the chemical process, if the equipment fails or is abnormal and is not repaired in time, extreme conditions, such as high temperature and high pressure, toxic gas release, and even explosion, may occur [1]. This will not only cause huge economic losses to the factory, but also greatly endanger the lives of workers. The emergence of fault diagnosis and monitoring technology has greatly ensured high-quality production in the chemical industry and reduced the frequency of safety accidents. At present, the mainstream research methods of fault diagnosis and monitoring technology mainly include data-based multivariate statistics and artificial intelligence-based deep learning methods [2,3,4,5].
Process monitoring technology based on multivariate statistics mainly reduces the dimension of offline data and online data generated in the production process through multivariate statistical analysis. Deep learning mainly uses the powerful learning ability of the computer, the multi-layer hidden layer network, and the layer-by-layer learning method to extract the main information from the input data, and, through a lot of training, to improve the fault accuracy diagnosis [6]. In recent years, many scholars at home and abroad have proposed a variety of fault diagnosis methods and achieved good results. However, industrial process monitoring and fault diagnosis remain a challenging problem due to the large number of complex properties in industrial processes, such as dynamics, nonlinearity, multimodality, and heavy coupling [7,8,9,10].
Common multivariate statistics-based methods include principal component analysis (PCA) [11], linear discriminant analysis (LDA) [12], independent component analysis (ICA) [13], and partial least squares (PLS) [14,15]. Aiming at the dynamic features appearing in chemical process data, Ku et al. proposed a dynamic principal component analysis (DPCA) method by extending the time augmentation matrix [16]. To address the nonlinear features present in the data, Kernel Principal Component Analysis (KPCA) is proposed. This method uses a nonlinear mapping function to map the original data space to a higher-dimensional space, and then performs dimensionality reduction processing through PCA [17,18].
Most of the dimensionality reduction algorithms belong to feature extraction. Through the dimensionality reduction matrix, a new set of feature variables is obtained, which changes the original feature space, often resulting in a loss of some important feature variables in the original feature space. Feature selection, on the other hand, selects a subset of the original feature space that best represents the original feature space without destroying the original data. For data with too large a latitude, finding the optimal feature subset can be regarded as an optimal problem. Through the heuristic optimization algorithm, each feature subset can be uniformly evaluated, and the optimal solution can be finally obtained.
Aiming at certain shortcomings of some existing algorithms, this paper proposes a fault diagnosis method that is built on GA feature selection based on dynamic principal component analysis (GA-DPCA). The main improvements of this paper are as follows:
(1)
First, the feature selection based on genetic algorithm is adopted, which can quickly find the optimal feature subset, which not only refines the original data, but also loses the integrity of the data as little as possible.
(2)
The data matrix after feature selection is extended with time delay and the residual space of feature selection is added, which effectively considers the autocorrelation and integrity of the data.
(3)
Sliding window filtering technology is adopted to remove the noise existing in the data itself.
(4)
An experimental comparative study of the proposed method and existing methods is carried out using the Tennessee Eastman process and the actual coking process.

2. PCA and DPCA

PCA is the most widely used data dimensionality reduction algorithm. The main idea of PCA is to map n-dimensional features onto k-dimensions, which is a brand-new orthogonal feature called principal components. It is a k-dimensional feature reconstructed on the basis of the original n-dimensional feature. After mapping, only the features with large variance are selected, and the features with almost zero variance are ignored to achieve dimensionality reduction of the original data features.
Assuming the normalized data are X = x 1 , x 2 , x n T R n × m , the formula has n variables and m samples. p is the projection vector, y i = p T x i , and the objective function of PCA is:
J P C A p = max ( 1 n i ( y i ) ( y i ) T ) s . t . p T p = 1
The data formula can be decomposed into:
X = T P T + E
C = 1 n X X T
where X is the normalized data; P m × k is the loading matrix, which can be obtained by eigen decomposition of C ; T n × k is the score matrix; and E is the residual matrix. k represents the number of principal components, which can be generally calculated by the cumulative variance contribution rate [19]:
i = 1 k λ i / i = 1 m λ i × 100 % σ
where λ i is the eigenvalue of the C eigenvalue decomposition and arranges it from large to small, and σ is the ratio of the sum of the maximum eigenvalues to all the eigenvalues, generally with a value of 85%.
Fault diagnosis mainly uses two statistics, T 2 and Q, to judge whether a fault has occurred. T 2 statistic, also known as Hotelling’s T 2 , mainly reflects the changes of the spatial characteristics of the principal components, which are defined as follows [20]:
T 2 = x i T P S 1 P T x i
The Q-statistic, also known as the Squared Prediction Error Index (SPE), mainly reflects the changes in the subspace characteristics of residuals, which are defined as follows:
S P E = e T e
The control limits for T 2 and Q statistics are D C and Q C , a , respectively, and the formula is as follows:
D C ~ l ( n 2 1 ) n ( n 1 ) F α ( l , n l )
Q C , α = θ 1 1 θ 2 h 0 1 h 0 θ 1 2 + z α 2 θ 2 h 0 2 θ 1 1 / h 0 θ i = j = l + 1 m λ j i i = 1 , 2 , 3 h 0 = 1 2 θ 1 θ 3 3 θ 2 2
Among them, n is the number of modeled data samples; l is the number of principal components retained in the principal components; α is the level of significance; the critical value of F distribution under l , n l is found in the statistical table; C α is the critical value of the normal distribution at the significant level α ; and λ j is the eigenvalue of the smaller data covariance matrix.
The DPCA algorithm requires time-delay expansion of the original data space matrix. Assuming that the current data are related to observations from previous L-step data, the expansion matrix is [21]:
X L = X ( t ) , X ( t 1 ) , , X ( t L )

3. GA-DPCA

GA feature selection can quickly obtain the optimal feature subset of the original data matrix and reduce the original data without destroying it. The DPCA algorithm can take into account the dynamic characteristics of the data and further reduce the dimension of the data matrix at the same time. Therefore, this paper proposes a dynamic principal component analysis method based on GA feature selection for process monitoring and fault diagnosis through TE datasets. The method flow is shown in Figure 1. The first part is sliding window denoising, the second part is feature selection based on a genetic algorithm, the third part is offline modeling and determination of control limits, and the last is online monitoring.

3.1. Sliding Window Removal Noise

In industrial process production, data acquisition is often accompanied by a large amount of noise interference, which makes the acquired data inaccurate, thus affecting the accuracy of modeling. In order to suppress noise interference and improve modeling accuracy, we need to use appropriate filtering algorithms to filter the data. This paper mainly adopts the filtering algorithm of sliding window denoising [22]. As shown in Figure 2, the basic idea is to set a sliding window with a fixed width, slide along the time series, and use the arithmetic mean of the data in the window as the output value of the new series, that is, the filtered series. N is set to the width of the sliding window. If N = 2k + 1, and the input and output are x n and y n , respectively, then:
y n = 1 2 k + 1 i = k i = k x ( n + i )
A piece of data is taken from the training set of the TE dataset and filtered through a sliding window. The results are shown in Figure 3, which shows that the sliding window filtering algorithm can effectively suppress high-frequency noise.

3.2. Feature Selection Based on GA

Feature selection is an important question in feature engineering. Its goal is to find the optimal feature subset. Feature selection can eliminate irrelevant or redundant features, thereby reducing the number of features and improving model accuracy. According to the form of feature selection, it can be divided into three categories: filtering, encapsulation and embedding.
The general process of feature selection is: (1) generating subsets; (2) evaluating functions; (3) stopping criteria; (4) validation process [23,24]. For high-latitude data, finding the optimal solution can take a lot of time. Genetic algorithm is an adaptive global optimization search algorithm, which is a computational model that simulates the natural selection and genetic mechanism of Darwin’s theory of biological evolution. Therefore, the feature selection based on genetic algorithm will greatly increase the speed of obtaining the optimal solution. As shown in Figure 4, the general steps of GA-based feature selection are mainly divided into the following steps: initializing the population, calculating the fitness, setting the genetic operator, and updating the iterations [25,26].

3.2.1. Initializing the Population

Genetic algorithm follows the law of survival of the fittest and elimination of the weak, that is, retaining individuals with strong fitness and eliminating individuals with weak fitness. Before starting the genetic algorithm, there are several parameters that need to be set in advance: the first is the size of the population. If the population size is too small, inbreeding will obviously occur, resulting in genetic disease. Conversely, if the population size is too large, the results will be difficult to converge, resources will be wasted, and the robustness will be reduced. The second parameter is the number of iterations. If the evolutionary algebra is too small, the algorithm is not easy to converge, and the population is immature. If the evolutionary algebra is too large, the population may converge too early, and it is pointless to continue evolution, which would only increase time consumption and waste resources. Further, there is the probability of mutation. If the probability of mutation is too small, the diversity of population genes cannot be guaranteed. If the probability of mutation is too large, the genetic algorithm will degenerate into random search. Similar to the mutation probability, if the mating probability is too large, the existing beneficial individuals will be destroyed, the randomness will increase, and the optimal individual will be easily lost. If the mating probability is too small, the population cannot be effectively updated. The last is the catastrophe count. When the population has not updated the optimal solution for N consecutive generations, it is necessary to set the catastrophe to prevent the population from falling into the local optimal solution. If the value of N is too small, the search path will be greatly increased or even impossible. If the value of N is too large, the function of the mutation operator will become small or even useless, so an appropriate value should be set.
Genetic coding is required before an individual can be produced. The coding method directly affects the operation of genetic operators, such as the crossover operator and mutation operator of the genetic algorithm, thus determining the efficiency of genetic evolution to a large extent. In this post, feature retention and loss is essentially a 0–1 problem. Assuming that the processed data are X = x 1 , x 2 , , x n , where x n is the feature of the data, if 1 means that this eigenvector is retained, then 0 means rounding off that eigenvector, the entire genotype of the individual is a binary encoded symbol string. For binary encoding, its advantages are that it is simple, it easily implements operations such as crossover and mutation, and it conforms to the principle of minimum character set encoding. The algorithm can be theoretically analyzed by using the mode theorem. Since the feature selection problem belongs to the type of planning problem, it is very suitable for binary encoding.

3.2.2. Calculating Fitness

Individual fitness refers to the degree of dominance of an individual in the survival of the population, and is used to distinguish the “good” and “bad” individuals. The fitness of an individual can be calculated using the fitness function. The fitness function, also known as the evaluation function, is mainly used to judge the fitness of an individual through individual characteristics. The general process for evaluating individual fitness is as follows:
  • After decoding the individual code string, the individual phenotype can be obtained.
  • The objective function value of the corresponding individual can be calculated from the individual’s phenotype.
  • According to the type of optimization problem, the fitness of individuals can be calculated from the objective function value according to certain transformation rules.
The purpose of feature selection is to extract the most representative feature subset from the original feature vector data. This subset should satisfy the following two conditions as much as possible:
The correlation between the feature subset vectors after feature selection should be as small as possible.
The correlation between the feature subset after feature selection and the residual data matrix should be as large as possible.
Assuming that the selected data of an individual feature are X 1 = x 1 , x 2 , , x m , the data in the residual space of the individual is then X 2 = x 1 , x 2 , , x k , k + m = n . If there are two groups of data, A = a 1 , a 2 , , a N , B = b 1 , b 2 , , b N , the Pearson correlation coefficient shows that the correlation coefficients of these two groups of data are:
ρ A B = C o v A , B σ A σ B = i = 1 n A i E A σ A B i E B σ B n σ A = i = 1 n ( A i E ( A ) ) n , σ B = i = 1 n ( B i E ( B ) ) n
where σ A , σ B is the standard deviation of A , B , respectively. At the same time, it is obvious that ρ A B 1 . The fitness function is as follows:
f F i t = p i = 1 m j = 1 m ρ X 1 ( i ) X 1 ( j ) i m 1 i + ( 1 p ) i = 1 m j = 1 k ρ X 1 ( i ) X 2 ( j ) i = 1 , 2 , , m j = 1 , 2 , , k
where f F i t is the fitness function, and p is the fitness proportion of the above two indicators, 0 p 1 .

3.2.3. Setting up Genetic Strategies

The genetic operators of the genetic algorithm proposed in this paper include the elite selection operator, gene crossover operator, gene mutation operator and mutation operator. The elite retention strategy [27,28] aims to prevent the loss of the best individuals in the current population in the next generation, thereby preventing the genetic algorithm from converging to the global optimal solution. The specific process is to directly copy the best individual currently existing in the population to the next generation and replace the worst individual in the next generation when the individual is selected. The elite retention strategy improves the global convergence of the standard genetic algorithm, and it is theoretically proved that the standard genetic algorithm with elite retention has global convergence. Catastrophic strategies [29,30] aim to further improve global search performance. Its essence is to simulate the biological evolution process. When the external environment undergoes major changes, such as volcanoes and tsunamis, biological populations will face mass death or even extinction.
(1)
Elite selection strategy
The most common selection operator in genetic algorithms is the roulette method, in which the probability of selecting an individual is directly proportional to its fitness value. Essentially, the strategy of elite retention is essentially the same as that of keeping individuals as well-adapted as possible. The idea is to have the most adapted individuals directly retain and replicate their offspring while eliminating the less adapted ones, which converges faster than roulette. In the genetic algorithm proposed in this paper, the specific strategies for elite preservation are to retain the first 25% of the fitness of the individuals and duplicate them to enter the next generation directly, retain 25%–50% of the individuals to enter the next generation, and directly eliminate the last 50%.
(2)
Gene crossover strategy
Crossover is the main process of generating new individuals in genetic algorithms. It exchanges some chromosomes between two individuals with a certain probability. This paper adopts the method of single-point chromosome segment crossover. The specific procedure is as follows: first, the populations are paired randomly; second, the position of the intersection is randomly set. Finally, some genes are exchanged between pairs of chromosomes. As shown in Figure 5, at the n-1 position of the chromosome, individual A crosses with individual B, where A changes from 0 to 1 and B changes from 1 to 0.
(3)
Gene mutation strategy
The mutation operation is to change the value of a gene with a small probability at one or a certain locus of an individual. It reflects the diversity of genes. In this paper, we use the single-point mutation method to perform mutation operations. The specific operation process is to first determine the mutation position of each individual. Then the original gene value of the mutation point is inverted according to a certain probability. As shown in Figure 6, the gene of individual A at the n-1 position mutates from 0 to 1.
(4)
Catastrophe Strategy
The genetic algorithm has strong local search ability, but it is easy to fall into local optimal solution. In order to jump out of the local optimum, all excellent individuals must be killed, so that there is enough evolution space for points far from the current extreme value. This is the idea of cataclysm. Holocaust is killing the best individuals, thereby producing more good species. A countdown to when the disaster will happen is required. In the genetic algorithm of this paper, if there is not a better individual than before in n consecutive generations, a disaster occurs, and the disaster makes the count n equal to 0.

3.2.4. Update Iteration

After one selection and cross-mutation, if the current disaster count does not reach the threshold, it will directly enter the next generation, otherwise, the disaster operation will be performed, and the disaster count will return to 0 and enter the next generation. Finally, if the current algebra reaches the number of iterations, the iteration is stopped.

3.3. Offline Modeling

The steps of offline modeling and monitoring are as follows:
(1)
Collect the data of normal operation of the chemical process and conduct sliding window denoising on the data to obtain X t r a i n . Then, due to the inconsistency of different variable units of process data, the data needs to be standardized. The standardization process is as follows:
X ˜ t r a i n = X t r a i n X ¯ t r a i n 1 n i = 1 n X t r a i n X ¯ t r a i n 2
where X ¯ t r a i n represents the mean matrix established by the mean of each variable in the data and X ˜ t r a i n represents the normalized process data matrix.
(2)
Overlay the standardized data with observations from previous L moments to construct a new data matrix, X T r a i n = X ˜ t r a i n , 1 ( 0 ) , X ˜ t r a i n , 1 ( 1 ) , , X ˜ t r a i n , 1 ( L ) , X ˜ t r a i n , 2 , for DGLPP. L is the delay parameter, which is generally 1 or 2 depending on the actual situation. X ˜ t r a i n , 1 ( L ) represents data delayed to L steps and selected for features. X ˜ t r a i n , 2 represents residual data for feature selection.
(3)
Establish a monitoring model based on DPCA according to Formulas (1)–(4).
(4)
Calculate T2 and Q statistics according to Formulas (5) and (6).

3.4. Determining Control Limits

The control limits are established to determine whether a failure has occurred at any point in time, so the control limits for the T2 and Q statistics need to be calculated. When the statistic value calculated by the online collected data is greater than the control limit, it is regarded as a failure. The control limit formulas for determining the statistics are shown in Formulas (7) and (8).

3.5. Online Monitoring

The steps of online monitoring are as follows:
(1)
Collect abnormal data in the process of chemical industry and obtain X t e s t by sliding window denoising, then standardize the data by using Formula (12). Finally, the new data sample required for DGLPP is constructed as X T e s t = X ˜ t e s t , 1 ( 0 ) , X ˜ t e s t , 1 ( 1 ) , , X ˜ t e s t , 1 ( L ) , X ˜ t e s t , 2 .
(2)
Using the projection matrix A obtained from the offline process and the newly acquired data, an online monitoring model based on DPCA is established:
X T e s t = A y T e s t + e T e s t y T e s t = ( A T A ) 1 A T X T e s t e T e s t = X T e s t A y T e s t
where y T e s t is the vector after projection, e T e s t is the residual vector, and A is the projection matrix.
(3)
Calculate T2 and Q statistics according to Formulas (5) and (6).
(4)
Compare with the control limit established in offline modeling to determine whether there is a fault.

3.6. Fault Diagnosis

After statistics are found to be abnormal, fault diagnosis is required. In this paper, the improved contribution graph is used for fault diagnosis, and the formula is as follows:
c o n t j = i = 1 d t j , i T t j , i λ η ( j , i )
q j = e j 2
where d is the number of principal elements, j is the number of variables, and e j represents the residual value of element j .
Considering the dynamic nature of the data, traditional contribution graphs cannot be used directly. The contribution diagram of the improvement used in this paper is as follows:
C O N T j = c o n t j + c o n t j + m + c o n t j + 2 × m + + c o n t j + L × m j m Q j = q j + q j + m + q j + 2 × m + + q j + L × m j m
Here, m is the number of variables in the original data before matrix expansion, and the variable with the largest result is the cause of the failure.

4. Simulation

In order to prove the effectiveness of the algorithm proposed in this paper, we used the Eastman process simulation data in Tennessee to verify it, and compared the PCA and DPCA algorithms to show the superiority of the algorithm. We also set the latency parameter L to 1 with a confidence level of 99%.
TE dataset is produced by Tennessee Eastman, an open and challenging simulation platform for chemical models developed by Eastman Chemical Company in the United States. It has time-varying, strong coupling, and nonlinear properties, and has been widely used to compare various monitoring methods [31]. The four reactants, A, C, D, and E, and inert B are fed into the reactor where products G and H are formed, and by-product F is formed [32]. The control structure is shown in Figure 7. The process has 22 continuous process measurements, 19 component measurements, and 11 control variables for a total of 52 variables. Table 1 lists the 52 variables and Table 2 describes the failure of the 21 sets of test data. The training and testing datasets for each failure are represented by 480 and 960 observations, respectively.
The proposed GA-DPCA is compared with PCA, DPCA. The cumulative variance contribution method is used to determine the number of principal components. The control limits for all monitoring statistics for the three methods were set to 99%. The monitoring results of the three algorithms PCA, DPCA and GA-DPCA are shown in Table 3. It can be seen from Table 3 that the average diagnostic rate of the T2 statistic of GA-DPCA is higher than that of the other two algorithms, and the Q statistic of GA-DPCA is slightly lower than that of DPCA, but higher than that of PCA. In the cases of faults 4, 11, 17, 18 and 21, the diagnostic rate of the T2 statistic of GA-DPCA was significantly higher than that of the other two methods. In the cases of faults 5, 10, 11 and 20, the diagnostic rate of the Q statistic of GA-DPCA is significantly higher than that of the other two methods. In other cases, the diagnostic rate of GA-DPCA is about the same as the other two algorithms. These situations reflect that the GA-DPCA algorithm has a significant improvement in the diagnosis rate of these small faults that are not easily detected by PCA or DPCA. In order to further illustrate the superiority of the GA-DPCA monitoring method, the comparison of the detection results of PCA, DPCA and GA-DPCA algorithms in faults 4 and 11 is shown in Figure 8 and Figure 9, respectively.
It can be seen from Figure 8 that in the case of fault 4, for PCA under T2 statistic monitoring, the effect is good before the fault is introduced, and after the fault is introduced, the T2 control limit cannot effectively identify the fault. For PCA under the monitoring of the Q statistic, the diagnosis effect before the introduction of the fault is average. After the introduction of the fault, the Q control limit can effectively identify the fault, but it can be seen from the figure that there is a large fluctuation. For DPCA, the monitoring effect of the T2 statistic is very poor, and most of the statistical values are below the control limit, almost losing the monitoring effect. For DPCA under the monitoring of the Q statistic, there are errors in the diagnosis effect before and after the fault is introduced. For GA-DPCA under the monitoring of the T2 statistic, the diagnosis effect is good before and after the introduction of the fault, and the accuracy rate is close to 100%. For GA-DPCA under the monitoring of the Q statistic, there are some errors in the diagnosis before the fault is introduced, but the diagnosis rate after the fault is introduced is 100%.
It can be seen from Figure 9 that in the case of fault 11, for PCA under T2 statistic monitoring, the effect is good before the fault is introduced, and after the fault is introduced, the T2 statistic monitoring effect has some errors. For PCA under monitoring of the Q statistic, the diagnosis effect before and after the fault is introduced is average, and the statistics fluctuate greatly. For DPCA under the monitoring of T2 statistic, the diagnosis effect before the introduction of the fault is good, and after the introduction of the fault, most of the fault data cannot be identification since the diagnosis rate is low. For DPCA under monitoring of the Q statistic, before the fault is introduced, the misjudgment rate of the diagnostic effect is relatively large, and after the fault is introduced, the accuracy rate is close to 100%. For GA-DPCA under monitoring of the T2 statistic, the diagnostic effect before and after the introduction of the fault is good, and the accuracy is close to 95%. For GA-DPCA under monitoring of the Q statistic, there are some errors in the diagnosis before the fault is introduced, but the diagnosis rate after the fault is introduced is 100%.
After the fault detection is completed, fault diagnosis was carried out by the contribution diagram. Failure 4 and Failure 11 were selected for testing. The contribution of the Q-statistic based on GA-DPCA to the sample is shown in Figure 10 and Figure 11. In the contribution diagram, the red represents the variable causing the fault, while the other variables are represented in blue. Therefore, variable 51 (reactor cooling water flow) was determined to be the cause of failures 4 and 11. Table 2 was checked to make sure they are correct. Table 4 shows how much GA-DPCA improves the average diagnostic accuracy of the T2 and Q statistics. Through the TE simulation process simulation results shown in Table 4, the average diagnostic accuracy of GA-DPCA-based fault diagnosis on T2 statistics is 70.01%, and the average diagnostic accuracy on Q statistics is 82.87%. Compared with PCA, it increased by 4.52% in T2 and by 6.09% in Q statistics. Compared with DPCA, it increased by 6.84% in T2 statistics and decreased by 0.78% in Q statistics. Therefore, compared with PCA and DPCA, GA-DPCA has a better monitoring and diagnosis effect in practical applications. In addition, the improved contribution map can accurately diagnose which variables caused the failure, which has practical application value for actual chemical process monitoring.

5. Conclusions

The method proposed in this paper is essentially based on the PCA method. Due to the wide application of the PCA method, many scholars at home and abroad have studied it to improve the monitoring effect. This paper considers the data feature selection aspect and puts forward a DPCA fault monitoring and diagnosis method based on feature selection by using the feature of fast optimal solution of the genetic algorithm. Firstly, sliding window denoising is introduced for removing noise and increasing modeling accuracy. Secondly, feature selection is performed based on a genetic algorithm to find the optimal feature subset and eliminate irrelevant or redundant features. Thirdly, offline modeling and determination of control limits are introduced for process monitoring. This method not only preserves the dimensionality reduction of principal component analysis, but also considers the dynamics and integrity of the data, which can effectively improve the monitoring effect compared with typical methods shown in [33,34,35].
Although the method presented in this paper provides improved results compared with traditional PCA methods, there is still much room for improvement, for example, diagnostic accuracy under T2 statistics is still less than 80%. At the same time, due to the non-linear nature of the chemical process data, the effect of traditional PCA is greatly reduced, so the effect of some unknown fault diagnosis is not ideal. In future research, the authors will consider existing algorithms to propose more efficient monitoring methods for chemical processes in view of the non-linear characteristics of high-dimensional data.

Author Contributions

Conceptualization, C.L. and F.W.; methodology, C.L.; software, C.L.; validation, C.L., J.B. and F.W.; formal analysis, F.W.; writing—original draft preparation, C.L.; writing—review and editing, J.B.; visualization, C.L.; supervision, F.W.; project administration, F.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Zhejiang Provincial Natural Science Foundation of China under Grant (LZ22F030001).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, J.; Zhou, D.-H.; Si, X.; Chen, M.-Y.; Xu, C.-H. Review of incipient fault diagnosis methods. Kongzhi Lilun Yu Yingyong/Control Theory Appl. 2012, 29, 1517–1529. [Google Scholar]
  2. Zhang, R.; Peng, Z.; Wu, L.; Yao, B.; Guan, Y. Fault Diagnosis from Raw Sensor Data Using Deep Neural Networks Considering Temporal Coherence. Sensors 2017, 17, 549. [Google Scholar] [CrossRef] [PubMed]
  3. Yang, F.; Cui, Y.; Wu, F.; Zhang, R. Fault Monitoring of Chemical Process Based on Sliding Window Wavelet DenoisingGLPP. Processes 2021, 9, 86. [Google Scholar] [CrossRef]
  4. Liu, K.; Lu, N.; Wu, F.; Zhang, R.; Gao, F. Model Fusion and Multiscale Feature Learning for Fault Diagnosis of Industrial Processes. IEEE Trans. Cybern. 2022, 1–14. [Google Scholar] [CrossRef] [PubMed]
  5. Wang, N.; Yang, F.; Zhang, R.; Gao, F. Intelligent Fault Diagnosis for Chemical Processes Using Deep Learning Multimodel Fusion. IEEE Trans. Cybern. 2022, 52, 7121–7135. [Google Scholar] [CrossRef]
  6. Lei, Y.; Jiang, W.; Jiang, A.; Zhu, Y.; Niu, H.; Zhang, S. Fault Diagnosis Method for Hydraulic Directional Valves Integrating PCA and XGBoost. Processes 2019, 7, 589. [Google Scholar] [CrossRef] [Green Version]
  7. Kim, J.; Lee, T.; Lee, S.; Lee, J.; Lee, W.; Kim, Y.; Park, J. A Study on Deep Learning-Based Fault Diagnosis and Classification for Marine Engine System Auxiliary Equipment. Processes 2022, 10, 1345. [Google Scholar] [CrossRef]
  8. Xu, X.; Feng, J.; Wang, H.; Zhang, N.; Wang, X. Dynamics Analysis of Misalignment and Stator Short-Circuit Coupling Fault in Electric Vehicle Range Extender. Processes 2020, 8, 1037. [Google Scholar] [CrossRef]
  9. Sun, J.; Xiao, Z.; Xie, Y. Automatic multi-fault recognition in TFDS based on convolutional neural network. Neurocomputing 2017, 222, 127–136. [Google Scholar] [CrossRef]
  10. Pilario, K.E.; Shafiee, M.; Cao, Y.; Lao, L.; Yang, S.-H. A Review of Kernel Methods for Feature Extraction in Nonlinear Process Monitoring. Processes 2020, 8, 24. [Google Scholar] [CrossRef] [Green Version]
  11. Chang, Y.W.; Wang, Y.C.; Tao, L.; Wang, Z.J. Fault diagnosis of a mine hoist using PCA and SVM techniques. J. China Univ. Min. Technol. 2008, 18, 327–331. [Google Scholar] [CrossRef]
  12. Li, C.-H.; Kuo, B.-C.; Lin, C.-T. LDA-Based Clustering Algorithm and Its Application to an Unsupervised Feature Extraction. IEEE Trans. Fuzzy Syst. 2011, 19, 152–163. [Google Scholar] [CrossRef]
  13. Liu, C.; Wechsler, H. Independent component analysis of Gabor features for face recognition. IEEE Trans. Neural. Netw. 2003, 14, 919–928. [Google Scholar]
  14. Nomikos, P.; MacGregor, J.F. Monitoring batch processes using multiway principal component analysis. AIChE J. 1994, 40, 1361–1375. [Google Scholar] [CrossRef]
  15. Nomikos, P.; MacGregor, J.F. Multi-way partial least squares in monitoring batch processes. Chemom. Intell. Lab. Syst. 1995, 30, 97–108. [Google Scholar] [CrossRef]
  16. Ku, W.; Storer, R.H.; Georgakis, C. Disturbance detection and isolation by dynamic principal component analysis. Chemom. Intell. Lab. Syst. 1995, 30, 179–196. [Google Scholar] [CrossRef]
  17. Choi, S.W.; Lee, C.; Lee, J.-M.; Park, J.H.; Lee, I.-B. Fault detection and identification of nonlinear processes based on kernel PCA. Chemom. Intell. Lab. Syst. 2005, 75, 55–67. [Google Scholar] [CrossRef]
  18. Cao, L.J.; Chua, K.S.; Chong, W.K.; Lee, H.P.; Gu, Q.M. A comparison of PCA, KPCA and ICA for dimensionality reduction in support vector machine. Neurocomputing 2003, 55, 321–336. [Google Scholar] [CrossRef]
  19. Joe Qin, S.; Dunia, R. Determining the Number of Principal Components for Best Reconstruction. IFAC Proc. Vol. 1998, 31, 357–362. [Google Scholar] [CrossRef]
  20. MacGregor, J. Multivariate Statistical Approaches to Fault Detection and Isolation. IFAC Proc. Vol. 2003, 36, 549–554. [Google Scholar] [CrossRef]
  21. Lee, J.-M.; Yoo, C.; Lee, I.-B. Statistical process monitoring with independent component analysis. J. Process Control. 2004, 14, 467–485. [Google Scholar] [CrossRef]
  22. Chen, Y.; Zi, Y.; Cao, H.; He, Z.; Sun, H. A data-driven threshold for wavelet sliding window denoising in mechanical fault detection. Sci. China Technol. Sci. 2014, 57, 589–597. [Google Scholar] [CrossRef]
  23. Yongguo, L. Feature Subset Selection Based on Genetic Algorithm. Comput. Eng. 2003, 29, 19–20. [Google Scholar]
  24. ElAlami, M.E. A filter model for feature subset selection based on genetic algorithm. Knowl. Based Syst. 2009, 22, 356–362. [Google Scholar] [CrossRef]
  25. Iba, K. Reactive power optimization by genetic algorithm. IEEE Trans. Power Syst. 1994, 9, 685–692. [Google Scholar] [CrossRef]
  26. Lingaraj, H. A Study on Genetic Algorithm and its Applications. Int. J. Comput. Sci. Eng. 2016, 4, 139–143. [Google Scholar]
  27. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef] [Green Version]
  28. Xiaohang, C.; Huifeng, X. Design of Elitist Adaptive Genetic Algorithm in Arrival Aircrafts Scheduling. Comput. Commun. 2006, 24, 91–94. [Google Scholar]
  29. Liao, M. Study on the Effect of Cataclysm Operator on Genetic Algorithm. Comput. Eng. Appl. 2005, 41, 54–57. [Google Scholar]
  30. Peng, Z. A Partheno-genetic Algorithm Based on Cataclysm. J. Hubei Automot. Ind. Inst. 2007, 2, 19–21. [Google Scholar]
  31. Crammer, K.; Singer, Y. On the Learnability and Design of Output Codes for Multiclass Problems. Mach. Learn. 2002, 47, 201–233. [Google Scholar] [CrossRef] [Green Version]
  32. Lyman, P.R.; Georgakis, C. Plant-wide control of the Tennessee Eastman problem. Comput. Chem. Eng. 1995, 19, 321–331. [Google Scholar] [CrossRef]
  33. Wang, H.; Song, Z.; Hui, W. Statistical process monitoring using improved PCA with optimized sensor locations. J. Process Control. 2002, 12, 735–744. [Google Scholar] [CrossRef]
  34. Wang, H.; Song, Z.; Ping, L.I. Improved PCA with application to process monitoring and fault diagnosis. J. Chem. Ind. Eng. 2001, 52, 471–475. [Google Scholar]
  35. Wang, K.; Chen, J.; Song, Z. Performance Analysis of Dynamic PCA for Closed-Loop Process Monitoring and Its Improvement by Output Oversampling Scheme. IEEE Trans. Control. Syst. Technol. 2019, 27, 378–385. [Google Scholar] [CrossRef]
Figure 1. Framework diagram of the whole control system.
Figure 1. Framework diagram of the whole control system.
Processes 10 02570 g001
Figure 2. Sliding window denoising diagram.
Figure 2. Sliding window denoising diagram.
Processes 10 02570 g002
Figure 3. Comparison diagram before and after filtering.
Figure 3. Comparison diagram before and after filtering.
Processes 10 02570 g003
Figure 4. Genetic algorithm flow chart.
Figure 4. Genetic algorithm flow chart.
Processes 10 02570 g004
Figure 5. Gene crossover diagram.
Figure 5. Gene crossover diagram.
Processes 10 02570 g005
Figure 6. Gene mutation diagram.
Figure 6. Gene mutation diagram.
Processes 10 02570 g006
Figure 7. TE process.
Figure 7. TE process.
Processes 10 02570 g007
Figure 8. (a) Monitoring results of PCA in fault 4. (b) Monitoring results of DPCA in fault 4. (c) Monitoring results of GA-DPCA in fault 4.
Figure 8. (a) Monitoring results of PCA in fault 4. (b) Monitoring results of DPCA in fault 4. (c) Monitoring results of GA-DPCA in fault 4.
Processes 10 02570 g008aProcesses 10 02570 g008b
Figure 9. (a) Monitoring results of PCA in fault 11. (b) Monitoring results of DPCA in fault 11. (c) Monitoring results of GA-DPCA in fault 11.
Figure 9. (a) Monitoring results of PCA in fault 11. (b) Monitoring results of DPCA in fault 11. (c) Monitoring results of GA-DPCA in fault 11.
Processes 10 02570 g009aProcesses 10 02570 g009b
Figure 10. Contribution plots of fault 4.
Figure 10. Contribution plots of fault 4.
Processes 10 02570 g010
Figure 11. Contribution plots of fault 11.
Figure 11. Contribution plots of fault 11.
Processes 10 02570 g011
Table 1. Process monitoring variables in the Tennessee Eastman process.
Table 1. Process monitoring variables in the Tennessee Eastman process.
No.Variables No.Variables
1A feed (stream 1)27Ingredient E (stream 6)
2D feed (stream 2)28Ingredient F (stream 6)
3E feed (stream 3)29Ingredient A (stream 9)
4Total feed (stream 4)30Ingredient B (stream 9)
5Recycle flow (stream 8)31Ingredient C (stream 9)
6Reactor feed rate (stream 6)32Ingredient D (stream 9)
7Reactor pressure33Ingredient E (stream 9)
8Reactor level34Ingredient F (stream 9)
9Reactor temperature35Ingredient G (stream 9)
10Purge rate (stream 9) 36Ingredient H (stream 9)
11Product separator temperature37Ingredient D (stream 11)
12Product separator level38Ingredient E (stream 11)
13Product separator pressure39Ingredient F (stream 11)
14Product separator underflow (stream 10)40Ingredient G (stream 11)
15Stripper level41Ingredient H (stream 11)
16Stripper pressure42D feed flow valve (stream 2)
17Stripper underflow (stream 11)43E feed flow valve (stream 3)
18Stripper temperature44A feed flow valve (stream 1)
19Stripper steam Flow45Total feed flow valve (stream4)
20Compressor work46Compressor recycle valve
21Reactor cooling water outlet temperature47Purge valve (stream 9)
22Separator cooling water outlet temperature48Separator pot liquid flow valve (stream 10)
23Ingredient A (stream 6)49Stripper liquid product flow valve (stream 11)
24Ingredient B (stream 6)50Stripper steam valve
25Ingredient C (stream 6)51Reactor cooling water flow
26Ingredient D (stream 6)52Condenser cooling water flow
Table 2. Process faults for the TE process.
Table 2. Process faults for the TE process.
Fault NumberProcess VariableType
1A/C feed ratio, B composition constant (stream 4)Step
2B composition, A/C ratio constant (stream 4) Step
3D feed temperature (stream 2)Step
4Reactor cooling water inlet temperatureStep
5Condenser cooling water inlet temperature Step
6A feed loss (stream 1)Step
7C header pressure loss-reduced availability (stream 4)Step
8A, B, C feed composition (stream 4) Random variation
9D feed temperature (stream 2) Random variation
10C feed temperature (stream 4) Random variation
11Reactor cooling water inlet temperature Random variation
12Condenser cooling water inlet temperature Random variation
13Reaction kinetics Slow drift
14Reactor cooling water valve Sticking
15Condenser cooling water valveSticking
16UnknownUnknown
17UnknownUnknown
18UnknownUnknown
19UnknownUnknown
20UnknownUnknown
21Valve position constant (stream 4) Constant position
Table 3. Fault diagnosis rate of four methods for the TE Process.
Table 3. Fault diagnosis rate of four methods for the TE Process.
Fault NumberPCADPCAGA-DPCA
T2QT2QT2Q
199.58%98.02%99.58%91.14%99.27%92.91%
298.44%96.88%98.75%91.46%98.85%94.79%
318.54%28.75%17.19%55.92%21.79%46.56%
450.73%97.19%23.85%91.77%93.53%93.12%
538.44%49.80%37.29%77.29%51.30%91.97%
699.17%98.54%99.38%93.57%99.79%95.20%
7100.00%98.13%100.00%94.20%99.79%94.16%
897.81%96.88%97.71%92.10%97.39%93.01%
918.85%28.33%16.87%54.89%21.48%50.42%
1051.36%74.06%52.71%83.57%64.03%84.37%
1160.00%76.05%42.39%91.37%83.94%92.60%
1298.85%95.84%99.27%93.46%99.06%94.06%
1395.94%95.21%95.83%92.32%96.14%92.70%
1499.80%96.67%99.90%90.82%34.20%90.83%
1520.73%32.08%19.48%52.71%30.03%46.67%
1634.90%68.54%32.29%82.51%45.15%81.14%
1783.23%94.90%81.67%90.74%91.76%90.20%
1890.94%91.98%91.04%87.49%92.18%88.85%
1922.29%56.35%21.25%89.60%26.28%67.40%
2047.81%73.02%48.75%84.60%56.73%87.29%
2147.92%65.21%51.46%75.20%67.47%71.97%
Average65.49%76.78%63.17%83.65%70.01%82.87%
Table 4. GA-DPCA average improvement rate of accurate diagnosis.
Table 4. GA-DPCA average improvement rate of accurate diagnosis.
MethodT2Q
PCA4.52%6.09%
DPCA6.84%−0.78%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, C.; Bai, J.; Wu, F. Fault Diagnosis Using Dynamic Principal Component Analysis and GA Feature Selection Modeling for Industrial Processes. Processes 2022, 10, 2570. https://doi.org/10.3390/pr10122570

AMA Style

Liu C, Bai J, Wu F. Fault Diagnosis Using Dynamic Principal Component Analysis and GA Feature Selection Modeling for Industrial Processes. Processes. 2022; 10(12):2570. https://doi.org/10.3390/pr10122570

Chicago/Turabian Style

Liu, Chenpeng, Jianjun Bai, and Feng Wu. 2022. "Fault Diagnosis Using Dynamic Principal Component Analysis and GA Feature Selection Modeling for Industrial Processes" Processes 10, no. 12: 2570. https://doi.org/10.3390/pr10122570

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop