Next Article in Journal
Quantification of Carbon Emission and Solid Waste from Pottery Production by Using Life-Cycle Assessment (LCA) Method in Yunnan, China
Previous Article in Journal
Helical Foldamers and Stapled Peptides as New Modalities in Drug Discovery: Modulators of Protein-Protein Interactions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Nonlinear Dynamic Process Monitoring Based on Two-Step Dynamic Local Kernel Principal Component Analysis

1
School of Information and Control Engineering, Liaoning Petrochemical University, Fushun 113005, China
2
Institute of Intelligence Science and Engineering, Shenzhen Polytechnic, Shenzhen 518055, China
3
Faculty of Engineering, Technology & Built Environment, UCSI University, Kuala Lumpur 56000, Malaysia
*
Author to whom correspondence should be addressed.
Processes 2022, 10(5), 925; https://doi.org/10.3390/pr10050925
Submission received: 17 March 2022 / Revised: 27 April 2022 / Accepted: 27 April 2022 / Published: 7 May 2022

Abstract

:
Nonlinearity may cause a model deviation problem, and hence, it is a challenging problem for process monitoring. To handle this issue, local kernel principal component analysis was proposed, and it achieved a satisfactory performance in static process monitoring. For a dynamic process, the expectation value of each variable changes over time, and hence, it cannot be replaced with a constant value. As such, the local data structure in the local kernel principal component analysis is wrong, which causes the model deviation problem. In this paper, we propose a new two-step dynamic local kernel principal component analysis, which extracts the static components in the process data and then analyzes them by local kernel principal component analysis. As such, the two-step dynamic local kernel principal component analysis can handle the nonlinearity and the dynamic features simultaneously.

1. Introduction

With the development of modern industrialization, chemical processes tend to be large-scale and complex. As such, fault detection technology [1,2] becomes increasingly essential because it has the potential to decrease the economic loss caused by process faults.
The data-driven process monitoring method [3,4], which analyzes the process data without knowing the precise analytical model of the system, is an effective method for ensuring industrial safety due to its simplicity and adaptability. Typical data-driven approaches include principal component analysis (PCA) [5,6,7,8,9,10], partial least squares (PLS) [11,12], independent component analysis (ICA) [13,14], etc. Among the above data-driven approaches, PCA is the most commonly used method [15]. The basic principle of PCA is to extract the principal components (PCs) by projecting high-dimensional data into low-dimensional PC space with a linear transformation, and then, it maintains the majority variance of the original data [16].
However, PCA is primarily a linear approach, and hence, it cannot handle nonlinearity in the process data. To handle this issue, Lee et al. proposed kernel PCA (KPCA) [17,18,19,20,21], the main idea of which is to map the input space into a high-dimensional feature space via a nonlinear mapping kernel function. According to survey paper [22], KPCA has become the mainstream nonlinear method in recent decades, and many improved versions have been proposed in that time. Jiang et al. incorporated PCA and KPCA to handle the linear and nonlinear relationships in process data [18]. Lahdhiri et al. proposed a reduced rank scheme to reduce the computational complexity of KPCA [23]. Peng et al. combined KPCA with the GMM to handle the nonlinearity and multimode feature simultaneously [24].
However, for KPCA, all the training data are used for calculating the kernel function values without considering the inner relationships among the different data points. As such, KPCA is a global structure analysis technique, and it ignores the detailed local structure information, which is important for data dimensionality reduction and feature extraction. To address this issue, Deng et al. proposed a novel KPCA based on local structure analysis, referred to as local KPCA (LKPCA) [25]. LKPCA introduces the local data structure analysis into the global optimization function of KPCA and solves it by using generalized eigenvalue decomposition [26]. The local data structure in LKPCA is obtained by k-nearest neighbor (KNN) [27,28], which is based on the Euclidean distances between variables.
LKPCA is a static algorithm, which assumes that the expectation value of each variable is fixed and that it will not change over time. However, in a dynamic process, the expectation value of each variable changes over time, and hence, the Euclidean distances are not suitable for assessing the similarity between variables in a different sample time. As such, the local data structure obtained by KNN is wrong in a dynamic process, which causes a model deviation problem.
To handle the nonlinearity and the dynamic feature simultaneously, this paper introduces the two-step dynamic scheme [29] into LKPCA and names it as two-step dynamic LKPCA (TSD-LKPCA). The two-step dynamic scheme is adopted to calculate the dynamical structure of the process data, and then, the static components are extracted from the process data. As the static components are time-uncorrelated, they can be monitored by LKPCA, which addresses the model deviation problem of LKPCA.
The main contributions of this paper are as follows: (a) the drawback of the traditional LKPCA is analyzed; (b) a new two-step dynamic scheme is proposed for LKPCA, which handles the dynamic feature in the process and inherits LKPCA’s nonlinear processing ability. Tests results in a numerical model show that TSD-LKPCA achieves a 100% fault detection rates in three different types of faults, and it successfully addresses the false alarm problem caused by the dynamic feature. In addition, the test results in the Tennessee Eastman (TE) Process [30,31] show that TSD-LKPCA achieves the best fault detection rate in 14 of the 21 faults, and it achieves a 100% fault detection rate in fault 5. In both tests, TSD-LKPCA achieves a much better performance than LKPCA, PCA, and KPCA.
The remainder of this paper is organized as follows. KPCA and LKPCA are briefly discussed in Section 2. Then, TSD-LKPCA is presented in Section 3. The superiority of the proposed method is demonstrated in Section 4 by several tests. Finally, the conclusion, limitations, and future research are presented in Section 5.

2. Methods

2.1. Notations and Symbols

Table 1 and Table 2 list the symbol and acronyms used in this manuscript:

2.2. Kernel Principal Component Analysis (KPCA)

Assuming that the training dataset X = [ x ( 1 ) x ( 2 ) x ( n ) ] T R n × m represents the n samples of m variables, where ϕ ( · ) denotes the nonlinear translation function mapping from the original nonlinear data space to the new linear feature space. To avoid computing with the function ϕ ( · ) , KPCA defines a N × N kernel matrix K = { K i j } as
K i j = ϕ ( x ( i ) ) , ϕ ( x ( j ) ) = Γ ( x ( i ) , x ( j ) ) .
In Equation (1), Γ ( ) is the kernel function [32], i.e., the radial basis kernel as
Γ ( x ( i ) , x ( j ) ) = exp ( x ( i ) x ( j ) 2 σ )
where σ is a weight parameter.
The matrix K can be mean normalized as
K ¯ = K K K + K ,
where
= 1 n [ 1 1 1 1 ] R n × n .
Finally, applying eigenvalue decomposition to K ¯ as
K ¯ p i = α i p i ,
where p = ( p 1 , p 2 , , p k ) R n × k is the projection vector, α i is the variance of the i th PC, and k is the number of PCs. For a new data sample y R 1 × m , one gets
t = Γ ( y , X ) p ,
and t can be monitored by the T 2 and S P E indices [33].

2.3. Local Kernel Principal Component Analysis (LKPCA)

The high-dimensional data ϕ ( x ) are often embedded in the ambient space on a low-dimensional manifold, and the goal of a local structure analysis is to determine the best linear approximation that keeps the nearby points as close together as feasible [34]. In other words, if ϕ ( x ( i ) ) and ϕ ( x ( j ) ) are nearby points, then ϕ ( x ( i ) ) T p and ϕ ( x ( j ) ) T p in the feature space should be nearby points as well. As such, the following neighborhood matrix is proposed for describing whether two points are nearby:
S i j = { 1 ,   if   ϕ ( x ( i ) )   and   ϕ ( x ( j ) )   are   nearby 0 ,   else .
To determine whether ϕ ( x ( i ) ) and ϕ ( x ( j ) ) are nearby, the k-nearest neighbor (KNN) approach is utilized: if the Euclidean distances between variables x ( i ) and x ( j ) are small then the two samples are considered to be nearby.
As such, the local optimization goal is as follows:
min α J L = min α T K L K α .
In Equation (8), L is the Laplacian matrix, which can be calculated as L = U S , where S = { S i j } is the neighbor matrix and U is the diagonal matrix, with its i t h diagonal element as U i i = j = 1 n S i j .
Hence, the global–local optimization goal is as follows:
max α J L G = α T K K α α T K L K α .
The above equation can be solved by applying singular value decomposition to
K K α = λ ( K L K + ε I ) α .
where ε is a small regularization parameter to avoid matrix singularity problems, and I is the identity matrix.

2.4. Two-Step Dynamic Scheme

The two-step dynamic scheme was first proposed in paper [35]. Assume
x ( t ) = x ( t 1 ) A 1 + + x ( t q ) A q + x ˜ ( t ) = X ¯ ( t ) A ˜ + x ˜ ( t ) ,
where X ¯ ( t ) = [ x ( t 1 ) x ( t q ) ] and A ˜ = [ A 1 T A q T ] T . Parameter q is the time lag and x ˜ is the static component.
The first step calculates the difference between the two data samples x ( t ) and x ( t D ) as
Δ x ( t ) = x ( t ) x ( t D ) ,
where D is the time difference. Then, calculate the dynamic matrix A ˜ between Δ x ( t ) and [ Δ x ( t 1 ) Δ x ( t q ) ] with the least squares algorithm [36], and extract the static components as
x ˜ ( t ) = x ( t ) X ¯ ( t ) A ˜ .
The second step monitors x ˜ ( t ) with the traditional process monitoring methods.

3. The Proposed Method

3.1. Analysis the Performance of LKPCA in Dynamic Process

In addition to the nonlinearity, the dynamic feature [37,38] is also a common issue for process monitoring. When LKPCA is applied in a dynamic process, it achieves a low detection rate and a large false alarm rate. The reason for this phenomenon is that in LKPCA the nearby points are determined by the Euclidean distance of the variable value x ( i ) , rather than that of the deviation x ( i ) E ( x ( i ) ) .
As shown in Figure 1, for the static process, the expectation E ( x ( i ) ) remains unchanged, i.e., E ( x ( i ) ) = μ , and hence, the Euclidean distance of the two sample values x ( i ) and x ( j ) can used to represent the Euclidean distance of the fluctuation at sample time i and j , i.e.,
x ( i ) x ( j ) = ( x ( i ) μ ) ( x ( j ) μ ) .
As a result, the nearby points in Figure 1 (marked with red circles) have a close fluctuation, and hence, they are nearby points (clustered by KNN).
As shown in Figure 2, for the dynamic process the expectation E ( X ) is not fixed, and it changes over time, i.e., E ( x ( i ) ) E ( x ( j ) ) . In this situation, the Euclidean distance of the two sample values, x ( i ) and x ( j ) , is not equal to the Euclidean distance of the fluctuation at sample time i and j . In Figure 2a, the cluster result of nearby points based on the Euclidean distances of original variable values is wrong because it ignores the changing of E ( X ) , and hence, the data with close sample times are more likely to be clustered into one group. As the process goes on, the expectation values of the variables deviate a lot from those of the training data, and hence, the false alarm rate will be very large. In Figure 2b, the cluster result is right because each cluster consists of a data sample from a different sample time, and it is not influenced by the changing of the expectation value. As such, how to handle the changing expectation value is the key to addressing the dynamic process feature.

3.2. New Two-Step Dynamic Scheme

To handle the dynamic issue, the key step is to extract the static components from the dynamic data. One drawback of the traditional two-step dynamic scheme is that x ( t ) and x ( t D ) are related even if D is very large, and hence, the dynamic matrix A ˜ may be wrong. In this paper, we introduce another independent sample dataset as
y ( t ) = y ( t 1 ) A 1 + + y ( t q ) A q + Y ˜ ( t ) = Y ¯ ( t ) A ˜ + Y ˜ ( t ) .
and calculate the difference between two datasets as
Δ z = x ( t ) y ( t ) = ( X ¯ ( t ) Y ¯ ( t ) ) A ˜ + ( x ˜ ( t ) y ˜ ( t ) ) = Δ Z ¯ ( t ) A ˜ + Δ z ˜ ( t ) ,
where X ¯ ( t ) Y ¯ ( t ) is the component related to Δ z and Δ z ˜ ( t ) = x ˜ ( t ) y ˜ ( t ) is the static component which is independent of Δ z . As x ˜ ( t ) and y ˜ ( t ) follow the same statistical distribution, hence the expectation value of Δ z ˜ ( t ) is zero and A ˜ can be estimated with the least squares algorithm. The following steps are the same as in the traditional two-step dynamic scheme.

3.3. Two-Step Dynamic Local Kernel Principal Component Analysis (TSD-LKPCA)

As shown in Figure 3, one gets that the static components extracted from X are independent of the changing variable expectation value, and hence, x ˜ ( t ) can be used for LKPCA.
The details of TSD-LKPCA are as follows.
Step 1. Process the original data with the new two-step dynamic scheme and extract the static components x ˜ ( t ) with Equation (13).
Step 2. Cluster x ˜ ( t ) into several nearby groups by using KNN and then calculate the neighborhood matrix S i j by using Equation (7).
Step 3. Process x ˜ ( t ) with the kernel function and calculate the projection vector p by using Equation (10).
Step 4. Extract the PCs by using Equation (6) and then monitor them by the T 2 and S P E indices.

4. Simulation Results

In this section, we test the performance of TSD-LKPCA based on a numerically simulated dynamic nonlinear process and a TE process. The monitoring results acquired by this approach are compared to those obtained by PCA, KPCA, and LKPCA in terms of detection rate and false alarm rate.

4.1. Numerical Model Test

In order to verify the superiority of TSD-LKPCA in a dynamic nonlinear process, the following mathematical model is designed to test the effectiveness of the algorithm:
{ x 1 ( t ) = χ + e 1 + x 1 ( t 1 ) x 2 ( t ) = χ 2 3 χ + e 2 + 0.8 x 1 ( t ) x 3 ( t ) = χ 3 + 3 χ 2 + e 3 + 0.8 x 2 ( t ) + 0.6 x 1 ( t ) ,
where χ is a random uniform distribution, e 1 , e 2 , e 3 are Gaussian random noise with zero mean and variance 0.01, and x 1 , x 2 , x 3 are the monitoring variables. This process generates 860 samples of normal data for off-line training and another 960 samples for testing. These test data introduce faults at the 450th sampling points.
There are two types of faults, as follows:
  • Fault 1: a step fault occurs at variable x 1 with an amplitude of −6;
  • Fault 2: the coefficient 0.8 in variable x 2 changes to 0.5.
For the three nonlinear approaches, KPCA, LKPCA, and TSD-LKPCA, the kernel width parameter σ is set to 50,000 by cross-validation. The LKPCA and TSD-LKPCA methods’ neighborhood relation parameter k n is 15. The dynamic process description model’s lag parameter q is set to 2. The control limits for the four methods are based on a confidence limit of 99%. All these parameters are also used in the following Section 4.2.
Table 3 shows the false alarm rates and fault detection rates of four algorithms. Due to the nonlinear characteristics of the process, the average T 2 statistics of the false alarm rates of the three nonlinear methods, i.e., KPCA, LKPCA, and TSD-LKPCA, are much lower than that of the linear method PCA. Because KPCA and LKPCA are static methods, on the one hand they regard the change of expectation value in the training data as the normal data fluctuation, and hence, the control limits are very high in both methods, and they are insensitive to the faults; on the other hand, they regard the change of data expectation values in the testing data as the deviation caused by faults, resulting in large false detection. However, because TSD-LKPCA effectively handles the nonlinear and dynamic characteristics of the process, it achieves a 100% detection rate for both faults. The best result in each item is marked in underline and bold.
The monitoring curves for faults 1 and 2 are depicted in Figure 4 and Figure 5. The blue line indicates the statistics, the red line represents the corresponding control limit, and the orange line represents the failure time in these two graphs. The T 2 statistics of PCA, KPCA, and LKPCA change over time in both figures, while the T 2 statistics of TSD-LKPCA do not change because it monitors the static components, whose expectation value is fixed. Because there is a deviation between the initial process dataset and the mean value of the training data, this deviation is regarded as a fault, and hence, the T 2 and S P E statistics of the PCA, KPCA, and LKPCA algorithms initially exceeded the control limit, resulting in a false alarm. However, the TSD-LKPCA method has handled the dynamic characteristics and hence avoided this issue. For fault 1, as shown in Figure 4, when a fault occurs the PCA, KPCA, and LKPCA algorithms detect the fault after a period of time, which causes detection delay. For fault 2, as shown in Figure 5, the mean or variance of the T 2 statistics for the PCA, KPCA, and LKPCA methods change over time because of the dynamic feature, and hence, they are insensitive to the fault. In contrast, the TSD-LKPCA method detects the fault immediately, demonstrating that it has great sensitivity in fault detection in the process of nonlinear dynamic characteristics.

4.2. Tennessee Eastman (TE) Process Test

The Tennessee Eastman (TE) process is based on simulation models of actual industrial processes and is frequently used as a publicly available data source for testing process monitoring methods. The entire process system has 12 operational variables, 22 continuous variables, and 19 component variables. The training dataset consists of 960 normal data samples collected under normal operating circumstances. The TE process also produces 21 distinct types of faults (as shown in Table 4), which contain 960 samples of where the fault occurs, from the 161st data point to the end.
As the TE process is a nonlinear and dynamic simulation model, it is, hence, adopted to test the performance of PCA, KPCA, LKPCA, and TSD-LKPCA. For method comparison, 315 samples of a normal process dataset were utilized for model training. Table 5 displays the detection results of 21 faults. Obviously, TSD-LKPCA achieves the best fault detection rate in 14 of the 21 faults and it has much lower false alarm rates than PCA and KPCA. In particular, TSD-LKPCA achieves a 100% fault detection rate in fault 5, and those of the other methods are lower than 45%. TSD-LKPCA also achieves much better than other the methods in faults 10, 16, 19, and 20. For fairness, we take the difference between the fault detection rates and the false alarm rates as an index, which indicates that TSD-LKPCA can achieve a much higher fault detection rate with the same false alarm rate. The best result in each item is marked in underline and bold.
Figure 6 provides monitoring diagrams for several algorithms in fault 5. Fault 5 is the fault of a sudden change in the intake temperature of the cooling water, which occurs from the 161st data point to the end. According to the detection result curves of PCA (Figure 6a), KPCA (Figure 6b), and LKPCA (Figure 6c), the process is considered to have returned to normal after the 400th sample. In contrast, TSD-LKPCA (Figure 6d) still alarms the fault after that time, and hence, it achieves a 100% fault detection rate. This result also demonstrates that the TSD-LKPCA method outperforms PCA, KPCA, and LKPCA.

5. Conclusions, Limitations, and Future Research

This paper proposed a TSD-LKPCA method for handling the dynamic and nonlinear features simultaneously. Our novel contribution proposes a novel two-step dynamic scheme and integrates it into the LKPCA technique naturally. As such, TSD-LKPCA can successfully extract the static component in the data and handle its nonlinear feature by LKPCA. The testing results in a dynamic and nonlinear numerical model show that TSD-LKPCA can successfully detect all types of faults, and the testing result in the TE process shows that TSD-LKPCA achieves a higher fault detection rate than PCA, KPCA, and LKPCA by more than 9%. As such, TSD-LKPCA is a promising method.
Another contribution of this paper is that it analyzed the influence of the dynamic feature on the clustering results of nearby points in detail and proved that LKPCA is not applicable to dynamic processes.
However, it is worth noting that the selection of parameter σ in the Gaussian kernel function and parameter k n in the KNN method is still a problem to be solved. More accurate results can be obtained by using more advanced parameter optimization methods. After some modifications, the algorithm for parameter optimization can be designed, which will be considered for future work.

Author Contributions

Conceptualization, H.F.; methodology, H.F. and Z.L.; validation, W.T. and S.L.; formal analysis, H.F.; resources, Y.W. and Y.X.; writing—original draft preparation, H.F.; writing—review and editing, Z.L. and W.T.; visualization, H.F. and W.T.; supervision, S.L.; project administration, Z.L.; funding acquisition, S.L., Z.L. and Y.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Innovation Team of the Department of Education of Guangdong Province, China (2020KCXTD041), the School Level Scientific Research Project of SZPT, China (6022310005K), and Young Talents by the Department of Education of Guangdong Province, China (2021KQNCX210).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gao, Z.; Liu, X. An Overview on Fault Diagnosis, Prognosis and Resilient Control for Wind Turbine Systems. Processes 2021, 9, 300. [Google Scholar] [CrossRef]
  2. Aslam, M.; Bantan, R.A.R.; Khan, N. Monitoring the Process Based on Belief Statistic for Neutrosophic Gamma Distributed Product. Processes 2019, 7, 209. [Google Scholar] [CrossRef] [Green Version]
  3. Quiñones-Grueiro, M.; Prieto-Moreno, A.; Verde, C.; Llanes-Santiago, O. Data-driven monitoring of multimode continuous processes: A review. Chemom. Intell. Lab. Syst. 2019, 189, 56–71. [Google Scholar] [CrossRef]
  4. Qin, S.J. Survey on data-driven industrial process monitoring and diagnosis. Annu. Rev. Control 2012, 36, 220–234. [Google Scholar] [CrossRef]
  5. Abdi, H.; Williams, L.J. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
  6. Zhou, P.; Zhang, R.; Xie, J.; Liu, J.; Wang, H.; Chai, T. Data-driven monitoring and diagnosing of abnormal furnace conditions in blast furnace ironmaking: An integrated PCA-ICA method. IEEE Trans. Ind. Electron. 2020, 68, 622–631. [Google Scholar] [CrossRef]
  7. Lou, Z.; Wang, Y.; Lu, S.; Sun, P. Process Monitoring Using a Novel Robust PCA Scheme. Ind. Eng. Chem. Res. 2021, 60, 4397–4404. [Google Scholar] [CrossRef]
  8. Song, Y.; Liu, J.; Zhang, L.; Wu, D. Improvement of Fast Kurtogram Combined with PCA for Multiple Weak Fault Features Extraction. Processes 2020, 8, 1059. [Google Scholar] [CrossRef]
  9. Lei, Y.; Jiang, W.; Jiang, A.; Zhu, Y.; Niu, H.; Zhang, S. Fault diagnosis method for hydraulic directional valves integrating PCA and XGBoost. Processes 2019, 7, 589. [Google Scholar] [CrossRef] [Green Version]
  10. Fu, Y.; Gao, Z.; Liu, Y.; Zhang, A.; Yin, X. Actuator and Sensor Fault Classification for Wind Turbine Systems Based on Fast Fourier Transform and Uncorrelated Multi-Linear Principal Component Analysis Techniques. Processes 2020, 8, 1066. [Google Scholar] [CrossRef]
  11. Kresta, J.V.; MacGregor, J.F.; Marlin, T.E. Multivariate statistical monitoring of process operating performance. Can. J. Chem. Eng. 1991, 69, 35–47. [Google Scholar] [CrossRef]
  12. Dong, J.; Zhang, K.; Huang, Y.; Li, G.; Peng, K. Adaptive total PLS based quality-relevant process monitoring with application to the Tennessee Eastman process. Neurocomputing 2015, 154, 77–85. [Google Scholar] [CrossRef]
  13. Kano, M.; Tanaka, S.; Hasebe, S.; Hashimoto, I.; Ohno, H. Monitoring independent components for fault detection. AIChE J. 2003, 49, 969–976. [Google Scholar] [CrossRef] [Green Version]
  14. Fan, J.; Qin, S.J.; Wang, Y. Online monitoring of nonlinear multivariate industrial processes using filtering KICA–PCA. Control Eng. Pract. 2014, 22, 205–216. [Google Scholar] [CrossRef]
  15. Cartocci, N.; Napolitano, M.R.; Crocetti, F.; Costante, G.; Valigi, P.; Fravolini, M.L. Data-Driven Fault Diagnosis Techniques: Non-Linear Directional Residual vs. Machine-Learning-Based Methods. Sensors 2022, 22, 2635. [Google Scholar] [CrossRef]
  16. Dunia, R.; Qin, S.J. Joint diagnosis of process and sensor faults using principal component analysis. Control Eng. Pract. 1998, 6, 457–469. [Google Scholar] [CrossRef]
  17. Lee, J.-M.; Yoo, C.; Choi, S.W.; Vanrolleghem, P.A.; Lee, I.-B. Nonlinear process monitoring using kernel principal component analysis. Chem. Eng. Sci. 2004, 59, 223–234. [Google Scholar] [CrossRef]
  18. Jiang, Q.; Yan, X. Parallel PCA–KPCA for nonlinear process monitoring. Control Eng. Pract. 2018, 80, 17–25. [Google Scholar] [CrossRef]
  19. Zhang, Y. Enhanced statistical analysis of nonlinear processes using KPCA, KICA and SVM. Chem. Eng. Sci. 2009, 64, 801–811. [Google Scholar] [CrossRef]
  20. Zeng, L.; Long, W.; Li, Y. A Novel Method for Gas Turbine Condition Monitoring Based on KPCA and Analysis of Statistics T2 and SPE. Processes 2019, 7, 124. [Google Scholar] [CrossRef] [Green Version]
  21. Zhou, M.; Zhang, Q.; Liu, Y.; Sun, X.; Cai, Y.; Pan, H. An Integration Method Using Kernel Principal Component Analysis and Cascade Support Vector Data Description for Pipeline Leak Detection with Multiple Operating Modes. Processes 2019, 7, 648. [Google Scholar] [CrossRef] [Green Version]
  22. Wang, Y.; Si, Y.; Huang, B.; Lou, Z. Survey on the theoretical research and engineering applications of multivariate statistics process monitoring algorithms: 2008–2017. Can. J. Chem. Eng. 2018, 96, 2073–2085. [Google Scholar] [CrossRef]
  23. Lahdhiri, H.; Elaissi, I.; Taouali, O.; Harakat, M.F.; Messaoud, H. Nonlinear process monitoring based on new reduced Rank-KPCA method. Stoch. Environ. Res. Risk Assess. 2018, 32, 1833–1848. [Google Scholar] [CrossRef]
  24. Peng, G.; Huang, K.; Wang, H. Dynamic multimode process monitoring using recursive GMM and KPCA in a hot rolling mill process. Syst. Sci. Control Eng. 2021, 9, 592–601. [Google Scholar] [CrossRef]
  25. Deng, X.; Tian, X.; Chen, S. Modified kernel principal component analysis based on local structure analysis and its application to nonlinear process fault diagnosis. Chemom. Intell. Lab. Syst. 2013, 127, 195–209. [Google Scholar] [CrossRef] [Green Version]
  26. Parra, L.; Sajda, P. Blind source separation via generalized eigenvalue decomposition. J. Mach. Learn. Res. 2003, 4, 1261–1269. [Google Scholar]
  27. Kramer, O. K-nearest neighbors. In Dimensionality Reduction with Unsupervised Nearest Neighbors; Springer: Berlin/Heidelberg, Germany, 2013; pp. 13–23. [Google Scholar]
  28. Wang, J.; Zhou, Z.; Li, Z.; Du, S. A Novel Fault Detection Scheme Based on Mutual k-Nearest Neighbor Method: Application on the Industrial Processes with Outliers. Processes 2022, 10, 497. [Google Scholar] [CrossRef]
  29. Lou, Z.; Tuo, J.; Wang, Y. Two-step principal component analysis for dynamic processes. In Proceedings of the 2017 6th International Symposium on Advanced Control of Industrial Processes (AdCONIP), Taipei, Taiwan, 28–31 May 2017; pp. 73–77. [Google Scholar]
  30. Lou, Z.; Wang, Y.; Si, Y.; Lu, S. A novel multivariate statistical process monitoring algorithm: Orthonormal subspace analysis. Automatica 2022, 138, 110148. [Google Scholar] [CrossRef]
  31. Chiang, L.H.; Russell, E.L.; Braatz, R.D. Fault Detection and Diagnosis in Industrial Systems; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2000. [Google Scholar]
  32. Xu, Y.; Zhang, D.; Song, F.; Yang, J.-Y.; Jing, Z.; Li, M. A method for speeding up feature extraction based on KPCA. Neurocomputing 2007, 70, 1056–1061. [Google Scholar] [CrossRef]
  33. Lou, Z.; Shen, D.; Wang, Y. Preliminary-summation-based principal component analysis for non-Gaussian processes. Chemom. Intell. Lab. Syst. 2015, 146, 270–289. [Google Scholar] [CrossRef]
  34. Deng, X.; Cai, P.; Cao, Y.; Wang, P. Two-step localized kernel principal component analysis based incipient fault diagnosis for nonlinear industrial processes. Ind. Eng. Chem. Res. 2020, 59, 5956–5968. [Google Scholar] [CrossRef]
  35. Lou, Z.; Shen, D.; Wang, Y. Two-step principal component analysis for dynamic processes monitoring. Can. J. Chem. Eng. 2018, 96, 160–170. [Google Scholar] [CrossRef]
  36. Han, M.; Zhang, S.; Xu, M.; Qiu, T.; Wang, N. Multivariate chaotic time series online prediction based on improved kernel recursive least squares algorithm. IEEE Trans. Cybern. 2018, 49, 1160–1172. [Google Scholar] [CrossRef] [PubMed]
  37. Ma, X.; Si, Y.; Yuan, Z.; Qin, Y.; Wang, Y. Multistep dynamic slow feature analysis for industrial process monitoring. IEEE Trans. Instrum. Meas. 2020, 69, 9535–9548. [Google Scholar] [CrossRef]
  38. Cong, Y.; Zhou, L.; Song, Z.; Ge, Z. Multirate dynamic process monitoring based on multirate linear Gaussian state-space model. IEEE Trans. Autom. Sci. Eng. 2019, 16, 1708–1719. [Google Scholar] [CrossRef]
Figure 1. Cluster results for static process.
Figure 1. Cluster results for static process.
Processes 10 00925 g001
Figure 2. Cluster results for dynamic static process: (a) clustered by variable value and (b) clustered by variable deviation.
Figure 2. Cluster results for dynamic static process: (a) clustered by variable value and (b) clustered by variable deviation.
Processes 10 00925 g002
Figure 3. Static components extracted from the dynamic process data.
Figure 3. Static components extracted from the dynamic process data.
Processes 10 00925 g003
Figure 4. Monitoring charts of fault 1 in the numerical example: (a) PCA, (b) KPCA, (c) LKPCA (d) TSD-LKPCA.
Figure 4. Monitoring charts of fault 1 in the numerical example: (a) PCA, (b) KPCA, (c) LKPCA (d) TSD-LKPCA.
Processes 10 00925 g004
Figure 5. Monitoring charts of fault 2 in the numerical example: (a) PCA, (b) KPCA, (c) LKPCA (d) TSD-LKPCA.
Figure 5. Monitoring charts of fault 2 in the numerical example: (a) PCA, (b) KPCA, (c) LKPCA (d) TSD-LKPCA.
Processes 10 00925 g005
Figure 6. Monitoring charts of fault 5: (a) PCA, (b) KPCA, (c) LKPCA (d) TSD-LKPCA.
Figure 6. Monitoring charts of fault 5: (a) PCA, (b) KPCA, (c) LKPCA (d) TSD-LKPCA.
Processes 10 00925 g006
Table 1. Acronyms and notations used in the present work.
Table 1. Acronyms and notations used in the present work.
SymbolDescription
X Training data
ϕ ( · ) Nonlinear mapping function
Γ ( ) Kernel function
K Kernel matrix
σ Weight parameter
N × N matrix of each element
p Projection vector
α Eigenvalue
k Number of pivots
S i j Local neighborhood relationship
L Laplacian matrix
S Neighbor matrix
U Diagonal matrix
ε Small regularization parameter
I Identity matrix
q Time lag
D Time difference
χ Random uniform distribution
k n Neighborhood relation parameter
Table 2. Acronyms used in the present work.
Table 2. Acronyms used in the present work.
AcronymDescription
GMMGaussian mixture model
ICAIndependent component analysis
KNNK-nearest neighbor
KPCAKernel principal component analysis
LKPCALocal kernel principal component analysis
PCAPrincipal component analysis
PCsPrincipal components
PLSPartial least squares
TETennessee Eastman
TSD-LKPCATwo-step dynamic local kernel principal component analysis
Table 3. Monitoring results (%) of PCA, KPCA, LKPCA, and TSD-LKPCA.
Table 3. Monitoring results (%) of PCA, KPCA, LKPCA, and TSD-LKPCA.
MethodsPCAKPCALKPCATSD-LKPCA
Indices T 2 S P E T 2 S P E T 2 S P E T 2 S P E
Fault 198.0427.4596.6796.6796.6796.67100.00100.00
Fault 20.00100.000.00100.000.00100.00100.00100.00
Average49.0262.7348.4398.3448.4398.34100.00100.00
False alarm rates15.110.222.110.7752.111.671.110.45
Table 4. Fault descriptions for the Tennessee Eastman (TE) process.
Table 4. Fault descriptions for the Tennessee Eastman (TE) process.
No.DescriptionTypeNo.DescriptionType
1A/C feed ratio, B composition constantStep9D feed temperatureRandom variation
2B composition, A/C ratio constantStep10C feed temperatureRandom variation
3D feed temperatureStep11Reactor cooling water inlet temperatureRandom variation
4Reactor cooling water inlet temperatureStep12Condenser cooling water inlet temperatureRandom variation
5Condenser cooling water inlet temperatureStep13Reaction kineticsSlow drift
6A feed lossStep14Reactor cooling water valveSticking
7C header pressure loss—reduced availabilityStep15Condenser cooling water valveSticking
8A, B, C feed compositionRandom variation16–21UnknownUnknown
Table 5. Monitoring results on the TE process (%).
Table 5. Monitoring results on the TE process (%).
MethodsPCAKPCALKPCATSD-LKPCA
Indices T 2 S P E T 2 S P E T 2 S P E T 2 S P E
Fault199.1099.0099.2599.0099.2599.3899.6399.75
Fault298.4095.8098.6398.8898.5098.7599.1398.13
Fault30.902.603.0021.500.635.501.251.50
Fault420.90100.0054.3898.885.2583.7584.1376.50
Fault524.3020.9027.8840.8823.5027.63100.00100.00
Fault699.10100.0099.1399.5099.0099.38100.00100.00
Fault7100.00100.00100.00100.0099.25100.0088.63100.00
Fault896.9083.6097.5096.2597.0097.7597.1397.63
Fault91.801.803.6318.251.754.131.380.63
Fault1029.9025.8041.1353.6333.8833.2580.5082.13
Fault1140.6074.9056.5069.7519.1365.8869.3861.25
Fault1298.4089.5098.6396.5098.0095.3899.1399.38
Fault1393.6095.3094.5096.0094.2595.2594.7598.50
Fault1499.30100.0099.50100.0079.13100.0099.6399.88
Fault151.403.005.2522.132.636.881.131.63
Fault1613.5027.4025.0043.5017.5020.0080.3880.88
Fault1776.4095.4080.3893.2570.6384.7594.5096.25
Fault1889.3090.1089.0092.7588.6389.1389.8893.00
Fault1911.0012.5011.1330.630.5023.0057.0054.25
Fault2031.8049.8036.1361.2529.0038.0085.6363.88
Fault2139.3047.3040.0062.3830.0048.7550.8833.38
Average55.5262.6060.0371.1951.7862.6974.9673.26
False alarm rates2.3212.200.515.330.001.610.361.13
Difference53.2050.4059.5265.8651.7861.0874.6072.13
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Fang, H.; Tao, W.; Lu, S.; Lou, Z.; Wang, Y.; Xue, Y. Nonlinear Dynamic Process Monitoring Based on Two-Step Dynamic Local Kernel Principal Component Analysis. Processes 2022, 10, 925. https://doi.org/10.3390/pr10050925

AMA Style

Fang H, Tao W, Lu S, Lou Z, Wang Y, Xue Y. Nonlinear Dynamic Process Monitoring Based on Two-Step Dynamic Local Kernel Principal Component Analysis. Processes. 2022; 10(5):925. https://doi.org/10.3390/pr10050925

Chicago/Turabian Style

Fang, Hairong, Wenhua Tao, Shan Lu, Zhijiang Lou, Yonghui Wang, and Yuanfei Xue. 2022. "Nonlinear Dynamic Process Monitoring Based on Two-Step Dynamic Local Kernel Principal Component Analysis" Processes 10, no. 5: 925. https://doi.org/10.3390/pr10050925

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop