Next Article in Journal
A 3D Keypoints Voting Network for 6DoF Pose Estimation in Indoor Scene
Next Article in Special Issue
Auxiliary Model-Based Multi-Innovation Fractional Stochastic Gradient Algorithm for Hammerstein Output-Error Systems
Previous Article in Journal
Machinery Foundations Dynamical Analysis: A Case Study on Reciprocating Compressor Foundation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Process Monitoring Method Based on Dynamic Autoregressive Latent Variable Model and Its Application in the Sintering Process of Ternary Cathode Materials

1
School of Automation, Central South University, Changsha 410083, China
2
State Key Laboratory of High Performance Complex Manufacturing, Central South University, Changsha 410083, China
3
Hunan Shanshan Energy Technology Co., Ltd., Changsha 410205, China
*
Authors to whom correspondence should be addressed.
Machines 2021, 9(10), 229; https://doi.org/10.3390/machines9100229
Submission received: 8 September 2021 / Revised: 2 October 2021 / Accepted: 3 October 2021 / Published: 7 October 2021
(This article belongs to the Special Issue Deep Learning-Based Machinery Fault Diagnostics)

Abstract

:
Due to the ubiquitous dynamics of industrial processes, the variable time lag raises great challenge to the high-precision industrial process monitoring. To this end, a process monitoring method based on the dynamic autoregressive latent variable model is proposed in this paper. First, from the perspective of process data, a dynamic autoregressive latent variable model (DALM) with process variables as input and quality variables as output is constructed to adapt to the variable time lag characteristic. In addition, a fusion Bayesian filtering, smoothing and expectation maximization algorithm is used to identify model parameters. Then, the process monitoring method based on DALM is constructed, in which the process data are filtered online to obtain the latent space distribution of the current state, and T2 statistics are constructed. Finally, by comparing with an existing method, the feasibility and effectiveness of the proposed method is tested on the sintering process of ternary cathode materials. Detailed comparisons show the superiority of the proposed method.

1. Introduction

To ensure production safety and product quality, process monitoring technology has become an indispensable ingredient for industrial processes in recent years. It is commonly divided into model-based methods and data-driven methods. Compared with the former ones, the later ones can take advantage of the routine measurement and do not rely on process prior knowledge and precise mechanism models, which are unavailable or cost-intensive to obtained at times [1,2]. Therefore, they are widely used in modern industrial process.
During the past decades, many data-driven process monitoring methods have been published [3,4,5,6]. Kim et al. [7] proposed a probabilistic PCA to monitoring industrial processes, which firstly extracts redundant information from the variables and constructs feature distribution for monitoring, but it only extracts features of the input space. Zhao et al. [8] proposed the probabilistic PLSR process monitoring method to monitor quality-related faults, which can simultaneously consider the fault characteristics of the input and output spaces for monitoring. Furthermore, Chen et al. [9] proposed a probability-related PCA method for detecting incipient faults, which can greatly improve the detection ability of minor faults. Probabilistic framework modeling can overcome process noise [10]. However, the process monitoring methods currently proposed are all static methods, and the actual production processes are dynamical featured with variable time lag [11,12].
Process dynamics could refer to the mutual influence before and after the current sampling [13]. To deal with the dynamics of process, Ku et al. [14] built an augmented matrix and extended the static PCA model to the dynamic PCA (DPCA) for process monitoring. However, the introduction of augmented matrix increases the parameter dimensions called the curse of dimensionality [15]. Motivated by DPCA, Li et al. [16] proposed a dynamic latent variable model for monitoring the Tennessee Eastman process. In this model, the autoregressive model is used to extract data dynamic information, and PCA is performed to reduce redundancy between variables. It divides variable order reduction and dynamic information extraction into two stages, which makes the system complex and not easy to tune. In addition, compared with process variables, quality variables also contain useful fault information [17]. For this reason, Ge et al. [18] proposed a supervised linear dynamic system model process monitoring method. This method uses a first-order autoregressive equation to simulate the first-order dynamic [19] but does not take the variable time lag into account.
Variable time lag refers to the delay between the effects of variables [20]. The existing monitoring methods considering time lag are usually divided into two categories [21]. One is to find the time lag between variables and translate the data to eliminate the time lag and then establish a static process monitoring model for the processed data. For example, Wang et al. [22] proposed a spatial reconstruction method to identify system time lag, then aligned the data and established a monitoring model, but the alignment operation will destroy the data structure and cause data loss. The other idea is to use time lag as an unknown parameter of the process monitoring model and identify the parameters through a data-driven method. For example, Huber et al. [23] proposed to take the time lag as a parameter of a high-order state space system model and then solve it uniformly with the model parameters, but this method relies on the setting of the time lag parameter and the parameter identification method.
From the above discussions, it can be observed that the variable time lag characteristic of a process makes the previous work unfavorable. However, this characteristic is common in industrial processes [24,25]. To deal with this problem, this paper proposes a process monitoring method based on a dynamic autoregressive latent variable model. Firstly, from the data point of view, a linear dynamic model is constructed between process variables and quality variables, and the dynamic information of process input and process output is compressed to latent variables, and then a dynamic autoregressive latent variable model (DALM) is constructed for latent variables to extract variable time lag information. In addition, a fusion Bayesian filtering, smoothing and expectation maximization algorithm is used to identify model parameters. Then, the DALM is applied to the industrial monitoring process. The process variables are filtered through improved Bayesian filtering technology to obtain the latent space distribution of the current state, and the T2 statistics of the latent space are constructed and monitored [26] to realize the process monitoring task. The main contribution can be concluded as (1) a process monitoring method based on dynamic autoregressive latent variable model is proposed in this paper; (2) a dynamic autoregressive latent variable model (DALM) is developed to extract variable time lag information; (3) a fusion Bayesian filtering, smoothing and expectation maximization algorithm is improved to identify model parameters; (4) based on the DALM, the T2 statistics of the latent space are constructed to realize the process monitoring task.
The main structure of the paper is arranged as follows. In the second section, a dynamic autoregressive latent variable model is proposed, and the parameter identification algorithm of the model is derived in detail. A process monitoring method based on DALM is proposed in the third chapter. The fourth section uses the monitoring method to monitor the sintering process of the ternary cathode material to verify the monitoring performance of the proposed method. Finally, the last section concludes.

2. Modeling Method Based on Dynamic Autoregressive Latent Variable Model

This section proposes a dynamic autoregressive latent variable modeling method for the accurate modeling of dynamical industrial processes. Bayesian filtering and smoothing inference were used to obtain the spatial distribution of latent variables, and the parameters of the model were identified by combining with the EM algorithm.

2.1. Dynamic Autoregressive Latent Variable Model Structure

In order to consider the dynamic characteristics of the process, the traditional probability latent variable model [27] establishes the relationship between the current moment and the previous moment data, as shown in (1).
z t = A z t 1 + η t z , x t = B x z t + η t x , y t = B y z t + η t y ,
where the structure consists of a linear Gaussian dynamic equation and two linear Gaussian observation equations, z t is the latent variable of the process state at time t;  x t and y t are the process variable and quality variable at time t, respectively; A , B x and B y are their own load matrix. The Gaussian dynamic equation is used to describe the dynamic relationship of the process data. The observation equation compresses the information of the process data into low-dimensional latent variables. Therefore, an accurate mathematical model can be established for the dynamic process, but the structure does not consider the time lag characteristics of the process.
In order to further consider the characteristics of process time lag, on the basis of dynamic probabilistic latent variable model (DPLVM) [13], the trend similarity analysis algorithm [22] was first used to obtain the time lag coefficient L of the current process, and then an autoregressive equation was constructed for the latent variables to describe the variable time lag information. Among them, the autoregressive equation models the process dynamics and time lag characteristics, and the linear observation equation models the cross-correlation of data. The probability graph model of the model is shown in Figure 1, and the mathematical expression is shown in (2).
z t = A h t 1 + η t z , x t = B x z t + η t x , y t = B y z t + η t y ,
where z t R d represents the latent variable of the process state at time t, h t 1 = [ z t 1 T z t 2 T z t L T ] T R d L is the augmented state variable containing the latent variables at time L in the past, x t R v is the observed value of the process variable at time t, y t R k is the observed value of the quality variable at time t and d ,   v ,   k , respectively, correspond to the dimensions of latent variables, process variables and quality variables. A R d × d L is the state transition matrix, L is the time lag value of the process, and B x R v × d , B y R k × d are the state divergence matrices. η t z R d , η t x R v , and η t y R k are Gaussian noise terms of latent variables, process variables and quality variables, respectively. Assuming that the noises are independent of each other, the distributions obeyed are η t z ~ N ( 0 , Σ z ) , η t x ~ N ( 0 , Σ x ) and η t y ~ N ( 0 , Σ y ) , respectively. Latent variables represent the current state of the process, and this model is an extension of the traditional DPLVM.

2.2. Parameter Identification Based on EM Algorithm

Since only x t and y t can be observed in the process, and latent variables are abstracted to describe the state of the process and are unobservable, the EM algorithm was used to identify the parameters of the model [28]. Each iteration of the EM algorithm consisted of two steps: E step, seeking expectation (exception); M step, seeking maximization (maximization). This section uses Bayesian filtering and smoothing to infer the spatial distribution of latent variables, so as to solve the difficult problem of calculating latent variable statistics.
Under the framework of probability, the model assumed that the latent variables at the initial moment obeyed a Gaussian distribution with mean u 0 and variance V 0 , that is, z 0 ,   z 1 ,   ,   z 1 L ~ N ( u 0 , V 0 ) . From the knowledge of probability theory [16], it is easy to get that the distribution of the latent variable z t , the process variable x t and the quality variable y t obey the Gaussian distribution, as shown in (3).
z t | z t 1 , z t 2 , , z t L ~ N ( A 1 z t 1 + A 2 z t 2 + + A L z t L , Σ z ) , x t | z t ~ N ( B x z t , Σ x ) , y t | z t ~ N ( B y z t , Σ y ) .
The parameters that needed to be identified were denoted as Θ = { A , B x , B y , u 0 , V 0 , Σ z , Σ x , Σ y } , of which A = [ A 1 , A 2 , , A L ] . According to the naive what you see is what you get thought, the parameter identification problem was transformed into the maximum observation data x 1 : T , y 1 : T . The log-likelihood function on the parameter Θ is shown in (4), where x 1 : T represents the observation sequence of the process variable x 1 ,   x 2 ,   ,   x T , y 1 : T represents the observation sequence y 1 ,   y 2 ,   ,   y T of the quality variable, where T is the total number of training samples, namely,
Θ n e w = arg   max Θ   log P ( x 1 : T , y 1 : T | Θ ) .
The EM algorithm [29] was used to solve the optimization problem of Equation (4). In the E step of the EM algorithm, the log-likelihood function log P ( x 1 : T , y 1 : T , z 1 L : T | Θ ) of the complete data had to be calculated with respect to the conditional expectation of the latent variable z 1 L : T to obtain the objective cost function (Q function), as shown in (5).
Q ( Θ | Θ o l d ) = E z 1 L : T | ( x 1 : T , y 1 : T , Θ o l d ) { log P ( x 1 : T , y 1 : T , z 1 L : T | Θ ) } .
Actually, the likelihood function can be formulated by the application of the product rule of probability. From the model structure, the log-likelihood function of the complete data was expanded, as expressed by (6).
log P ( x 1 : T , y 1 : T , z ( L + 1 ) : T | Θ ) = log { P ( z 0 , z 1 , , z L + 1 ) t = 1 T P ( z t | z t 1 , z t 2 , , z t L ) P ( x t | z t ) P ( y t | z t ) } = log P ( z 0 , z 1 , , z L + 1 ) + t = 1 T log P ( z t | z t 1 , z t 2 , , z t L ) + t = 1 T log P ( x t | z t ) + t = 1 T log P ( y t | z t ) .
For clear writing, we denote E z T ( f ( x t , y t , z t ) ) = E z 1 L : T | ( x 1 : T , y 1 : T , Θ o l d ) ( f ( x t , y t , z t ) ) , then the expectation of the complete data likelihood function log P ( x 1 : T , y 1 : T , z ( L + 1 ) : T | Θ ) with respect to the latent variable distribution P ( z ( L + 1 ) : T | x 1 : T , y 1 : T , Θ o l d ) is shown in (7).
Q ( Θ | Θ o l d ) = E z T { log p ( x 1 : T , y 1 : T , z 1 L : T | Θ ) } = 1 2 { log | V 0 | + E z T ( [ z 0 z L + 1 ] T V 0 1 [ z 0 z L + 1 ] ) 2 E z T ( [ z 0 z L + 1 ] T ) V 0 1 u 0 + u 0 T V 0 1 u 0 } 1 2 { T log | Σ z | + t = 1 T { E z T ( z t T Σ z 1 z t ) 2 E z T ( [ z t 1 z t L ] T A T Σ z 1 z t ) + E z T ( [ z t 1 z t L ] T A T Σ z 1 A [ z t 1 z t L ] ) } } 1 2 { T log | Σ x | + t = 1 T { x t T Σ x 1 x t 2 E z T ( z t T ) B x T Σ x 1 x t + E z T ( z t T B x T Σ x 1 B x z t ) } } 1 2 { T log | Σ y | + t = 1 T { y t T Σ y 1 y t 2 E z T ( z t T ) B y T Σ y 1 y t + E z T ( z t T B y T Σ y 1 B y z t ) } } + c o n s tan t .
Appendix A provides a detailed update of all parameters at step M. From (7), the related statistics of latent variables in the Q function include E z T ( z t ) , E z T ( z t z t T ) and E z T ( z t z t i T ) , where t = 0 , 1 , , T , i = 1 , 2 , , L , in fact, these statistics can be passed. The posterior probability distribution of the latent variables obtained in the E step of the EM algorithm was obtained. The calculation results of these statistics are shown in (8). The detailed derivation process is shown in Appendix B.
{ E z T ( z t ) = E ( z t | x 1 : T , y 1 : T , Θ o l d ) = m t 1 E z T ( z t z t T ) = E ( z t z t T | x 1 : T , y 1 : T , Θ o l d ) = M t 11 + m t 1 ( m t 1 ) T E z T ( z t z t i T ) = E ( z t z t i T | x 1 : T , y 1 : T , Θ o l d ) = M t 1 ( i + 1 ) + m t 1 ( m t ( i + 1 ) ) T E z T ( z t z t L T ) = E ( z t z t L T | x 1 : T , y 1 : T , Θ o l d ) = i = 1 L A i ( M t 1 i L + m t 1 i ( m t 1 L ) T ) ,
where m t and M t are the mean and covariance of the posterior probability distribution of the latent variable. Therefore, through E step and M step iterative update until the parameters converge, the optimized parameter set Θ o p t = { A , B x , B y , u 0 , V 0 , Σ z , Σ x , Σ y } can be obtained.

3. Process Monitoring Method Based on Dynamic Autoregressive Latent Variable Model

In this section, the established dynamic autoregressive latent variable model is used for industrial process monitoring. At first, DALM was used to model the process data so that the current state information was reflected in the latent variables, and then the latent space at the current time was obtained by filtering the process data distribution, constructing statistics and monitoring them. Let us introduce the monitoring process in detail below.
Although the latent space was unobservable, the establishment of a data-driven DALM model based on the characteristics of the process data extracted the information of the process variables to the spatial distribution of the latent variables. The process input X = [ x 1 , x 2 , , x N ] and output Y = [ y 1 , y 2 , , y N ] needed to be pre-processed by the normalization method, as shown in (9).
X q = ( x u x ) · s t d x 1 , Y q = ( y u y ) · s t d y 1 ,
where u x and u y are the means of the variables X and Y , s t d x and s t d y are the variances of the variables X and Y . Preprocessed data were filtered through the filtering algorithm to obtain the spatial distribution of the latent variables, as shown in (10).
z t q , z t 1 q , , z t L + 1 q | x 1 : t q , y 1 : t q ~ N ( u t q , V t q ) = N ( [ u t 1 q u t L q ] , [ V t 11 q V t 1 L q V t ( L ) 1 q V t L L q ] ) .
Among them, u t 1 q and V t 11 q are the mean and variance of the latent space distribution, respectively, which were obtained from (11). The detailed derivation process is shown in (A12)–(A16).
u t 1 q = g t 1 + [ G t 1 ( L + 1 ) G t 1 ( L + 2 ) ] [ G t ( L + 1 ) ( L + 1 ) G t ( L + 1 ) ( L + 2 ) G t ( L + 2 ) ( L + 1 ) G t ( L + 2 ) ( L + 2 ) ] [ x t g t L + 1 y t g t L + 2 ] , V t 11 q = [ G t 11 G t 1 L ] [ G t 1 ( L + 1 ) G t 1 ( L + 2 ) ] [ G t ( L + 1 ) ( L + 1 ) G t ( L + 1 ) ( L + 2 ) G t ( L + 2 ) ( L + 1 ) G t ( L + 2 ) ( L + 2 ) ] 1 [ G t 1 ( L + 1 ) G t 1 ( L + 2 ) G t L ( L + 1 ) G t L ( L + 2 ) ] T .
It can be seen from (A24) that the information of the data X t q = [ x 1 q , x 2 q , , x t q ] and Y t q = [ y 1 q , y 2 q , , y t q ] at the current and previous moments was filtered into the current latent variable, and the latent variable distribution at the current moment is shown in (12).
z t q | x 1 : t q , y 1 : t q ~ N ( u t 1 q , V t 11 q ) .
Because the latent space contains the current state of the process dynamics and variable time lag information, the process statistic T2 was constructed for the current latent variable at time t, as shown in (13).
T t , q 2 = E ( z t q | x 1 : t q , y 1 : t q ) T c o v a r i a n c e ( z t q | x 1 : t q , y 1 : t q ) 1 E ( z t q | x 1 : t q , y 1 : t q ) .
Among them, the mathematical expectation and variance of the latent variables on the observation data at the current moment are shown in (14).
E ( z t q | x 1 : t q , y 1 : t q ) = u t 1 q , c o v a r i a n c e ( z t q | x 1 : t q , y 1 : t q ) = V t 11 q .
The probability of the latent variable obeyed the Gaussian distribution. Therefore, according to the definition of chi-square distribution, this statistic obeyed the chi-square distribution χ α 2 ( d ) after data preprocessing. Then, combining to the latent variable dimension d of the model and the significance level α required by the industry, the control threshold T lim 2 of the process monitoring method was obtained, and then the statistics of each time data were calculated online and compared with the control threshold, to determine whether the process deviated from the normal state. The process monitoring logic is determined by (15).
T t , q 2 < T lim 2 = χ α 2 ( d ) .
Too large an α value will lead to a high false alarm rate, and too low an α will lead to a high false alarm rate; therefore, in practice, it is a balance between false alarms and missed alarms. This paper chose α as 0.01, which means that the false positive rate of normal data was 0.01. If T t , q 2 < T lim 2 , the system was in a normal state. Otherwise, the process located in a fault state, and further diagnosis and identification of the fault was required for process maintenance. The process of DALM modeling and online process monitoring is shown in Figure 2.
The main steps of the process monitoring method based on the DALM model were as follows:
Step 1: Collect process data, divide the training and test data sets and standardize them.
Step 2: Use the training data set to learn the parameters of the DALM model.
Step 3: Build the model and determine the control threshold.
Step 4: Filter the process data online to get the latent space distribution at the current moment.
Step 5: Calculate statistics and compare with the control threshold to determine whether the process is abnormal.

4. Case Study on the Sintering Process of Ternary Cathode Materials

In this section, the proposed process monitoring method based on the dynamic autoregressive latent variable model is used to monitor the sintering process of ternary cathode materials to verify the effectiveness of the method. First the sintering process technology of the ternary cathode material was introduced, then the model structure and parameter determination were introduced in detail, and finally the performance of the model was evaluated.

4.1. Introduction to the Sintering Process of Ternary Cathode Materials

The rapid development of the new energy industry has led to an extremely urgent demand for high-quality ternary cathode materials, and the sintering process of battery materials is the core and key process of battery preparation. This process consists of a series connection of a heating section, a constant temperature section and a cooling section, as shown in Figure 3. The optimal production state of a single temperature section cannot guarantee that the product performance indicators of the entire sintering process are within the optimal range; at the same time, changes in the sintering process, such as environmental humidity or temperature, also affect the stability of product performance indicators. In order to ensure the stability of product performance indicators as much as possible, while reducing energy consumption and material consumption, it is necessary to adjust the sintering parameters of the kiln according to the sintering state in real time, which leads to many variables in each temperature zone and series coupling, which makes the process data present complex process characteristics [29].
The temperature field in the sintering process has a significant effect on the material properties. Over-firing will cause changes in the material morphology and internal structure, and under firing will not provide sufficient activation energy for chemical reactions. However, the decomposition reaction that occurs in the heating section is an endothermic process and requires sufficient heat supply, otherwise a reverse reaction will occur, resulting in inefficient water removal, which will affect the subsequent oxidation reaction. Therefore, the state of the heating section is very important to the sintering process. At the same time, the residual lithium content can directly reflect the quality of the product. In order to monitor the process status in real time, a monitoring model is established for the temperature and residual lithium content of the heating section.
Huang et al. [30] established a temperature field monitoring model based on the PBF equipment equation to monitor the dynamic sintering process of parts, but this method requires precise grinding tool structure parameters and can only monitor uniformly distributed temperature fields. Egorova et al. [31] tried to combine neural networks and PCA diagnosis method monitor and diagnose the sintering process. This method can locate the fault and diagnose the cause of the fault. However, the introduction of neural networks increases the time and space complexity of the system and ignores the system dynamic and time lag problems.
Due to the severe temperature interval coupling, the process variables exhibit complex characteristics, making the traditional static monitoring methods unable to achieve accurate monitoring results. The dynamic autoregressive latent variable model proposed in this section considers the dynamic and time lag information of the process at the same time, so it is more in line with the sintering process.

4.2. Determination of Model Parameters

This section establishes a monitoring model for the temperature and product quality in the heating section of the sintering process. The heating section contained seven temperature zones, and each temperature zone had two upper and lower temperature measuring points, but the temperature changes in the 4th to 7th temperature zones were not obvious. The temperature of the first three temperature zones was selected as the process variable x t of the model. At the same time, the residual lithium content of the product reflects the quality of the battery, as does the quality variable y t of the model, Table 1 lists the physical meaning of these variables.
To test the monitoring effect of the model under different faults, a total of 2200 continuous time data samples were collected on site with a sampling period of five minutes. The process included a total of three types of faults such as over-temperature, under-temperature and shutdown. For detailed status information, see Table 2.
First, analyze the dynamics of the data and the time lag characteristics of the variables from the data point of view. Figure 4 shows the autocorrelation and cross-correlation diagrams of process data.
Figure 4 shows the correlation and cross-correlation between the first four process variables. The value at time 0 in each figure represents the cross-correlation between variables; the value at non-zero time shows the autocorrelation between variables under different time lags. It is worth mentioning that the cross-correlation index can measure the redundancy of variable information, and the autocorrelation index can indirectly measure the dynamic and time delay information between variables. It can be seen that the cross-correlation performance between the variables was above 0.5, indicating that there was strong redundant information between the variables. At the same time, even if there was a difference of 10 sampling times, the autocorrelation between the variables was still very high, indicating that there were time lags and dynamic characteristics between the variables. Therefore, the establishment of a DALM model for the process can be considered. The emission equation of the model extracts the redundant information of the data, and the autoregressive equation of the model extracts the dynamic and time lag information of the variables. This paper uses the trend similarity algorithm, which constructs the trend similarity function according to the time lag feature and solves it, to determine the time lag coefficient, that is, L = 3.
To verify the rationality of the time lag coefficient, under different time lag coefficients, a dynamic autoregressive latent variable monitoring model was established respectively. Note: In order to avoid the latent variable dimension from interfering with the selection of the time lag coefficient the latent variable dimension selected by Akaike information criterion (AIC) was temporarily used [32]. The false alarm rate (false alarm rate, FAR) and fault detection rate (fault detection rate, FDR) were defined to evaluate and monitor performance indicators, as defined in (16).
F A R = N F A R N n F D R = N F D R N f .
N F A R represents the number of normal samples that were mistakenly detected as abnormal by the monitoring method, and Nn is the number of all normal samples. N F D R represents the number of fault samples correctly monitored by the monitoring method, N f is the number of all abnormal samples. Therefore, the closer the FAR is to the significance level, the better, and the closer the FDR is to 1, the better. The significance level of this work was set to 0.01.
The first 1000 normal samples were selected to train the model, and the data type fault 1 was used to test the monitoring effect of the model. Table 3 shows the indicators of the monitoring results of the new method under different time lag coefficients.
The model did not converge when the time lag coefficient was 5, and when the model time lag coefficient was 3, the error and false alarm rate of the model were the best. Therefore, when the time lag coefficient was 3, the model gave the best performance. In order to visually see the monitoring results of the model, Figure 5 shows the monitoring T2 diagram when the model’s time lag coefficient was 2, 3 and 4.
It can be seen from Figure 5 that when the model time lag coefficient was 2 and 4, it was easy to misclassify the sample. Especially in the fault interval of 201st–400th: the divided normal samples and abnormal samples were close to the monitoring threshold, which shows that the robustness of the model with this time lag is low; when the model had a time lag coefficient of 3, it is insensitive to the noise and the false alarms are the smallest. Hence, its N F A R and N F D R were the best. Therefore, the time lag coefficient obtained by the trend similarity identification algorithm enabled the model to obtain a better monitoring effect.
Next, the latent variable dimension was determined. The latent variable dimension is the result of comprehensively considering the complexity and accuracy of the model. The root mean square error (RMSE) is an indicator to measure the accuracy of the model. The expression is shown in (17).
R M S E = i = 1 N | y i y ^ i | N ,
where N is the number of test samples, y ^ i is the prediction of the true value y i and y ¯ i is the mean value of the true value of the test sample. Samples from the 1st to the 600th were used to train the model, and samples from the 601st to the 1000th were used as the test set. Table 4 shows the root mean square error of model prediction under different latent variable dimensions.
Table 4 shows that the prediction performance of the model tends to be stable after the latent variable dimension increased to 3, which was the balance point between model complexity and accuracy. It is worth mentioning that under the time lag coefficient, the latent variable dimension selected by the AIC algorithm was also 3, so the latent variable dimension was determined to be 3.

4.3. Model Performance Test

This section verifies the effect of the proposed monitoring method, and constructs a first-order dynamic process monitoring method: DPLVM [18] and static process monitoring method: PPLSR [33], which were used to compare with the proposed method. The latent variable dimensions of the model were adjusted to 3.
The first 1000 normal samples were used to train the parameters of the model, and the trained model was monitored for three types of different fault samples. In order to distinguish between normal and abnormal samples, the first 200 samples of each type of failure test set were normal samples, and the last 200 samples were their respective failure samples. Table 5 shows the FAR and FDR of different monitoring methods under different failure test sets, and the last line calculates the average value of different indicators.
It can be seen from Table 5 that the monitoring performance of the proposed method was better than that of the static model PPLSR and the first-order dynamic model DPLVM. Therefore, the detection performance was greatly improved after the autoregressive equation was added to the model to extract the dynamic and time lag information. Compared with the basic first-order dynamic DPLVM fault detection method, DALM considered the time lag characteristics, so the model performance was further improved. The detailed monitoring results of the three methods for the three types of faults are shown in Figure 6, Figure 7, Figure 8.
For each type of fault test set, the first 200 samples were in a normal state, and the last 200 samples were fault samples. It can be seen from Figure 8 that the static model PPLSR easily mistakenly classified normal samples into faulty samples, and it also easily classified faulty samples into normal samples. The error rate of the first-order dynamic model DPLVM was reduced a lot. Furthermore, the FAR based on the DALM fault detection method proposed in this paper was close to the significance level and the FDR was close to 1, verifying that its monitoring performance was greatly improved.

5. Conclusions

A process monitoring method based on the dynamic autoregressive latent variable model was proposed in this paper. Compared with the traditional DPLVM monitoring method, this method not only considered the dynamic characteristics of the process but also considered the complex time lag characteristics, integrated the time lag information into the model, and greatly improved the monitoring performance of the model in the time lag process. First, from the point of data, this method established a dynamic autoregressive latent variable model to adopt the characteristics of dynamics and variable time lag. Then a fusion Bayesian filtering, smoothing and expectation maximization algorithm was used to identify model parameters. Then, on the basis of the identified model, the improved Bayesian filtering technique was used to infer the latent variable distribution of the process state, and the T2 statistic was constructed for the latent space and online monitoring is performed. Finally, the proposed method was applied to the monitoring of the sintering process of ternary cathode materials. Through industrial case studies, the modeling and monitoring results of the proposed method show that the DALM model was better than the static and first-order dynamic modeling process monitoring methods.
An important issue for process monitoring application in industrial processes is the multisampling rate problem. The method proposed in this paper assumed that the input and output data had the same sampling rate. If the sampling rate was inconsistent, some data were deleted by down-sampling. However, a more worthwhile way to try would be to combine semi-supervised learning methods, which can train data on unbalanced input and output data, thereby improving data utilization. Another practical problem is the non-linear relationship between process data, which is very common in industrial processes. How to effectively deal with this problem is worthy of further research in the near future to make the monitoring method more applicable.

Author Contributions

Conceptualization, N.C. and W.G.; methodology, F.H. and J.C.; software, F.H. and Z.C.; validation, J.C.; writing—review and editing, N.C., F.H., J.C., Z.C. and X.L.; supervision, N.C., Z.C. and W.G.; project administration, W.G. and X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded in part by the Key Program of the National Natural Science Foundation of China (62033014) and in part by the Application Projects of Integrated Standardization and New Paradigm for Intelligent Manufacturing from the Ministry of Industry and Information Technology of China in 2016 and in part by the Fundamental Research Funds for the Central Universities of Central South University(2021zzts0700), in part by the Project of State Key Laboratory of High Performance Complex Manufacturing (#ZZYJKT2020-14).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available upon request from the first author. The data are not publicly available due to intellectual property protection.

Acknowledgments

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Detailed Derivation of the M-Step

According to the EM algorithm, all the parameters of the DALM model can be updated in M steps. By maximizing the cost function Q ( Θ | Θ o l d ) , the estimated value Θ n e w of the next iteration parameter was determined, which is shown in (A1).
Θ n e w = arg max Θ Q ( Θ | Θ o l d ) .
The Q function was applied to the partial derivative of the model parameters and the derivative was set to zero.
Q ( Θ | Θ o l d ) Θ = 0 .
The updated value of the model parameter Θ n e w was obtained, as shown in (A3)–(A10).
u 0 n e w = E z T ( [ z 0 z L + 1 ] ) ,
V 0 n e w = E z T ( [ z 0 z L + 1 ] [ z 0 z L + 1 ] T ) E z T ( [ z 0 z L + 1 ] ) E z T ( [ z 0 z L + 1 ] ) T ,
A n e w = t = 1 T E z T ( z t [ z t 1 z t L ] T ) [ t = 1 T E z T ( [ z t 1 z t L ] [ z t 1 z t L ] T ) ] 1 ,
Σ z n e w = 1 T t = 1 T ( E z T ( z t z t T ) 2 A n e w E z T ( [ z t 1 z t L ] z t T ) + A n e w E z T ( [ z t 1 z t L ] [ z t 1 z t L ] T ) A n e w T ) ,
B x n e w = t = 1 T x t E z T ( z t T ) [ t = 1 T E z T ( z t z t T ) ] 1 ,
B y n e w = t = 1 T y t E z T ( z t T ) [ t = 1 T E z T ( z t z t T ) ] 1 ,
Σ z n e w = 1 T t = 1 T ( x t x t T 2 B x n e w E z T ( z t ) x t T + B x n e w E z T ( z t z t T ) B x n e w T ) ,
Σ y n e w = 1 T t = 1 T ( y t y t T 2 B y n e w E z T ( z t ) y t T + B y n e w E z T ( z t z t T ) B y n e w T ) .
The updated parameter set Θ n e w = { A n e w , B x n e w , B y n e w , u 0 n e w , V 0 n e w , Σ z n e w , Σ x n e w , Σ y n e w } , E steps and M steps were iterated until the parameter Θ matrix converged, that is, satisfied (A11), where ς is a sufficiently small constant, and the model parameter identification was completed.
| | Θ n e w Θ o l d | | < ς .
Among them, Θ o l d is the parameter of the last iteration, and Θ n e w is the parameter after this round of iteration. Only when the parameters obtained by two adjacent identifications converged did the algorithm stop calculating. Therefore, the parameter convergence can be guaranteed by the EM algorithm itself.

Appendix B. Detailed Derivation of the E-Step

In order to determine the statistics E z T ( z t ) , E z T ( z t z t T ) and E z T ( z t z t i T ) , the forward and backward algorithm were employed. This is an iterative calculation method, which includes the forward filtering and backward correction step.
In the Bayesian filtering stage, the goal was to calculate the posterior probability of the latent variable [ z t , z t 1 , , z t L + 1 ] with respect to the variable x 1 : t , y 1 : t at time t, given the posterior distribution z t 1 , z t 2 , , z t L | x 1 : t 1 , y 1 : t 1 ~ N ( u t 1 , V t 1 ) of the latent variable [ z t 1 , z t 2 , , z t L ] at the previous time t − 1 on the variable x 1 : t 1 , y 1 : t 1 , as shown in (A12) where 1 t T ,
z t , z t 1 , , z t L + 1 | x 1 : t , y 1 : t ~ N ( u t , V t ) = N ( [ u t 1 u t L ] , [ V t 11 V t 1 L V t L 1 V t L L ] ) .
The joint probability distribution of the latent variables [ z t , z t 1 , , z t L + 1 ] and x t , y t with respect to the variable x 1 : t 1 , y 1 : t 1 is shown in (A13).
z t , z t 1 , , z t L + 1 , x t , y t | x 1 : t 1 , y 1 : t 1 ~ N ( g t , G t ) = N ( [ g t 1 g t L + 2 ] , [ G t 11 G t 1 ( L + 2 ) G t ( L + 2 ) 1 G t ( L + 2 ) ( L + 2 ) ] ) .
The parameters of (A13) can be calculated by (A15)
{ g t 1 = E ( z t | x 1 : t 1 , y 1 : t 1 ) = A u t 1 g t i = E ( z t i + 1 | x 1 : t 1 , y 1 : t 1 ) = u t 1 i 1 g t L + 1 = E ( x t | x 1 : t 1 , y 1 : t 1 ) = C g t 1 g t L + 2 = E ( y t | x 1 : t 1 , y 1 : t 1 ) = P g t 1 ,
{ G t 11 = cov ( z t , z t | x 1 : t 1 , y 1 : t 1 ) = A u t 1 A T + Σ z G t 1 i = cov ( z t , z t i + 1 | x 1 : t 1 , y 1 : t 1 ) = A [ V t 1 1 ( i 1 ) V t 1 L ( i 1 ) ] G t i 1 = ( G t 1 i ) T G t i j = cov ( z t i + 1 , z t j + 1 | x 1 : t 1 , y 1 : t 1 ) = V t 1 ( i 1 ) ( j 1 ) G t ( L + 1 ) k = cov ( x t , z t k + 1 | x 1 : t 1 , y 1 : t 1 ) = B x G t 1 k G t ( L + 1 ) k = ( G t k ( L + 1 ) ) T G t ( L + 2 ) k = cov ( y t , z t k + 1 | x 1 : t 1 , y 1 : t 1 ) = B y G t 1 k G t ( L + 2 ) k = ( G t k ( L + 2 ) ) T G t ( L + 1 ) ( L + 1 ) = cov ( x t , x t | x 1 : t 1 , y 1 : t 1 ) = B x G t 11 B x T + Σ x G t ( L + 1 ) ( L + 2 ) = cov ( x t , y t | x 1 : t 1 , y 1 : t 1 ) = B x G t 11 B y T G t ( L + 2 ) ( L + 1 ) = ( G t ( L + 2 ) ( L + 1 ) ) T G t ( L + 2 ) ( L + 2 ) = cov ( y t , y t | x 1 : t 1 , y 1 : t 1 ) = B y G t 11 B y T + Σ y ,
where i = 2 , 3 , , L ; j = 2 , 3 , , L ; k = 1 , 2 , , L , therefore, according to the knowledge of conditional probability [34] and Appendix C, the mean value and variance of the latent variable filter distribution z t , z t 1 , , z t L + 1 | x 1 : t , y 1 : t ~ N ( u t , V t ) were calculated as shown in (A16).
u t = [ g t 1 g t L ] + [ G t 1 ( L + 1 ) G t 1 ( L + 2 ) G t L ( L + 1 ) G t L ( L + 2 ) ] [ G t ( L + 1 ) ( L + 1 ) G t ( L + 1 ) ( L + 2 ) G t ( L + 2 ) ( L + 1 ) G t ( L + 2 ) ( L + 2 ) ] 1 [ x t g t L + 1 y t g t L + 2 ] , u t = [ G t 11 G t 1 L G t L 1 G t L L ] [ G t 1 ( L + 1 ) G t 1 ( L + 2 ) G t L ( L + 1 ) G t L ( L + 2 ) ] [ G t ( L + 1 ) ( L + 1 ) G t ( L + 1 ) ( L + 2 ) G t ( L + 2 ) ( L + 1 ) G t ( L + 2 ) ( L + 2 ) ] 1 [ G t 1 ( L + 1 ) G t 1 ( L + 2 ) G t L ( L + 1 ) G t L ( L + 2 ) ] T .
In the Bayesian smoothing stage, the goal was to calculate the posterior probability z t + 1 , z t , , z t L + 2 | x 1 : T , y 1 : T ~ N ( m t + 1 , M t + 1 ) of the latent variable [ z t + 1 , z t , , z t L + 2 ] with respect to the variable x 1 : T , y 1 : T at time t + 1 to calculate the posterior probability of the latent variable [ z t , z t 1 , , z t L + 1 ] with respect to the variable x 1 : T , y 1 : T at time t, as shown in (A17).
z t , z t 1 , , z t L + 1 | x 1 : T , y 1 : T ~ N ( m t , M t ) = N ( [ m t 1 m t L ] , [ M t 11 M t 1 L M t L 1 M t L L ] ) ,
where 0 t T , T, there is m T = u T , M T = V T . In order to calculate the distribution, first, the posterior distribution of the latent variable [ z t + 1 , z t , , z t L + 1 ] was calculated with respect to the variable x 1 : t , y 1 : t , as shown in (A18).
z t + 1 , z t , , z t L + 1 | x 1 : t , y 1 : t ~ N ( d t , D t ) = N ( [ d t 1 d t L + 1 ] , [ D t 11 D t 1 ( L + 1 ) D t ( L + 1 ) 1 D t ( L + 1 ) ( L + 1 ) ] ) .
The parameter calculation of (A18) is shown in (A19)–(A20).
{ d t 1 = E ( z t + 1 | x 1 : t , y 1 : t ) = A u t d t i = E ( z t i + 2 | x 1 : t , y 1 : t ) = u t i 1 D t 11 = cov ( z t + 1 , z t + 1 | x 1 : t , y 1 : t ) = A V t A T + Σ Q D t 1 i = cov ( z t + 1 , z t i + 2 | x 1 : t , y 1 : t ) = A [ V t 1 ( i 1 ) V t L ( i 1 ) ] D t i 1 = ( D t 1 i ) T D t i j = cov ( z t i + 2 , z t j + 2 | x 1 : t , y 1 : t ) = V t ( i 1 ) ( j 1 ) ,
where i = 2 ,   3 ,   ,   L + 1 ; j = 2 ,   3 ,   ,   L + 1 , and then the following distribution was calculated.
P ( z t L + 1 | z t + 1 , z t , , z t L + 2 , x 1 : T , y 1 : T ) = P ( z t L + 1 | z t + 1 , z t , , z t L + 2 , x 1 : t , y 1 : t ) = N ( r t , R t ) .
According to the knowledge of conditional probability [34] and Appendix C. The calculation of its mean and variance is shown in (A21).
r t = d t L + 1 + [ D t 1 ( L + 1 ) D t L ( L + 1 ) ] T [ D t 11 D t 11 D t L 1 D t L L ] 1 [ z t + 1 d t 1 z t L + 2 d t L ] , R t = D t ( L + 1 ) ( L + 1 ) [ D t 1 ( L + 1 ) D t L ( L + 1 ) ] T [ D t 11 D t 11 D t L 1 D t L L ] 1 [ D t 1 ( L + 1 ) D t L ( L + 1 ) ] .
The posterior probability of the latent variable [ z t + 1 , z t , , z t L + 1 ] was obtained with respect to the variable x 1 : T , y 1 : T as shown in (A22).
P ( z t + 1 , z t , , z t L + 1 | x 1 : T , y 1 : T ) = P ( z t + 1 , z t , , z t L + 2 | x 1 : T , y 1 : T ) P ( z t 1 | z t + 1 , z t , , z t L + 2 , x 1 : T , y 1 : T ) = N ( h t , H t ) = N ( [ h t 1 h t L + 1 ] , [ H t 11 H t 1 ( L + 1 ) H t ( L + 1 ) 1 H t ( L + 1 ) ( L + 1 ) ] ) .
The mean and variance of the distribution were calculated as shown in (A23).
h t = [ m t + 1 K [ m t + 1 1 d t 1 m t + 1 L d t L ] + d t L + 1 ] , H t = [ M t + 1 M t + 1 K T K M t + 1 K M t + 1 K T + R t ]   where   K = [ D t 1 ( L + 1 ) D t L ( L + 1 ) ] T [ D t 11 D t 11 D t L 1 D t L L ] 1 .
According to the Bayesian smoothing rule [31], the smooth distribution of the latent variable was obtained, as shown in (A24).
z t , z t 1 , , z t L + 1 | x 1 : T , y 1 : T ~ N ( m t , M t ) .
The calculation of its mean and variance is shown in (A25).
m t = [ h t 2 h t L + 1 ] M t = [ H t 22 H t 2 ( L + 1 ) H t ( L + 1 ) 2 H t ( L + 1 ) ( L + 1 ) ] .

Appendix C. Properties of Gaussian Distribution

Definition A1.
(Gaussian distribution) A random variable x R n has a Gaussian distribution with mean m R n and covariance P R n × n if its probability density has the form.
N ( x | m , P ) = 1 ( 2 π ) n / 2 | P | 1 / 2 exp ( 1 2 ( x m ) T P T ( x m ) ) ,
where | P | is the determinant of the matrix P .
Lemma A1.
(Joint distribution of Gaussian variables) If random variables x R n and y R m have the Gaussian probability distributions.
x ~ N ( m , P ) , y | x ~ N ( H x + u , R ) .
then the joint distribution of x , y and the marginal distribution of y are given as (A28).
( x y ) ~ N ( ( m H m + u ) , ( P P H T H P H P H T + R ) ) , y ~ N ( H m + u , H P H T + R ) .
Lemma A2.
(Conditional distribution of Gaussian variables) If the random variables x and y have the joint Gaussian probability distribution.
( x y ) ~ N ( ( a b ) , ( A C C T B ) ) .
then the marginal and conditional distributions of x and y are given as follows:
x ~ N ( a , A ) , y ~ N ( b , B ) , x | y ~ N ( a + C B 1 ( y b ) , A C B 1 C T ) , y | x ~ N ( b + C T A 1 ( x a ) , B C T A 1 C ) .

References

  1. Chen, Z.W.; Cao, Y.; Ding, S.X.; Zhang, K.; Koenings, T.; Peng, T.; Yang, C.H.; Gui, W.H. A Distributed Canonical Correlation Analysis-Based Fault Detection Method for Plant-Wide Process Monitoring. IEEE Trans. Ind. Inf. 2019, 15, 2710–2720. [Google Scholar] [CrossRef]
  2. Qin, S.J. Survey on data-driven industrial process monitoring and diagnosis. Annu. Rev. Control 2012, 36, 220–234. [Google Scholar] [CrossRef]
  3. Venkatsubramanian, V.; Rengaswamy, R.; Yin, K.; Kavuri, S.N. A review of process fault detection and diagnosis Part I: Quantitative model-based methods. Comput. Chem. Eng. 2003, 27, 293–311. [Google Scholar] [CrossRef]
  4. Venkatasubramanian, V.; Rengaswamy, R.; Kavuri, S.N. A review of process fault detection and diagnosis Part II: Quantitative model and search strategies. Comput. Chem. Eng. 2003, 27, 313–326. [Google Scholar] [CrossRef]
  5. Venkatasubramanian, V.; Rengaswamy, R.; Kavuri, S.N.; Yin, K. A review of process fault detection and diagnosis Part III: Process history based methods. Comput. Chem. Eng. 2003, 27, 327–346. [Google Scholar] [CrossRef]
  6. Chen, H.; Jiang, B.; Ding, S.; Huang, B.S. Data-driven fault diagnosis for traction systems in high-speed trains: A survey, challenges, and perspectives. IEEE Trans. Intell. Transp. Syst. 2020. aarly access. [Google Scholar] [CrossRef]
  7. Kim, D.S.; Lee, I.B. Process monitoring based on probabilistic PCA. Chemometr. Intell. Lab. Syst. 2003, 67, 109–123. [Google Scholar] [CrossRef]
  8. Kong, X.Y.; Cao, Z.H.; An, Q.S.; Gao, Y.B.; Du, B.Y. Quality-Related and Process-Related Fault Monitoring With Online Monitoring Dynamic Concurrent PLS. IEEE Access 2018, 6, 59074–59086. [Google Scholar] [CrossRef]
  9. Chen, H.; Jiang, B.; Chen, W.; Yi, H. Data-driven Detection and Diagnosis of Incipient Faults in Electrical Drives of High-Speed Trains. IEEE Trans. Ind. Electron. 2019, 66, 4716–4725. [Google Scholar] [CrossRef]
  10. Ge, Z.Q. Process Data Analytics via Probabilistic Latent Variable Models: A Tutorial Review. Ind. Eng. Chem. Res. 2018, 57, 12646–12661. [Google Scholar] [CrossRef]
  11. Si, Y.B.; Wang, Y.Q.; Zhou, D.H. Key-Performance-Indicator-Related Process Monitoring Based on Improved Kernel Partial Least Squares. IEEE Trans. Ind. Electron. 2021, 68, 2626–2636. [Google Scholar] [CrossRef]
  12. Jiang, Y.; Fan, J.L.; Chai, T.; Lewis, F.L.; Li, J.N. Tracking Control for Linear Discrete-Time Networked Control Systems With Unknown Dynamics and Dropout. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 4607–4620. [Google Scholar] [CrossRef] [PubMed]
  13. Ge, Z.Q.; Chen, X.R. Dynamic Probabilistic Latent Variable Model for Process Data Modeling and Regression Application. IEEE Trans. Control Syst. Technol. 2019, 27, 323–331. [Google Scholar] [CrossRef]
  14. Kruger, U.; Zhou, Y.Q.; Irwin, G.W. Improved principal component monitoring of large-scale processes. J. Process Control 2004, 14, 879–888. [Google Scholar] [CrossRef]
  15. Chen, J.H.; Liu, K.C. On-line batch process monitoring using dynamic PCA and dynamic PLS models. Chem. Eng. Sci. 2002, 57, 63–75. [Google Scholar] [CrossRef]
  16. Li, G.; Qin, S.J.; Zhou, D.H. A New Method of Dynamic Latent-Variable Modeling for Process Monitoring. IEEE Trans. Ind. Electron. 2014, 61, 6438–6445. [Google Scholar] [CrossRef]
  17. Ge, Z.Q.; Song, Z.H.; Gao, F.R. Review of Recent Research on Data-Based Process Monitoring. Ind. Eng. Chem. Res. 2013, 52, 3543–3562. [Google Scholar] [CrossRef]
  18. Ge, Z.Q.; Chen, X.R. Supervised linear dynamic system model for quality related fault detection in dynamic processes. J. Process Control 2016, 44, 224–235. [Google Scholar] [CrossRef]
  19. Zhurabok, A.N.; Shumsky, A.E.; Pavlov, S.V. Diagnosis of Linear Dynamic Systems by the Nonparametric Method. Autom. Remote Control 2017, 78, 1173–1188. [Google Scholar] [CrossRef]
  20. Ren, X.M.; Rad, A.B.; Chan, P.T.; Lo, W.L. Online identification of continuous-time systems with unknown time delay. IEEE Trans. Autom. Control 2005, 50, 1418–1422. [Google Scholar] [CrossRef]
  21. Yang, Z.J.; Iemura, H.; Kanae, S.; Wada, K. Identification of continuous-time systems with multiple unknown time delays by global nonlinear least-squares and instrumental variable methods. Automatica 2007, 43, 1257–1264. [Google Scholar] [CrossRef]
  22. Yalin, W.; Haibing, X.; Xiaofeng, Y. Multi-delay identification by trend-similarity analysis and its application to hydrocracking process. Chem. Eng. News 2018, 69, 1149–1157. [Google Scholar]
  23. Drakunov, S.V.; Perruquetti, W.; Richard, J.P.; Belkoura, L. Delay identification in time-delay systems using variable structure observers. Annu. Rev. Control 2006, 30, 143–158. [Google Scholar] [CrossRef] [Green Version]
  24. Shen, B.B.; Ge, Z.Q. Supervised Nonlinear Dynamic System for Soft Sensor Application Aided by Variational Auto-Encoder. IEEE Trans. Instrum. Meas. 2020, 69, 6132–6142. [Google Scholar] [CrossRef]
  25. Prosper, H.B. Deep Learning and Bayesian Methods. In Proceedings of the 12th Conference on Quark Confinement and the Hadron Spectrum, Thessaloniki, Greece, 29 August–3 September 2016. [Google Scholar]
  26. Alcala, C.F.; Qin, S.J. Reconstruction-based contribution for process monitoring. Automatica 2009, 45, 1593–1600. [Google Scholar] [CrossRef]
  27. Lataire, J.; Chen, T. Transfer function and transient estimation by Gaussian process regression in the frequency domain. Automatica 2016, 72, 217–229. [Google Scholar] [CrossRef]
  28. Panic, B.; Klemenc, J.; Nagode, M. Improved Initialization of the EM Algorithm for Mixture Model Parameter Estimation. Mathematics 2020, 8, 373. [Google Scholar] [CrossRef] [Green Version]
  29. Chen, J.Y.; Gui, W.H.; Dai, J.Y.; Jiang, Z.H.; Chen, N.; Li, X. A hybrid model combining mechanism with semi-supervised learning and its application for temperature prediction in roller hearth kiln. J. Process Control 2021, 98, 18–29. [Google Scholar] [CrossRef]
  30. Huang, X.K.; Tian, X.Y.; Zhong, Q.; He, S.W.; Huo, C.B.; Cao, Y.; Tong, Z.Q.; Li, D.C. Real-time process control of powder bed fusion by monitoring dynamic temperature field. Adv. Manuf. 2020, 8, 380–391. [Google Scholar] [CrossRef]
  31. Egorova, E.G.; Rudakova, I.V.; Rusinov, L.A.; Vorobjev, N.V. Diagnostics of sintering processes on the basis of PCA and two-level neural network model. J. Chemom. 2018, 32, 2. [Google Scholar] [CrossRef]
  32. de Waele, S.; Broersen, P.M.T. Order selection for vector autoregressive models. IEEE Trans. Signal Process. 2003, 51, 427–433. [Google Scholar] [CrossRef]
  33. Wang, Z.Y.; Liang, J. A JITL-Based Probabilistic Principal Component Analysis for Online Monitoring of Nonlinear Processes. J. Chem. Eng. Jpn. 2018, 51, 874–889. [Google Scholar] [CrossRef]
  34. Särkkä, S. Bayesian Filtering and Smoothing; Cambridge University Press: London, UK, 2013. [Google Scholar]
Figure 1. Probability graph model of dynamic autoregressive latent variable model.
Figure 1. Probability graph model of dynamic autoregressive latent variable model.
Machines 09 00229 g001
Figure 2. Flowchart based on DALM process monitor.
Figure 2. Flowchart based on DALM process monitor.
Machines 09 00229 g002
Figure 3. Structure diagram of sintering furnace for battery sintering.
Figure 3. Structure diagram of sintering furnace for battery sintering.
Machines 09 00229 g003
Figure 4. Correlation plots for the first four process variables.
Figure 4. Correlation plots for the first four process variables.
Machines 09 00229 g004
Figure 5. Monitoring performance under different time lag. (a) Monitoring result at order 2; (b) monitoring result at order 3; (c) monitoring result at order 4.
Figure 5. Monitoring performance under different time lag. (a) Monitoring result at order 2; (b) monitoring result at order 3; (c) monitoring result at order 4.
Machines 09 00229 g005
Figure 6. Monitoring results of fault 1. (a) T2 of PPLSR in fault 1; (b) T2 of DPLVM in fault 1; (c) T2 of DALM in fault 1.
Figure 6. Monitoring results of fault 1. (a) T2 of PPLSR in fault 1; (b) T2 of DPLVM in fault 1; (c) T2 of DALM in fault 1.
Machines 09 00229 g006
Figure 7. Monitoring results of fault 2. (a) T2 of PPLSR in fault 2; (b) T2 of DPLVM in fault 2; (c) T2 of DALM in fault 2.
Figure 7. Monitoring results of fault 2. (a) T2 of PPLSR in fault 2; (b) T2 of DPLVM in fault 2; (c) T2 of DALM in fault 2.
Machines 09 00229 g007
Figure 8. Monitoring results of fault 3. (a) T2 of PPLSR in fault 3; (b) T2 of DPLVM in fault 3; (c) T2 of DALM in fault 3.
Figure 8. Monitoring results of fault 3. (a) T2 of PPLSR in fault 3; (b) T2 of DPLVM in fault 3; (c) T2 of DALM in fault 3.
Machines 09 00229 g008
Table 1. Selected variables in the sintering process.
Table 1. Selected variables in the sintering process.
No.Measured Variables
1Below temperature of 1st zone
2Upper temperature of 1st zone
3Below temperature of 2nd zone
4Upper temperature of 2nd zone
5Below temperature of 3rd zone
6Upper temperature of 3rd zone
7Lithium loss coefficient
Table 2. Process data description.
Table 2. Process data description.
Date TypesData DescriptionTime Durations
NormalNormal data1st–1000th
Fault 1Normal samples in 1001st–1200th and abnormal samples of 3rd zone temperature rise in 1201st–1400th1001st–1400th
Fault 2Normal samples in 1401st–1600th and abnormal samples of 3rd zone temperature drop in 1601st–1800th1401st–1800th
Fault 3Normal samples in 1801st–2000th and abnormal samples of downtime fault in 2001st–2200th1801st–2200th
Table 3. FAR and FDR under different time lag.
Table 3. FAR and FDR under different time lag.
Time Lag2345
FAR0.1650.0500.270\
FDR0.6651.0000.905\
Table 4. RMSE under different latent variable dimensions.
Table 4. RMSE under different latent variable dimensions.
Number of Latent Variables12345
RMSE0.1230.080.0450.0420.042
Table 5. FAR and FDR of the three methods under different fault cases.
Table 5. FAR and FDR of the three methods under different fault cases.
FaultsPPLSRDPLVMDALM
FARFDRFARFDRFARFDR
Fault 10.2400.1200.2100.7950.0501.000
Fault 20.1000.3500.1550.8100.0451.000
Fault 30.3150.9800.0800.7700.0451.000
Average0.2180.4830.1480.7920.0471.000
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, N.; Hu, F.; Chen, J.; Chen, Z.; Gui, W.; Li, X. A Process Monitoring Method Based on Dynamic Autoregressive Latent Variable Model and Its Application in the Sintering Process of Ternary Cathode Materials. Machines 2021, 9, 229. https://doi.org/10.3390/machines9100229

AMA Style

Chen N, Hu F, Chen J, Chen Z, Gui W, Li X. A Process Monitoring Method Based on Dynamic Autoregressive Latent Variable Model and Its Application in the Sintering Process of Ternary Cathode Materials. Machines. 2021; 9(10):229. https://doi.org/10.3390/machines9100229

Chicago/Turabian Style

Chen, Ning, Fuhai Hu, Jiayao Chen, Zhiwen Chen, Weihua Gui, and Xu Li. 2021. "A Process Monitoring Method Based on Dynamic Autoregressive Latent Variable Model and Its Application in the Sintering Process of Ternary Cathode Materials" Machines 9, no. 10: 229. https://doi.org/10.3390/machines9100229

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop