Next Article in Journal
Ordered Regions within a Nonlinear Time Series Solution of a Lorenz Form of the Townsend Equations for a Boundary-Layer Flow
Previous Article in Journal
Quantitative Analysis of Dynamic Behaviours of Rural Areas at Provincial Level Using Public Data of Gross Domestic Product
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Function Based Fault Detection for Uncertain Multivariate Nonlinear Non-Gaussian Stochastic Systems Using Entropy Optimization Principle

1
School of Information and Control, Nanjing University of Information Science & Technology, Ninglu Road No. 219, Pukou District, Nanjing, 210044, China
2
School of Automation Science and Electrical Engineering, Beijing University of Aeronautics & Astronautics, Xueyuan Road No. 37, Haidian District, Beijing, 100191, China
*
Author to whom correspondence should be addressed.
Entropy 2013, 15(1), 32-52; https://doi.org/10.3390/e15010032
Submission received: 26 June 2012 / Revised: 16 December 2012 / Accepted: 18 December 2012 / Published: 21 December 2012

Abstract

:
In this paper, the fault detection in uncertain multivariate nonlinear non-Gaussian stochastic systems is further investigated. Entropy is introduced to characterize the stochastic behavior of the detection errors, and the entropy optimization principle is established for the fault detection problem. The principle is to maximize the entropies of the stochastic detection errors in the presence of faults and to minimize the entropies of the detection errors in the presence of disturbances. In order to calculate the entropies, the formulations of the joint probability density functions (JPDFs) of the stochastic errors are presented in terms of the known JPDFs of both the disturbances and the faults. By using the novel performance indexes and the formulations for the entropies of the detection errors, new fault detection design methods are provided for the considered multivariate nonlinear non-Gaussian plants. Finally, a simulation example is given to illustrate the efficiency of the proposed fault detection algorithm.

1. Introduction

Fault detection (FD) has long been regarded as an important and integrated part in modern control systems for improving reliability [1,2,3,4,5,6,7,8,9,10]. The so-far obtained approaches include the system identification technique [2] and the statistic approaches based on the Bayesian theorem, likelihood methods and hypothesis test techniques [11]. Besides, filters or observers have been widely used to generate the residual signal for fault detection and estimation [2,6,8,9,12,13,14,15,16,17,18,19,20].
Among the above results, for dynamic stochastic systems, the filter-based approach has been shown as an effective way where generally the noises or disturbances are supposed to be Gaussian [21,22] and the filter design objective is to realize a minimum variance residual. However, in many chemical and manufacturing processes, the inputs involved are non-Gaussian[19,20,23,24,25,26]. For example, in paper making, the length of fibre and the size of filling molecular weight are important input variables, but none of them would obey a Gaussian distribution. This is simply because most random variables and random processes are bounded in paper web formation system. Actually, even for stochastic systems with Gaussian inputs, nonlinearities in the system may lead to non-Gaussian outputs. For the non-Gaussian variables (vectors), it is well-known that the expectation and variance are insufficient to characterize their statistics. As a result, new measures should be adopted. Entropy is a known measure that describes the average information contained in the given PDF (JPDF), which has been widely used in information, thermodynamics and control fields [6,8,27,28,29,30]. By minimizing the entropy, higher order moments can be minimized with respect to the random variables [13,26,27,31,32,33].
Nevertheless, since entropy is an integral operation of PDF while a PDF is a non-linear function with an integral constraint and a positive constraint, it is not easy to find the dynamical relationships between the PDFs of the input and the concerned output and further to formulate the entropies of the distribution of outputs. Besides, for random vectors, the formulation for the JPDF of the output becomes much more complicated than the case of random variables even if the transform matrix is linear. For example, in [6], π k + 1 [ a , b ] n is supposed to be a non-Gaussian continuous random vector, D k R m × n , after the multivariate mapping θ k + 1 = D k π k + 1 , the JPDF of θ k + 1 should be discussed for three different cases based on D k and θ k + 1 is not continuous any more except that D k is invertible. Thus, the existing concept of entropy has to be extended. In fact, the approach presented in [6] holds only for linear systems with special form of system matrices when applied to FD.
In this paper, the filter-based FD approach for uncertain multivariate nonlinear non-Gaussian stochastic system is further investigated using the entropy optimization principle. Firstly, a novel filtering and filter-based FD framework is established to construct a residual such that the fault can be detected from the changes of the residual. Secondly, the formulations between the JPDFs of the disturbances and faults and those of the detection errors are established. The entropy optimization principles are then presented to calculate the gain matrix of the optimal FD filter.
The remainder of this paper is organized as follows. Section 2 presents the preliminaries. In Section 2.1 and Section 2.2, the nonlinear difference model and the corresponding filter model are described. In Section 2.3, the definition of entropy is introduced for the detection errors and the basic relationships are formulated between the entropy of the input and the output. The main results are given in Section 3, where the entropy optimization performance index is proposed in Section 3.1 and its corresponding formulations for entropies are established in Section 3.2. In order to calculate the JPDFs of the detection errors, the simplified algorithms are provided in Section 3.3 and the FD filtering algorithms are finally given in Section 3.4 to compute the optimal FD filter gain using the proposed entropy optimization principle. A simple simulation example is provided in Section 4 to demonstrate the effectiveness of the main results. Conclusion is given in Section 5.
In the following, unless stated otherwise, matrices are assumed to have compatible dimensions. The identity and zero matrices are denoted by I and 0 respectively with appropriate dimensions. For a square matrix M, its inverse and determinant are denoted by M - 1 and det M respectively. For two real vectors v 1 and v 2 , the notation v 1 v 2 means that every element of v 1 is no less than the corresponding one of v 2 and v 1 2 is used to denote the Euclidean norm. For a multivariate nonlinear smooth function y = g ( x ) , g ( x ) x denotes its Jacobian matrix. For a random vector z , the formula P { z τ } represents the joint probability of event z τ , F z ( · ) denotes its joint probability distribution function, γ z ( · ) denotes the corresponding joint probability density function (JPDF), and H ( z ) , ε ( z ) are used to denote its entropy and expectation, respectively.

2. Preliminary

2.1. Plant Models

Consider a multivariate nonlinear stochastic discrete-time system described by
x k + 1 = F 0 ( x k , δ k , w k ) + F y k = G ( x k , v k )
where x k R n is the state, y k R m is the output, w k R p is the stochastic disturbance input, v k R m is the disturbance influencing on the output and δ k R q is the fault to be detected. F 0 ( . , . , . ) is a known multivariate Borel measurable and smooth nonlinear functions of their arguments, F represents the uncertainty satisfying F 2 δ 0 and | det F 0 ( . , . , . ) w k | 0 holds for any w k [ a , b ] p .
It should be pointed out that δ k , w k and v k are supposed to be arbitrary bounded independent random vectors rather than Gaussian ones, which is different from the existing FD approaches based on the Kalman filtering theory. It is noted that δ k can also represent the abrupt change of the model parameters. This model is actually a generalization of those where only additive faults or unexpected changes of model parameters are concerned.
To simplify the FD design procedures, the following assumptions are required, which can be satisfied by many practical processes.
Assumption 1. x 0 , δ k , v k and w k ( k = 0 , 1 , 2 , ) are bounded, mutually independent random vectors with known JPDFs denoted as γ x 0 ( τ ) , γ δ ( τ ) , γ v ( τ ) and γ w ( τ ) , respectively. γ w ( τ ) = γ w 0 ( τ ) + γ w ( τ ) where γ w 0 ( τ ) is a known function defined on [ a , b ] p and γ w ( τ ) satisfies γ w ( τ ) 2 δ 1 .
Generally speaking, two groups of approaches can be used to determine γ x 0 ( τ ) , γ δ ( τ ) , γ v ( τ ) , and γ w ( τ ) . One is the direct measurement using some advanced instruments such as the digital camera. For example, with the developments in image processing, several digital cameras have been used to measure the distribution of the flame in the flame combustion system, which can be further transformed into the temperature distribution. The other is the kernel estimation technique based on the open loop test [34]. In some practical processes (e.g., paper web formation control and particle size distribution control), enormous data can be stored and used to analyze the model of both the disturbances and the faults, with which some probabilistic properties can also be obtained, where the involved random vectors (including both the disturbances and the faults) may also obey the non-Gaussian statistic behaviour [14,15,18,32,35].
Assumption 2. F 0 ( · , · , · ) is a known multivariate B o r e l measurable and smooth nonlinear functions of their arguments, where F 0 ( 0 , 0 , 0 ) = 0 holds.
Assumption 3. G ( · , · ) = C k x k + D k v k , where C k R m × n and m n hold. Further, C k is with a full row rank at every sample time, and its first m columns also have full rank (otherwise, the column can be re-arranged to guarantee this assumption). D k R m × m is an invertible matrix.

2.2. Filter and Error Dynamics

For the nonlinear dynamic system given by (1), the filter can be described by
x ^ k + 1 = F 0 ( x ^ k , 0 , 0 ) + U k ( y k - y ^ k ) y ^ k = C k x ^ k
where U k R n × m is the filter matrix gain to be determined. Combining (1) with (2), the resulting estimation error e k = x k - x ^ k satisfies
e k + 1 = Φ ( x k , δ k , w k )
= F 0 ( x k , δ k , w k ) + F - U k y k - F 0 ( x ^ k , 0 , 0 ) + U k y ^ k
The residual signal for the fault detection is therefore defined by
e ^ k = y k - y ^ k = C k e k + D k v k
Related to (4), the vector denoted by
θ k = - U k y k - F 0 ( x ^ k , 0 , 0 ) + U k y ^ k
can be regarded as a deterministic term with an unknown matrix gain U k . The main difficulty is that the term F 0 ( x k , δ k , w k ) in (4) is both nonlinear and multivariate at each sample time. As the detection error, it is supposed that e ^ k is defined on [ α , β ] m , where α and β can also be respectively chosen as - and + . In the following, we will establish the relationships recursively between the JPDFs of x 0 , δ k , v k and w k with e ^ k .
Remark 1. In [8], e k is considered instead of e ^ k for simplicity based on a rigorous assumption, which may lead to considerable conservations. As such, in this note, e ^ k will be concerned directly for which e ^ k should be affected primarily by the fault δ k and minimally by the disturbances v k and w k .

2.3. Entropy and Its Formulation

H ( z ) can be regarded as a measure of randomness of z, which is defined as follows.
Definition 1. The entropy for a continuous random vector z is defined by
H ( z ) = - [ α , β ] l γ z ( τ ) ln γ z ( τ ) d τ γ z ( τ ) > 0 0 γ z ( τ ) = 0
where γ z ( τ ) is denoted as the JPDF of z, z [ α , β ] l .
Based on Remark 1, the main task for fault detection is to find U k such that H ( e ^ k ) can be influenced by δ k maximally and by w k or v k minimally.
It can be shown that H ( e ^ k ) , as the conditional entropies of x k - 1 , δ k , v k , w k , F and U k , is actually a functional of γ δ ( τ ) , γ v ( τ ) , γ w ( τ ) and γ x k - 1 ( τ ) as well as the underdetermined gain U k , which can be further represented by H ( e ^ k x k - 1 , δ k , w k , v k , F , U k ) . In order to calculate the entropies, the JPDFs of the errors have to be formulated in advance. The following result reveals the relationship between the JPDFs of the input and output, subject to a multivariate nonlinear mapping.
Lemma 1. For a multivariate B o r e l smooth function φ = Γ ( σ ) , if σ is a random vector with known JPDF γ σ ( τ ) , the JPDFs of φ can be given by
γ φ ( τ ) = d Ω k ( τ ) γ σ k ( ρ ) d ρ d ρ
where for a given τ , Ω k ( τ ) is the definition domain of the random vector σ with Ω k ( τ ) = { ρ Γ ( ρ ) τ } .
Proof: For a given τ, Ω k ( τ ) is the set consisting of all events ρ such that Γ ( ρ ) τ . For the concerned function, φ is also a random vector. Based on Probability theory [27], the joint distribution function of φ can be given by
F φ ( τ ) = P { φ τ } = Ω k ( τ ) γ σ k ( ρ ) d ρ
Hence, Equation (8) can be obtained by taking derivatives on both sides of (9).      Q.E.D
The next step is to consider the special relationship between e ^ k and the terms e k , v k in (5), which is related to the linear mapping φ = Θ σ , where Θ R m × n is a compatible real matrix.
In the following, Lemma 2 and Lemma 3 will be given respectively according to whether the concerned matrix Θ is invertible.
Lemma 2. If m = n and Θ is invertible, the following relationship holds
γ φ ( τ ) = γ σ Θ - 1 τ det Θ - 1
Proof: If Θ is invertible, it can be verified that for given τ, there exists τ 0 such that τ 0 = Θ - 1 τ . Thus, by using Lemma 1, we have
γ φ ( τ ) d τ = P { τ φ τ + d τ } = P { τ 0 σ τ 0 + d τ 0 } = γ σ ( τ 0 ) d τ 0 = γ σ ( Θ - 1 τ ) det Θ - 1 d τ
This process also can be regarded as a generalization of the classical relationship for random variables (see, e.g., Chapter 6, [27]).               Q.E.D
In the other case, Θ is a singular matrix. Corresponding to the structure of C k , supposing Θ is with a full row rank. In this case, there exist a low-triangle invertible matrix T 1 and an upper-triangle invertible matrix T 2 such that
T 1 Θ T 2 = I m 0
where I m represents an m - dimensional identical matrix and T 2 is invertible with positive diagonal elements.
To simplify the presentation, we denote
φ ˜ : = φ ˜ ( 1 ) φ ˜ ( 2 ) = T 1 φ , σ ˜ : = T 2 - 1 σ = σ ˜ ( 1 ) σ ˜ ( 2 ) , τ ˜ : = T 1 τ = τ ˜ ( 1 ) τ ˜ ( 2 )
with compatible dimensions. This implies
φ ˜ = I m 0 σ ˜ = σ ( 1 )
Lemma 3. If Θ R m × n is with full row rank and m < n , then for the linear mapping φ = Θ σ , the following relationship holds
γ φ ( τ ) = [ a , b ] n - m γ σ ( η ) det T 1 det T 2 d τ ˜ ( 2 )
with the auxiliary integral argument η being denoted by η : = ( T 2 T 1 τ ) T ( T 2 τ ˜ ( 2 ) ) T T , where T 1 and T 2 are denoted by (11) and τ ˜ ( 2 ) is denoted by (12).
Proof: To complete the required proof, there are three tasks to be achieved: (1) find γ σ ˜ in terms of γ σ ; (2) formulate γ φ ˜ in terms of γ σ ˜ ; (3) represent γ φ in terms of γ φ ˜ .
Based on Lemma 1, task 1 and 3 can be solved as follows. From (12), it can be shown that the JPDFs of φ ˜ and σ ˜ should satisfy
γ φ ( τ ) = γ φ ˜ T 1 τ det T 1 , γ σ ˜ ( ϵ ) = γ σ T 2 ϵ det T 2
On the other hand, (13) implies that
P φ ˜ τ ˜ ( 1 ) = P { σ ˜ ( 1 ) τ ˜ ( 1 ) , σ ˜ ( 2 ) [ a , b ] n - m }
It is noted that both τ ˜ ( 1 ) and σ ˜ ( 1 ) are m - dimensional sub-vectors as defined by (12). Thus, we have
F φ ˜ ( τ ˜ ( 1 ) ) = F σ ˜ ( 1 ) ( τ ˜ ( 1 ) ) = [ a , b ] n - m F σ ˜ ( τ ˜ ) d τ ˜ ( 2 )
which means that
γ φ ˜ ( τ ˜ ( 1 ) ) = F φ ˜ ( τ ˜ ( 1 ) ) τ ˜ ( 1 ) = F σ ˜ ( 1 ) ( τ ˜ ( 1 ) ) τ ˜ ( 1 )
= γ σ ˜ ( 1 ) ( τ ˜ ( 1 ) ) = [ a , b ] n - m γ σ ˜ ( τ ˜ ) d τ ˜ ( 2 )
As such, task 2 is completed. Combining (15) with (17) yields (14).       Q.E.D
Remark 2. In Lemma 3, if m = n , according to matrix theory, T 2 = T 1 holds. In this case, Θ = ( T 2 T 1 ) - 1 and
γ φ ( τ ) = γ σ ( η ) det T 1 det T 2 = γ σ ( T 2 T 1 τ ) det T 2 T 1 = γ σ ( Θ - 1 τ ) det Θ - 1
which implies that Lemma 3 is consistent with Lemma 2 when both the row rank and the column rank of Θ is m.

3. Main Results

3.1. Performance Indexes

Following the above procedures, the fault detection objective can be performed by judging the changes of e ^ k at every sample time, where the entropies of e ^ k only in the presence of v k and w k are minimized, but the entropies resulting from δ k are maximized. After the corresponding performance indexes are established, the involved filter gain can be designed based on the entropy optimization principle.
Generally speaking, there are two main tasks to finish in the following procedures. The first one is to provide appropriate optimization principles to achieve the FD objectives, which are represented by the entropies of e ^ k . The second one is to formulate the entropies and PDFs of e ^ k in terms of the PDFs of x 0 , v k , w k and x k , δ k , as well as Δ F .
In this subsection, we will focus on the entropy optimization principle. The JPDF of e ^ k ( e k ) in the presence of v k , w k and Δ F and in the absence of δ k can be represented by γ e ^ k 0 ( γ e k 0 ), while that in the presence of δ k , Δ F and in the absence of v k , w k can be represented by γ e ^ k 1 ( γ e k 1 ). Similarly, we denote H 0 ( e ^ k x k - 1 , v k , w k , U k , Δ F ) ( H 1 ( e ^ k x k - 1 , δ k , U k , Δ F ) ) as the entropy of e ^ k in the presence of v k , w k , Δ F and in the absence of δ k (in the presence of δ k , Δ F and in the absence of v k , w k ). Concretely it can be further denoted that for i = 0 , 1 ,
H i ( e ^ k ) = - [ α , β ] m γ e ^ k i ( τ ) ln γ e ^ k i ( τ ) d τ γ e ^ k i ( τ ) > 0 0 γ e ^ k i ( τ ) = 0
where H i ( e ^ k ) ( i = 0 , 1 ) represents the conditional entropies H 0 ( e ^ k x k - 1 , v k , w k , U k , Δ F ) and H 1 ( e ^ k x k - 1 , δ k , U k , Δ F ) respectively for brevity.
Based on (2), (4), (5) and Assumption 3, it can be claimed that only the first m rows of U k are related to H i ( e ^ k + 1 ) while the last n - m rows are irrelevant to the entropy. This means that in this case, U k , m + 1 , , U k , n in (19) are redundant vectors for the optimization, where U k , m + 1 = = U k , n = 0 can be selected under this circumstance.
Since U k is an n × m matrix, in order to use conventional optimization techniques, we denote
U k = [ U k 1 T U k n T ] T , u k = [ U k 1 U k n ] T
where U k i is the row vector of U k and thus u k R m n × 1 is a stretched column vector.
The entropy optimization performance index is proposed as follows
J N ( e ^ k ) : = k = 1 N J ( e ^ k , x k - 1 , δ k , w k , v k , u k ) = k = 1 N - H 0 ( e ^ k ) + 1 2 u k T R 1 u k + R 2 H 1 ( e ^ k )
where R 1 and R 2 are pre-specified matrix weight and constant weight respectively.
If there exists a filter such that J N ( e ^ k ) is minimized for each sample time N , then it is called as an entropy optimization FD filter.
Remark 3. It is noted that entropy optimization is corresponding to variance optimization for Gaussian signals [32]. For example, the entropy optimization principle for FD problems is in parallel to the main results in [2], where linear Gaussian systems were studied and the minimax technology was applied to the variance of errors.
Remark 4. To enhance the FD filtering performance, the expectation vector ε ( e ^ k ) can also be included in the cost function described in J N ( e ^ k ) . We can consider the more general performance indexes than J N ( e ^ k ) where the entropies of the error resulting from x 0 , v k , w k and δ k are considered simultaneously. However, noting that there is no added difference between the two performance indexes in optimization context, we will focus on J N ( e ^ k ) for the FD problem in the following subsections.
Remark 5. In the entropy optimization performance index, R 1 and R 2 are two pre-specified weights. R 1 corresponds to the energy. If there is no restriction on the energy, R 1 can be selected as 0. Generally, the smaller R 1 is, the easier the fault can be detected. R 3 can also be added as the weight of H 0 ( e ^ k ) . In that case R 2 and R 3 are relative, so we set R 3 = 1 and regulate R 2 . It is noted that the cumulative performance index is adopted in this paper. If there is no sum operation, the index corresponds to the “greedy” one [36]. However, it is shown that the use of long-range prediction can overcome the problem of stabilizing a non-minimum phase plant with unknown or variable dead-time. As such, in this paper, (20) is employed. Moreover, Entropy optimization is equivalent to the variance optimization for Gaussian signals. It can be verified that the entropy optimization is consistent with the results for linear Gaussian systems.

3.2. Formulations for the Error JPDFs

In this subsection, the formulations for the JPDFs related to the FD performance index will be discussed. The JPDFs γ e k i have to be calculated in order to formulate γ e ^ k i based on Lemmas 1–3, and then to provide the representation of H i ( e ^ k ) ( i = 0 , 1 ).
For a given τ [ α , β ] n , the domains of definition for several involved functions are defined as follows
Π k 0 : = { ( τ 1 , τ 2 ) Φ ( τ 1 , 0 , τ 2 ) τ } , Π k 1 : = { ( τ 1 , τ 2 ) Φ ( τ 1 , τ 2 , 0 ) τ }
Ξ k 0 : = { ( τ 1 , τ 2 ) F ( τ 1 , 0 , τ 2 ) τ } , Ξ k 1 : = { ( τ 1 , τ 2 ) F ( τ 1 , τ 2 , 0 ) τ }
where the notation ⪯ have been defined in the ending of the Introduction section.
To summarize, the following result can be obtained.
Theorem 1. Under Assumptions 1-3, γ e k 0 ( τ ) and γ e k 1 ( τ ) ( k = 1 , 2 , ) can be calculated from the differentials on the multiple integrations as follows:
γ e k 0 ( τ ) = d Π k 0 ( τ ) γ x k - 1 , w ( τ 1 , τ 2 ) d τ 1 d τ 2 d τ , γ e k 1 ( τ ) = d Π k 1 ( τ ) γ x k - 1 , δ ( τ 1 , τ 2 ) d τ 1 d τ 2 d τ
where
γ x k 0 ( τ ) = d Ξ k 0 ( τ ) γ x k - 1 , w ( τ 1 , τ 2 ) d τ 1 d τ 2 d τ , γ x k 1 ( τ ) = d Ξ k 1 ( τ ) γ x k - 1 , δ ( τ 1 , τ 2 ) d τ 1 d τ 2 d τ ,
and Π k i ( τ ) , Ξ k i ( τ ) ( i = 0 , 1 ) are denoted respectively by (21) and (22).
Proof: This result can be obtained similarly to Lemma 1 using the independence property in probability theory.                 Q.E.D
For the FD problem, e ^ k is used as the residual to detect the fault. In order to calculate the entropies of γ e ^ k i ( τ ) ( i = 0 , 1 ), we denote two auxiliary vectors as follows
s 1 k = C k e k , s 2 k = D k v k
and correspondingly to (11), we suppose that there exist T 1 k and T 2 k such that
T 1 k C k T 2 k = I m 0
Theorem 2. Under Assumptions 1–3, γ e ^ k i ( τ ) ( i = 0 , 1 ; k = 0 , 1 , 2 , ) can be calculated from the following
γ e ^ k i ( τ ) = [ α , β ] m γ s 1 k i ( τ - τ 1 ) γ s 2 k ( τ 1 ) d τ 1
where
γ s 1 k i ( τ ) = [ α , β ] n - m γ e k i ( η ) det T 1 k det T 2 k d τ ˜ ( 2 )
and
γ s 2 k ( τ ) = γ v D k - 1 τ det D k - 1
Proof: The formulations of γ s 1 k i ( τ ) can be obtained using Lemma 2 and Lemma 3. Noting e ^ k = s 1 k + s 2 k , the result can be provided.           Q.E.D

3.3. Simplified Calculation of the Error JPDFs

In the above results, the formulations of the entropies have been reduced to the differentials of some multiple integrals that depend on their integral domains. To simplify the design methods, the following assumption is introduced.
Assumption 4. det [ F ( x k , δ k , w k ) x k ] 0 holds and det [ Φ ( x k , δ k , w k ) x k ] 0 holds for any x k in their operation domain.
Different from Theorem 1, under Assumption 4, simplified algorithms can be further obtained to calculate γ e k i ( τ ) and then to formulate γ e ^ k i ( τ ) ( i = 0 , 1 ) by constructing the following auxiliary multivariate functions.
Under Assumption 4, for functions Φ ( · , · , · ) and F ( · , · , · ) and a given vector τ 1 , there exist the inverse functions denoted by
δ = Φ - 1 ( τ 1 , τ , 0 ) , w = Φ - 1 ( τ 1 , 0 , τ )
respectively, which satisfy
τ = Φ ( τ 1 , δ , 0 ) , τ = Φ ( τ 1 , 0 , w )
Construct the auxiliary vectors for γ e k + 1 as
η k + 1 0 = w k + 1 e k + 1 , Ψ 0 ( w k + 1 , x k ) = w k + 1 Φ ( x k , 0 , w k + 1 )
η k + 1 1 = δ k + 1 e k + 1 , Ψ 1 ( δ k + 1 , x k ) = δ k + 1 Φ ( x k , δ k + 1 , 0 )
And for γ x k + 1 , we denote
ξ k + 1 0 = w k + 1 x k + 1 , Υ 0 ( w k + 1 , x k ) = w k + 1 F ( x k , 0 , w k + 1 )
ξ k + 1 1 = δ k + 1 x k + 1 , Υ 1 ( δ k + 1 , x k ) = δ k + 1 F ( x k , δ k + 1 , 0 )
where Φ ( · , · , · ) is denoted by (3). With (25) and (24), the dynamics of estimation errors in the presence of δ k + 1 ( w k + 1 ) and in the absence of w k + 1 ( δ k + 1 ) can also be represented respectively by
η k + 1 0 = Ψ 0 ( w k + 1 , x k ) , η k + 1 1 = Ψ 1 ( δ k + 1 , x k ) , ξ k + 1 0 = Υ 0 ( w k + 1 , x k ) , ξ k + 1 1 = Υ 1 ( δ k + 1 , x k )
Using the auxiliary vectors and functions given by (25), (24), (27), (26) and (28), the following results can be obtained.
Theorem 3. Under Assumptions 1–4, γ e k 0 ( τ ) and γ e k 1 ( τ ) can be calculated by ( k = 0 , 1 , 2 , )
γ e k + 1 0 ( τ ) = [ α , β ] p γ w ( τ 1 ) γ x k 0 ( Φ - 1 ( τ 1 , 0 , τ ) ) ψ 0 ( τ 1 , τ ) d τ 1
and
γ e k + 1 1 ( τ ) = [ α , β ] q γ δ ( τ 1 ) γ x k 1 ( Φ - 1 ( τ 1 , τ , 0 ) ) ψ 1 ( τ 1 , τ ) d τ 1
where
γ x k + 1 0 ( τ ) = [ α , β ] p γ w ( τ 1 ) γ x k 0 ( F - 1 ( τ 1 , 0 , τ ) ) ψ 0 ( τ 1 , τ ) d τ 1
γ x k + 1 1 ( τ ) = [ α , β ] q γ δ ( τ 1 ) γ x k 1 ( F - 1 ( τ 1 , τ , 0 ) ) ψ 1 ( τ 1 , τ ) d τ 1
and
ψ 0 ( τ 1 , τ ) : = det F ( τ 1 , 0 , τ ) τ 1 - 1 , ψ 1 ( τ 1 , τ ) : = det F ( τ 1 , τ , 0 ) τ 1 - 1
Proof: Firstly, from (25) and (28), it can be claimed that γ e k + 1 1 ( τ ) = [ α , β ] q γ η k + 1 1 ( τ 1 , τ ) d τ 1 . It is noted that Ψ 1 ( δ k + 1 , x k ) is an one-to-one mapping from ( δ k + 1 , x k ) to ( δ k + 1 , e k + 1 ) under Assumption A.4. Thus, the PDF of ( δ k + 1 , e k + 1 ) can be represented by that of ( δ k + 1 , x k ) (see e.g., [27]). On the other hand, random vectors x k and δ k + 1 can be regarded as mutually independent ones under Assumption A.1. As such, based on the special structure described by (25) and the notations in (23), it can be verified that
γ η k + 1 1 ( τ 1 , τ ) = γ δ k + 1 x k ( τ 1 , δ ) det F ( τ 1 , τ , 0 ) τ 1 - 1 = γ δ k + 1 ( τ 1 ) γ x k ( Φ - 1 ( τ 1 , τ , 0 ) ) det F ( τ 1 , τ , 0 ) τ 1 - 1
= γ δ k + 1 ( τ 1 ) γ x k ( Φ - 1 ( τ 1 , τ , 0 ) ) ψ 1 ( τ 1 , τ )
which means that (30) holds. Similarly, by considering the auxiliary vector and function described by η k + 1 0 = Ψ 0 ( w k + 1 , x k ) , we can obtain (32). This procedure can also be used to prove (29) and (31). Q.E.D
With γ e k i ( τ ) ( i = 0 , 1 ), γ e ^ k i ( τ ) can be obtained by using Theorem 2 so that H i ( e ^ k ) can also be formulated for J N ( e ^ k ) .

3.4. Optimal FD Filter Design Strategy

Based on the above procedures, it can be claimed that the optimization is required for the performance index J N ( e ^ k ) .
It is noted that (20) leads to
J k ( e ^ k ) = J k - 1 ( e ^ k ) + Ψ ( u k ) + 1 2 u k T R 1 u k , k = 0 , 1 , 2 , , +
where
Ψ ( u k ) : = Ψ ( e ^ k , u k ) = - H 0 ( e ^ k ) + R 1 H 1 ( e ^ k )
The optimal filtering strategy can be obtained through
Ψ ( e ^ k , u k ) + 1 2 u k T R 1 u k u k = 0
where an explicit function from other arguments to u k can be further determined. The principle of optimality can thus result in the optimal FD filtering law for the whole process. To simplify the filter structure, a recursive design procedure is formulated in the following.
Denote
u k = u k - 1 + Δ u k , k = 0 , 1 , 2 , , N , , +
As a function of e k , w k and τ , it can be approximated to give
Ψ ( e ^ k , u k ) = Ψ k 0 + Ψ k 1 Δ u k + 1 2 Δ u k T Ψ k 2 Δ u k + o ( Δ u k T u k )
where
Ψ k 0 : = Ψ k ( u k ) u k = u k - 1 , Ψ k 1 : = Ψ k ( u k ) u k u k = u k - 1 , Ψ k 2 : = 2 Ψ k ( u k ) u k 2 u k = u k - 1
The recursive algorithm can be provided to determine the gain of the entropy optimization FD filter.
Theorem 4. An entropy optimization FD filtering strategy for J N subject to nonlinear error model (4) is given by
Δ u k * = - Ψ k 1 + R 1 u k - 1 Ψ k 2 + R 1
for a weight R 1 > 0 satisfying
Ψ k 2 + R 1 > 0
Proof: Firstly, it can be seen that
R 1 u k 2 = u k - 1 T R 1 u k - 1 + 2 u k - 1 T R 1 Δ u k + Δ u k T R 1 Δ u k
Substituting (39) and (42) into (37) yields recursive suboptimal control law (40) for all k = 0 , 1 , 2 , , + , under condition (41). It should be pointed out that the above algorithm given by (40) results from a necessary condition for optimization. To guarantee the sufficiency, the following second-order derivative should also be satisfied
2 Ψ k ( u k ) + 1 2 u k T R 1 u k Δ u k 2 = Ψ k 2 + R 1
which can be guaranteed if R 1 is selected sufficiently large.         Q.E.D
Remark 6. The real-time suboptimal FD filter design algorithm can be summarized as follows:
  • Initialize x 0 , x ^ 0 = ε ( x 0 ) and u 0 ;
  • At the sample time k , compute γ e k + 1 i ( τ ) , τ [ α , β ] n based on Theorem 3;
  • At the sample time k , compute γ e ^ k + 1 i ( τ ) , τ [ α , β ] m and then obtain H i ( e ^ k + 1 ) ( i = 0 , 1 ) via (18);
  • Calculate Δ u k and u k using equation (40) and (38);
  • Increase k by 1 to the next step.
The schematic diagram and flow chart are given in Figure 1 and Figure 2 to illustrate how to apply our function based fault detection method to multivariate uncertain non-Gaussian system step by step.
Figure 1. Function based fault detection.
Figure 1. Function based fault detection.
Entropy 15 00032 g001
Figure 2. Flow chart of function based fault detection.
Figure 2. Flow chart of function based fault detection.
Entropy 15 00032 g002

4. Simulation Results

To demonstrate the FD algorithm, we consider a nonlinear non-Gaussian model described as
x 1 , k + 1 x 2 , k + 1 = 1 0 0 a 0 ( k ) x 1 , k x 2 , k + sin x 1 , k cos 2 x 2 , k + 1 b 0 ( k ) δ k + 1 + 1 1 ω k + 1 + F y k = x 1 , k + x 2 , k + v k
where a 0 ( k ) = - arctan ( 1 + k ) - 1 + 0 . 5 , b 0 ( k ) = 0 . 02 ( k + 1 ) - 1 / 2 , the fault δ k and the disturbances ω k , v k ( k = 1 , 2 , ) are assumed to be mutually independent. The asymmetric PDF of δ k is defined by γ δ ( x ) = - 48 ( x 2 - x + 3 16 ) x [ 0 . 25 , 0 . 75 ] 0 x ( - , 0 . 25 ) ( 0 . 75 , + )
The PDF of the disturbance w k is supposed to have a perturbation, i.e., γ ω ( x ) = γ ω 0 ( x ) + γ ω where γ ω 0 ( x ) = - 3000 ( x 2 - 0 . 01 ) 4 x [ - 0 . 1 , 0 . 1 ] 0 x ( - , - 0 . 1 ) ( 0 . 1 , + ) and | γ ω | 0 . 01 .
The disturbance v k is a Gaussian variable with the distribution N ( 0 , 0 . 01 ) .
According to (2), the filter can be formed as follows:
x ^ 1 , k + 1 x ^ 2 , k + 1 = 1 0 0 a 0 ( k ) x ^ 1 , k x ^ 2 , k + sin x ^ 1 , k cos 2 x ^ 2 , k + u k 0 ( y k - y ^ k ) y ^ k = x ^ 1 , k + x ^ 2 , k
Thus, the estimation error system is driven by
e k + 1 = x 1 , k + sin x 1 , k a 0 ( k ) x 2 , k + cos 2 x 2 , k + 1 b 0 ( k ) δ k + 1 + 1 1 ω k + 1 - x ^ 1 , k + sin x ^ 1 , k a 0 ( k ) x ^ 2 , k + cos 2 x ^ 2 , k - u k 0 ( y k - y ^ k ) + F .
and thus
e ^ k + 1 = y k + 1 - y ^ k + 1 = x 1 , k + 1 + x 2 , k + 1 + v k + 1 - x ^ 1 , k + 1 - x ^ 2 , k + 1 = e 1 , k + 1 + e 2 , k + 1 + v k + 1
In the simulations, x 0 is set to be under uniform distribution on the interval [ 0 , 1 ] 2 and thus x ^ 0 = ε ( x 0 ) = [ 0 . 5 , 0 . 5 ] T , u 0 is set to be [ 0 , 0 ] T , the weights are set to be R 1 = 1 and R 2 = 0 . 1 . Figure 3, Figure 4 and Figure 5 show the signals for the fault and disturbances, while Figure 6, Figure 7 and Figure 8 show the corresponding pdfs of the signal. Figure 9 and Figure 10 are provided to demonstrate the residual signals e ^ k in the absence and in the presence of fault emerging respectively. When the random fault occurs, it is shown that the detection error increases significantly. Figure 11 is the 3-D mesh of the system output y, which shows a remarkable change of the pdf γ y when the fault occurs at sample time 50. Figure 12 is the optimization performance index along time t. The result shows that a satisfactory FD performance can be obtained through the error dynamics by optimizing the entropies of the detection errors.
Figure 3. The response of the fault δ.
Figure 3. The response of the fault δ.
Entropy 15 00032 g003
Figure 4. The response of the input disturbance ω k .
Figure 4. The response of the input disturbance ω k .
Entropy 15 00032 g004
Figure 5. The response of the output disturbance v k .
Figure 5. The response of the output disturbance v k .
Entropy 15 00032 g005
Figure 6. The PDF of the fault signal δ.
Figure 6. The PDF of the fault signal δ.
Entropy 15 00032 g006
Figure 7. The PDF of the input disturbance ω k .
Figure 7. The PDF of the input disturbance ω k .
Entropy 15 00032 g007
Figure 8. The PDF of the output disturbance v k .
Figure 8. The PDF of the output disturbance v k .
Entropy 15 00032 g008
Figure 9. The residual value when fault occurs.
Figure 9. The residual value when fault occurs.
Entropy 15 00032 g009
Figure 10. The residual value when no fault occurs.
Figure 10. The residual value when no fault occurs.
Entropy 15 00032 g010
Figure 11. 3-D mesh of the system output y.
Figure 11. 3-D mesh of the system output y.
Entropy 15 00032 g011
Figure 12. The optimization performance index J.
Figure 12. The optimization performance index J.
Entropy 15 00032 g012

5. Conclusions

Since the classical FD approach using the Kalman-filter theory is insufficient to apply to the stochastic systems with non-Gaussian variables, a new FD framework is proposed in this paper for dynamic multivariate nonlinear stochastic systems. The entropy optimization principle is established for the concerned nonlinear detection error system, as represented by a non-Gaussian stochastic system. The main design principle is to maximize the entropies of the residual errors when the faults occur and to minimize the entropies of the residual errors resulting from other stochastic noises. For this purpose, new relationships are provided to calculate the JPDFs of the detection error in terms of the JPDFs of both the disturbances and the faults. As such, recursive approaches can be constructed to formulate the entropies of the detection errors. Combining the formulations with the novel performance indexes based on the entropy optimization principles, the recursive algorithms are provided to calculate the gain of the optimal FD filters. The advantages derived from the proposed fault detection approach are summarized as following
  • The detected fault and system noises do not have to be Gaussian.
  • The entropy optimization principle for FD problems is in parallel to the main results to [2], where linear Gaussian systems were studied and the minimax technology was applied to the variance of errors. It has therefore generalized variance optimization for Gaussian signal.
  • The fault detection algorithm only uses the JPDF of the residual signal to calculate the filter gain matrix. There is no need to measure the system output pdf. This constitutes a major advantage compared with [37,38,39], which require the output pdf to be measurable.
  • This fault detection approach is applicable to multivariate and uncertain systems. It is a generalization of the method in [8] where only single-input-signal-output system is concerned.
However, several issues still need to be studied in the future, i.e., how to determine the threshold and how to perform fault diagnosis. Moreover, as a branch of stochastic distribution control (SDC), fault detection using entropy optimization principle still remains in the theoretic research stage [8,37,40]. Its real world application requires more research efforts.

Acknowledgment

This work is jointly supported by the National Science Foundation of China under grant 60925012,61104123,61104073,61105115 and China Postdoctoral Science Foundation funded project (2012M520141). These are gratefully acknowledged.

References

  1. Basseville, M. On-board component fault detection and isolation using the statistic local approach. Automatic 1998, 34, 1391–1415. [Google Scholar] [CrossRef]
  2. Chen, R.H.; Mingori, D.L.; Speyer, J.L. Optimal stochastic fault detection filter. Automatica 2003, 39, 377–390. [Google Scholar] [CrossRef]
  3. Chen, R.H.; Speyer, J.L. A generalized least-squares fault detection filter. Int. J. Adapt. Contr. Signal Process. 2000, 14, 747–757. [Google Scholar] [CrossRef]
  4. De Persis, C.; Isidori, A. A geometric approach to nonlinear fault detection and isolation. IEEE Trans. Automat. Contr. 2001, 46, 853–856. [Google Scholar] [CrossRef]
  5. Frank, P.M. Fault diagnosis in dynamic systems using analytical and knowledge-based redundancy a survey and some new results. Automatica 1990, 26, 459–474. [Google Scholar] [CrossRef]
  6. Guo, L.; Wang, H. Minimum entropy filtering for multivariate stochastic systems with non-Gaussian noises. IEEE Trans. Automat. Contr. 2006, 51, 695–700. [Google Scholar] [CrossRef]
  7. Shields, D.N.; Ashton, S.A.; Daley, S. Robust fault detection observers for nonlinear polynomial systems. Int. J. Syst. Sci. 2001, 32, 723–737. [Google Scholar] [CrossRef]
  8. Guo, L.; Wang, H.; Chai, T. Fault detection for non-linear non-Gaussian stochastic systems using entropy optimization principle. Trans. Inst. Meas. Contr. 2006, 28, 145–161. [Google Scholar] [CrossRef]
  9. Mao, Z.H.; Jiang, B.; Shi, P. H-infinity fault detection filter design for networked control systems modelled by discrete Markovian jump systems. IET Proc. Contr. Theor. Appl. 2007, 1, 1336–1343. [Google Scholar] [CrossRef]
  10. Jiang, B.; Staroswiecki, M.; Cocquempot, V. Fault accommodation for a class of nonlinear systems. IEEE Trans. Automat. Contr. 2006, 51, 1578–1583. [Google Scholar] [CrossRef]
  11. Basseville, M.; Nikiforv, I. Fault isolation and diagnosis: Nuisance rejection and multiple hypothesis testing. Annu. Rev. Contr. 2002, 26, 189–202. [Google Scholar] [CrossRef]
  12. Bar-Shalom, Y.; Li, X.R. Estimation and Tracking: Princples, Techniques, and Software; Artech House: Norwood, MA, USA, 1996. [Google Scholar]
  13. Zhang, J.; Cai, L.; Wang, H. Minimum entropy filtering for networked control systems via information theoretic learning approach. In Proceedings of the 2010 International Conference on Modelling, Identification and Control, Okayama, Japan, 17–19 July 2010; pp. 774–778.
  14. Orchardand, M.; Vachtsevanos, G. A particle filtering approach for on-line fault diagnosis and failure prognosis. Trans. Inst. Meas. Contr. 2009, 31, 221–246. [Google Scholar] [CrossRef]
  15. Chen, C.; Zhang, B.; Vachtsevanos, G.; Orchard, M. Machine condition prediction based on adaptive neuro-fuzzy and high-order particle filtering. IEEE Trans. Ind. Electron. 2011, 58, 4353–4364. [Google Scholar] [CrossRef]
  16. Wang, H.; Afshar, P. ILC-based fixed-structure controller design for output PDF shaping in stochastic systems using LMI techniques. IEEE Trans. Automat. Contr. 2009, 54, 760–773. [Google Scholar] [CrossRef]
  17. Zhou, J.; Zhou, D.; Wang, H.; Guo, L.; Chai, T. Distribution function tracking filter design using hybrid characteristic functions. Automatica 2010, 46, 101–109. [Google Scholar] [CrossRef]
  18. Chen, C.; Brown, D.; Sconyers, C.; Zhang, B.; Vachtsevanos, G.; Orchard, M. An integrated architecture for fault diagnosis and failure prognosis of complex engineering systems. Expert Syst. Appl. 2012, 39, 9031–9040. [Google Scholar] [CrossRef]
  19. Yin, L.; Guo, L. Fault isolation for dynamic multivariate nonlinear non-gaussian stochastic systems using generalized entropy optimization principle. Automatica 2009, 45, 2612–2619. [Google Scholar] [CrossRef]
  20. Yin, L.; Guo, L. Fault detection for NARMAX stochastic systems using entropy optimization principle. In Proceedings of the 2009 Chinese Control and Decision Conference, Guilin, China, June 2009; pp. 870–875.
  21. Guo, L.; Chen, W. Disturbance attenuation and rejection for systems with nonlinearity via DOBC approach. Int. J. Robust Nonlinear Contr. 2005, 15, 109–125. [Google Scholar] [CrossRef]
  22. Yang, F.; Wang, Z.; Hung, S. Robust kalman filtering for discrete time-varying uncertain systems with multiplicative noise. IEEE Trans. Automat. Contr. 2002, 47, 1179–1183. [Google Scholar] [CrossRef]
  23. Guo, L.; Wang, H.; Wang, A.P. Optimal probability density fuction control for NARMAX stochastic systems. Automatica 2008, 44, 1904–1911. [Google Scholar] [CrossRef]
  24. Yin, L.; Guo, L. Joint stochastic distribution tracking control for multivariate descriptor systems with non-gaussian variables. Int. J. Syst. Sci. 2012, 43, 192–200. [Google Scholar] [CrossRef]
  25. Guo, L.; Yin, L.; Wang, H. Robust PDF control with guaranteed stability for nonlinear stochastic systems under modeling errors. IET Contr. Theor. Appl. 2009, 3, 575–582. [Google Scholar] [CrossRef]
  26. Yin, L.; Guo, L.; Wang, H. Robust minimum entropy tracking control with guaranteed stability for nonlinear stochastic systems under modeling errors. In Proceedings of the 10th International Conference on Control, Automation, Robotics and Vision, Hanoi, Vietnam, 17–20 December 2008; pp. 1469–1474.
  27. Papoulis, A. Probablity, Random Variables and Stochastic Processes, 3rd ed.; McGraw-Hill: New York, NY, USA, 1991. [Google Scholar]
  28. Patton, R.J.; Frank, P.; Clark, R. Fault Diagnosis in Dynamic Systems: Theory and Application; Prentice Hall: Englewood Cliff, NJ, USA, 1989. [Google Scholar]
  29. Feng, X.B.; Loparo, K.A. Active probing for information in control system with quantized state measurements: A minimum entropy approach. IEEE Trans. Automat. Contr. 1997, 42, 216–238. [Google Scholar] [CrossRef]
  30. Wang, H. Minimum entropy control of non-Gaussian dynamic stochastic systems. IEEE Trans. Automat. Contr. 2002, 47, 398–403. [Google Scholar] [CrossRef]
  31. Saridis, G.N. Entropy formulation of optimal and adaptive control. IEEE Trans. Automat. Contr. 1988, 33, 713–720. [Google Scholar] [CrossRef]
  32. Wang, H. Bounded Dynamic Stochastc Systems: Modelling and Control; Springer-Verlag: London, UK, 2000. [Google Scholar]
  33. Zhang, J.; Du, L.; Ren, M.; Hou, G. Minimum error entropy filter for fault detection of networked control systems. Entropy 2012, 14, 505–516. [Google Scholar] [CrossRef]
  34. Silverman, B.W. Density Estimation for Statistics and Data Analysis; Chapman and Hall: London, UK, 1986. [Google Scholar]
  35. Wang, H.; Lin, W. Applying observer based FDI techniques to detect faults in dynamic and bounded stochastc distributions. Int. J. Contr. 2000, 73, 1424–1436. [Google Scholar] [CrossRef]
  36. Yue, H.; Wang, H. Minimum entropy control of closed-loop tracking errors for dynamic stochastic systems. IEEE Trans. Automat. Contr. 2003, 48, 118–121. [Google Scholar]
  37. Guo, L.; Zhang, Y.M.; Wang, H.; Fang, J.C. Observer-based optimal fault detection and diagnosis using conditional probability distributions. IEEE Trans. Signal Process. 2006, 54, 3712–3719. [Google Scholar] [CrossRef]
  38. Zhang, Y.M.; Guo, L.; Wang, H. Filter-based fault detection and diagnosis using output PDFs for stochastic systems with time delays. Int. J. Adapt. Contr. Signal Process. 2006, 20, 175–194. [Google Scholar] [CrossRef]
  39. Li, T.; Guo, L.; Wu, L.Y. Observer-based optimal fault detection using PDFs for time-delay stochastic systems. Nonlinear Anal. R. World Appl. 2008, 9, 2337–2349. [Google Scholar] [CrossRef]
  40. Guo, L.; Yin, L.; Wang, H.; Chai, T.Y. Entropy optimization filtering for fault isolation of nonlinear non-gaussian stochastic systems. IEEE Trans. Automat. Contr. 2009, 54, 804–810. [Google Scholar]

Share and Cite

MDPI and ACS Style

Yin, L.; Zhou, L. Function Based Fault Detection for Uncertain Multivariate Nonlinear Non-Gaussian Stochastic Systems Using Entropy Optimization Principle. Entropy 2013, 15, 32-52. https://doi.org/10.3390/e15010032

AMA Style

Yin L, Zhou L. Function Based Fault Detection for Uncertain Multivariate Nonlinear Non-Gaussian Stochastic Systems Using Entropy Optimization Principle. Entropy. 2013; 15(1):32-52. https://doi.org/10.3390/e15010032

Chicago/Turabian Style

Yin, Liping, and Li Zhou. 2013. "Function Based Fault Detection for Uncertain Multivariate Nonlinear Non-Gaussian Stochastic Systems Using Entropy Optimization Principle" Entropy 15, no. 1: 32-52. https://doi.org/10.3390/e15010032

APA Style

Yin, L., & Zhou, L. (2013). Function Based Fault Detection for Uncertain Multivariate Nonlinear Non-Gaussian Stochastic Systems Using Entropy Optimization Principle. Entropy, 15(1), 32-52. https://doi.org/10.3390/e15010032

Article Metrics

Back to TopTop