Next Article in Journal
Systematic Encoding and Shortening of PAC Codes
Previous Article in Journal
Combining Sparse and Dense Features to Improve Multi-Modal Registration for Brain DTI Images
Previous Article in Special Issue
On Higher Order Structures in Thermodynamics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On Energy–Information Balance in Automatic Control Systems Revisited

by
Vladimir Rubtsov
1,2,3
1
Département de Mathématiques, LAREMA UMR 6093 du CNRS, Université d’Angers, 49045 Angers, France
2
Theory Division, Institute for Theoretical and Experimental Physics, 117218 Moscow, Russia
3
Institute of Geometry And Physics, 34151 Trieste, Italy
Entropy 2020, 22(11), 1300; https://doi.org/10.3390/e22111300
Submission received: 9 November 2020 / Accepted: 13 November 2020 / Published: 15 November 2020
(This article belongs to the Special Issue Thermodynamics, Geometry and Control Theory)

Abstract

:
We revise and slightly generalize some variational problems related to the “informational approach” in the classical optimization problem for automatic control systems which was popular from 1970–1990. We find extremals for various degenerated (derivative independent) functionals and propose some interpretations of obtained minimax relations. The main example of such functionals is given by the Gelfand–Pinsker–Yaglom formula for the information quantity contained in one random process in another one. We find some balance relations in a linear stationary one-dimensional system with Gaussian signal and interpret them in terms of Legendre duality.

To Nikolay Petrovich Bukanov with my gratitude and admiration

1. Introduction

An informational approach to an opimality criteria for a control system synthesis was proposed by Bukanov in [1]. He has considered the following one-dimensional, control system.
Remark 1.
The “subtraction” and “addition” notation of the diagram nodes means that f.e. the input function X ( t ) = U ( t ) P ( t ) and Z ( t ) = Y ( t ) Y 0 ( t ) while V ( t ) = Z ( t ) + N ( t ) .
His choice of an informational criterion was motivated by a straightforward engineering application: studies of the automatic board control system of the passenger jet TU-154. Mathematically, he had used the classical Gelfand–Pinsker–Yaglom functional formula [2]
J U , Z = 1 4 π + log ( 1 R U Z 2 ) d ω
for the amount of information about the one random process (the defect or the error signal Z) contained in another such process (the control signal U), where R U Z denotes the mutual correlation function for these processes.
Pinsker had re-written this functional in the following form:
J U , Z = 1 4 π + log s U s Z s U s Z | s U Z | 2 d ω .
Here, S U , S Z are spectral densities of the powers of the control signal U and the defect signal Z, S U Z -is the mutual spectral density.
Bukanov has expressed the functional variables of (1) in terms of the object-regulator transfer functions and spectral densities of the perturbations and has obtained the following explicit integral:
J U , Z = 1 4 π + log ( R 0 s P ) 2 + ( R r s N ) 2 + ( 1 + R 0 2 R r 2 ) s N s P ( 1 + R 0 2 R r 2 + 2 R 0 R r cos ( ϕ 0 + ϕ r ) ) s N s P d ω .
The transfer functions of the object and the regulator are represented by the frequency responses K 0 ( j ω ) = R 0 exp ( j ω ) and K r ( j ω ) = R r exp ( j ω ) . (We shall describe below the detailed reduction of the Pinsker Formula (1) to the Bukanov’s expression (2) in a slightly general context.)
The aim of this paper to revise and to generalize this optimal regulator problem (for a class of linear stationary systems with non-zero program ( Y 0 )) as the classical optimization problem for automatic control system of Figure 1. We get the following results:
  • find extremals for a generalization of (1) and for some other degenerated (derivative independent) “informational” functionals (the main example of such functionals is given by the Pinsker formula for the entropy quantity of one random process in another.)
  • obtained some minimax relations and propose its new interpretations as an “energy–balance equilibrium” in a spirit of the Brillouin’s and Schrodinger’s ideas of “negentropy” (we remind briefly this notion in the Discussion section).
  • write our balance relations in a linear stationary one-dimensional system with Gaussian signals and interpret them in terms of Legendre duality.
We would like to stress and warn a mathematical purist reader that this manuscript is not a “mathematical paper” in a formal sense of this notion. We (almost) never discuss and almost never precise any “existence and finitude conditions”. We would rather prefer formal manipulations whenever they exist. Therefore, we decided not to overload these small notes with probability and measure theory precise statements, technicalities, and terminology. We refer all interested readers to the book [2] for all rigorous conditions and other pure mathematical details and statements.

1.1. The Origins

The famous Shannon’s formula [3] provides a facility to measure the capacity of a communication channel in a presence not only of white noise but also in the case when the transmission is perturbed by any Gaussian noise with the power spectrum function ϵ ( ω ) proportional to the square module of a filter transfer function y ( ω ) :
ϵ ( ω ) = κ | y ( ω ) | 2 .
We recall it because it is probably the first when implicitly the informational characteristics are stipulated by some “energy” constraint. Namely, let the power of the “transmitter” is bounded by some (positive) quantity A 2 such the spectral function of the transmitted signal is s ( ω ) :
ω 0 ω 1 s ( ω ) d ω = A 2 .
The Shannon formula we have mentioned above for the total capacity of s ( ω ) reads for the frequency bandwidth [ ω 0 , ω 1 ] as
C s , ϵ = ω 0 ω 1 log ( 1 + s ( ω ) ϵ ( ω ) ) d ω .
Now, it is easy to give a mathematical answer to the natural question: what should we do to get a maximal transmission rate for the given constraint (3)? Shannon proposes to consider the variational problem with the Lagrange multiplier:
max { ω 0 ω 1 [ log ( 1 + s ( ω ) ϵ ( ω ) ) + λ s ( ω ) ] d ω }
which gives the condition
λ + 1 s ( ω ) + ϵ ( ω ) = 0
and implies that s ( ω ) + ϵ ( ω ) is a constant. Going further, we put s ( ω ) = 1 λ ϵ ( ω ) in (3) and obtain
λ = ω 1 ω 0 A 2 + B 2 ,
where B 2 = ω 0 ω 1 ϵ ( ω ) d ω and the extremal s m a x ( ω ) = A 2 + B 2 ω 1 ω 0 ϵ ( ω ) , while
C s , ϵ m a x = ( ω 1 ω 0 ) log A 2 + B 2 ω 1 ω 0 ω 0 ω 1 ϵ ( ω ) ) d ω .
In this paper, we are going to apply these beautiful observations of the founding father of Theory of Information to some (degenerated) variational problems with power constraints arising in the optimal control of automatic systems.

1.2. Review of Some Previously Known Results

The informational criteria for a control system synthesis were very popular in applied automatic control and measure systems in 1970–1980. We should mention here relevant to our interests works of academician Petrov’s school researchers [4]. They were concentrated on the questions of correctness and regularization (in sense of Tikhonov) the control and statistical optimization problems. Their most important and most interesting (to our aims ) input has concluded in a study of a connection between the regularization and a correctness of problems in a one side and an approach motivated by the above Shannon channel capacity formulae in the other. Mathematically, their main tool was based on the Wiener filter theory and related integral equations of the Kolmogorov–Wiener type (see, for example, [5]).
It seems that the system considered by N. Bukanov was ideologically similar to models elaborated by Petrov’s school researchers [5,6]. In the same time, one should stress that Bukanov’s choice of the informational criterion has some advantages. Its mathematical toolbox does not appeal to integral equations and was strongly motivated as we have mentioned by engineering applications to numerical computations related with an optimization of the jet TU-154 automatic board control system.
We shall compare both approaches—[1,5,6] for some basic examples of control linear systems.
We remind that, in [1], a solution was proposed for the variational problem in the class of linear stationary systems with zero program ( Y 0 = 0 ) and Gaussian perturbation signal functions. Namely, Formula (1) makes sense exactly in this class of random functions admissible in [2].
More precisely, the main result of [1] is that the (degenerated) variational problem for the functional (2) has the following extremals:
R r = R 0 s P s N , ϕ r = ϕ 0 ,
which are minimized at the same time as the functional of the error signal dispersion:
D Z , Z = + R 0 2 ( s P + R r 2 s N ) 1 + R 0 2 R r 2 + 2 R 0 R r cos ( ϕ 0 + ϕ r ) d ω
and the functional of the control signal power dispersion:
D U , U = + R r 2 ( s N + R 0 2 s P ) 1 + R 0 2 R r 2 + 2 R 0 R r cos ( ϕ 0 + ϕ r ) d ω .
One can conclude that the extremal relations (5) between the object–regulator characteristics guarantee that the optimal regulator provides the minimal power to the object of control. Another interesting observation made in [1] is that, in spite of the functional (2) minimal value ( J U , Z ) m i n = 0 , it does not mean that the regulator does not use the information at all—it means only that there is a certain “informational balance”:
J U , Z = [ C 0 , P , N Q 0 ] + [ C r , P , N Q r ]
between the channel capacity
C 0 , P , N = 1 4 π + log ( 1 + R 0 2 s P s N ) d ω .
of the object channel in the presence of perturbations P and N, the channel capacity
C r , P , N = 1 4 π + log ( 1 + R r 2 s N s P ) d ω .
of the regulator channel in the presence of perturbations P and N, and the (half-)differences
Q 0 = 1 2 ( Q 0 Q 0 ) , Q r = 1 2 ( Q r Q r )
of the informational performances Q 0 , Q r of the output of the object and the regulator channels with open ( Q 0 , Q r ) and closed ( Q 0 , Q r ) loop–controllers:
Q 0 = 1 4 π + log ( R 0 2 s P s N ) d ω , Q r = 1 4 π + log ( R r 2 s N s P ) d ω ,
Q 0 = 1 4 π + log ( | W Z P | 2 s P s N ) d ω , Q r = 1 4 π + log ( | W U N | 2 s N s P ) d ω .
Here, W Z P and W U N are corresponding complex frequency responses (see details in [1]).

1.3. Further Generalizations and the Subject of the Paper

Further studies of this type optimal linear control systems were concerned to a natural generalization for the systems with a non-trivial program ( Y 0 0 ). At the same time, it was an interesting and mathematically natural question to study similar (degenerated) variational problems for various informational functionals similar to (1) with a “power” constraint of type (3).
The first steps in this direction were done by the author (in a collaboration with N. Bukanov and V. Kublanov). We have reported our partial results in the paper [7] which was, in fact, a short conference proceedings announcement. No details of those results were ever published (to the best of my knowledge). It should be mentioned that some results and the intrinsic ideology—to consider an “entropy criterion” of optimal regulation proposed in our paper [7] were rediscovered in 1986 in the announcement [8].
One of the aims of this manuscript is to revise and to generalize the results reported in [7] and to consider them in a general context of variational problems with Legendre–Fenchel dual functionals.
The paper is organized as follows: Section 2 contains a reminder and a generalization of basic results from [1]. We introduce the notations and remind main definitions of all important notions, objects, and formulas used in the manuscript—among them are: the Pinsker functionals of entropy and informational rates of one stationary Gaussian signal with respect to another, signal spectral densities, and transfer functions.
First, we remind readers about (and generalize in the case of a non-trivial Gaussian stationary program signal) the variational problem for the Pinsker functional in the form proposed by Bukanov. We observe and discuss (Proposition 1) an important formula that represents the Pinsker informational rate functional associated with the linear system in Figure 1 in a Shannon-like channel capacity form.
Theorem 2 in Section 3 solves two degenerate variational problems for channel capacities functionals related with the program channel and with the error signal perturbation channel. We compare the expressions for both channel capacities with the Petrov–Uskov Formula (6) from [6]. Theorem 3 (essentially reported in [7]) gives the main balance relations for the system on Figure 1. We want to stress that the observed similarity of these relations with the famous Brilliouin and Schrodinger “negentropy” notion appeared due to my numerous discussions with Bukanov, who also indicated a close connection with the information gain, or the Kullback–Leibler divergence (see, for example [9], for a modern geometric approach to “informational thermodynamics”). In Section 4, we consider another informational functionals with or without necessary connections with optimization of linear automatic control systems. Our main observation here is Theorem 4 which remarks that the formula for the second Pinsker main functional—the entropy rate of a random process relatively to another random process—can be interpreted like a Legendre duality relation for the first Pinsker informational rate functional.

2. Optimization of the Linear Control System with a Non-Trivial Stationary Gaussian Program

The results of the section (slightly) generalize and extend the material of the paper [1] and give a new interpretation of it. It is worth noting that our notations are different from the original and our motivations and reasonings do not base on possible engineering applications and are purely theoretical.

2.1. Useful Formulas and Notations

We collect here (for the convenience of readers) all basic theoretical formulas which we shall use throughout the text. We clarify (once and forever in the paper) that all random variables and random processes (discrete or continuous) we shall understand in the same sense as in the Pinsker book [2], and we refer a “mathematically oriented” reader to this book for precise definitions, description, and properties.

2.1.1. Formulae

In what follows, we shall use the following important formulas. Let ξ , η be two random functions (= stationary (generalized) Gaussian random processes).
  • The Gelfand–Pinsker–Yaglom functional (Formula (2.58) from [10]) for the amount of information about a random function contained in another such function:
    J ξ , η = 1 4 π + log s ξ s η s ξ s η | s ξ η | 2 d ω = 1 4 π + log ( 1 R ξ η 2 ) d ω
    where s ξ ( ω ) and s η ( ω ) are respectively the spectral densities of ξ and η and s ξ η ( ω ) —the mutual spectral density of ξ and η . Here,
    R ξ η 2 : = | s ξ η | 2 s ξ s η
    is the common correlation function for the processes s ξ and s η (Formula (10.21) from [2]).
  • Gelfand–Pinsker–Yaglom functional for Gaussian ξ , η = ξ + ν such that ξ and ν are non-correlated (Formula (2.61) from [10]):
    J ξ , η = 1 4 π + log ( 1 + s ν s ξ ) d ω
  • The entropy rate of a random variable ξ with respect to η was defined by Pinsker [2]. For two one-dimensional Gaussian stationary (in the wide sense) stochastic processes ξ and η , the formula for the entropy rate (Formula (10.5.7) from [2]) reads
    H ξ , η = 1 4 π + ( s ξ s η 1 log s ξ s η ) d ω
  • For Gaussian ξ , η = ξ + ν such that ξ and ν are non-correlated:
    H ξ , η = 1 4 π + ( s ν s ξ log ( 1 + s ν s ξ ) ) d ω

2.1.2. Generalized Variational Problem

We suppose now that the stationary Gaussian program Y 0 0 . Then, the following system of linear operator equations with respect to the variables U and Z in p coordinates of the Laplace transform for our control system in Figure 1 takes place:
K r ( p ) Z ( p ) U ( p ) = K r ( p ) N ( p ) , Z ( p ) + K 0 ( p ) U ( p ) = K 0 ( p ) P ( p ) Y 0 ( p ) .
The main determinant of the system
Δ =   K r 1 1 K 0   = 1 + K 0 K r 0 .
We remind readers about frequency representations K 0 ( j ω ) : = R 0 ( ω ) exp ( j ϕ 0 ( ω ) ) and K r ( j ω ) : = R r ( ω ) exp ( j ϕ r ( ω ) )

2.1.3. Notations

For the automatic control system in Figure 1, we introduce the following notations:
  • S 0 , P , N : = R 0 2 s P + s N
  • S r , N , P : = R r 2 s N + s P
  • S 0 , r , ϕ : = ( 1 + 2 R 0 R r cos ( ϕ 0 + ϕ r ) + R 0 2 R r 2 ) s P s N
In these notations, the functional (2) J ξ , η for the processes ξ = U and η = Z is written as
J U , Z = 1 4 π + log S 0 , P , N S r , N , P S 0 , r , ϕ d ω
Solving the system (10) and calculating the spectral density of signals U , Z and denoting the program signal s Y 0 , we obtain
s Z ( ω ) = | W Z Y 0 | 2 ( j ω ) s Y 0 ( ω ) + | W Z N | 2 ( j ω ) s N ( ω ) + | W Z P | 2 ( j ω ) s P ( ω ) , s U ( ω ) = | W U Y 0 | 2 ( j ω ) s Y 0 ( ω ) + | W U N | 2 ( j ω ) s N ( ω ) + | W U P | 2 ( j ω ) s P ( ω ) .
One can find the explicit expressions for the transfer functions W ( j ω ) in Appendix B. We have computed the full functional generalizing functional (2):
J U , Z g e n = 1 4 π + log s P [ ( s Y 0 + R 0 2 S r , N , P ) ( s Y 0 + S 0 , P , N ) ] S 0 , r , ϕ ( s Y 0 + R 0 2 s P ) d ω
We see that Y 0 = 0 (13) and (11) coincide.
Theorem 1.
The functional J U , Z g e n has the following extremals:
ϕ r = ϕ r ^ : = ϕ 0 , R r = R r ^ : = R 0 s P s N [ 1 + s Y 0 R 0 2 s p ]
Proof of Theorem 1.
We denote by L the “Lagrangian” density for the functional (13):
L ( ϕ r ( ω ) , R r ( ω ) ) = log s P [ ( s Y 0 + R 0 2 S r , N , P ) ( s Y 0 + S 0 , P , N ) ] S 0 , r , ϕ ( s Y 0 + R 0 2 s P )
The “degeneracy” of (13) means that the Lagrangian density does not depend on derivatives of ϕ r and R r and the Euler–Lagrange equations are replaced by easy and straightforward computations from the system of variations
δ J U , Z g e n δ ϕ r = 0 , δ J U , Z g e n δ R r = 0
(due to the basic lemma of the calculus of variations) by the usual function minimax conditions. The first variation gives
δ J U , Z g e n δ ϕ r = 0 2 R 0 R r sin ( ϕ 0 + ϕ r ) S 0 , r , ϕ ( R 0 2 s P + s Y 0 ) = 0
and we obtain the first condition in Theorem 1. The second condition is verifying with a bit more cumbersome computation and we put it in Appendix A. □
Corollary 1.
  • Taking Y 0 = 0 immediately obtains the extremals (5) of the functional (2);
  • The extremal minimal value of the functional (13) is
    J U , Z g e n ^ = 0 ,
    but, exactly like in the case of the zero program (see (15) in [1]), it does not mean that the regulator of the control system does not use the information. It just means that there is a balance of various channel capacities generalizing (8). We shall discuss and interpret these generalizations in the next sections.
  • The same extremals (14) minimize the (generalized) functional of “power” error signal:
    D Z , Z g e n = + s P s N ( s Y 0 + R 0 2 S r , N , P ) S 0 . r , ϕ d ω .
  • This extremal value easily computed
    D Z , Z g e n ^ = ( σ Z 2 ) m i n = + s N ( s Y 0 + R 0 2 s P ) s Y 0 + S 0 , P , N d ω .
    and coincides with the minimal value of D Z , Z (14.1) in [1] for zero–program case.
  • The same is true for the similar generalization of the functional (7) for the power of the regulator. We omit the corresponding evident formulas generalizing (14.2) in [1].
We shall obtain another remarkable representation for the functional (13) using the spectral densities of signals Z and Z ^ :
Proposition 1.
J U , Z g e n = 1 4 π + log s Z s Z ^ d ω
Proof. 
We shall compute the spectral density s Z from the first equation of (12):
s Z ( ω ) = ( s Y 0 + R 0 2 S r , N , P ) s P s N S 0 , r , ϕ ,
the spectral density s Z ^ putting the extremals (14) in the above formula and substituting both spectral densities in (17). □
This formula will be very useful in further interpretations in Section 4.

3. Capacity of Channels and Balance Theorems for the Linear Stationary Control System

Now, using the generalization for Y 0 0 , we can also generalize and extend the relations which we have reported in [7]. In the same time, we shall propose some new interpretations of these relations. We shall assume throughout this section that all signals are stationary and Gaussian.

3.1. Capacity of Channels via Differential Entropy

We recall that, for the capacity of the channel for the linear control system in Figure 1, the given program Y 0 can be computed via differential entropies (“reduced to a degree of freedom”) of this program and the error (defect) Z (see, for example, (ch. IV (12.5) in [11]):
C Y 0 = H ( Y 0 ) H ( Z )
where
H ( Y 0 ) = 1 4 π + log ( 2 π e ) s Y 0 ( ω ) d ω , H ( Z ) = 1 4 π + log ( 2 π e ) s Z ( ω ) d ω
and the expression (18) is written as
C Y 0 = 1 4 π + log s Y 0 ( ω ) s Z ( ω ) d ω .
Analogously, one can obtain the capacity of the channel for the linear control system relatively the perturbations P and N:
C P , N = H ( Z ) H ( P , N )
and
H ( P , N ) = 1 4 π + log ( 2 π e ) S 0 , P , N ( ω ) d ω
such that
C P , N = 1 4 π + log s Z ( ω ) S 0 , P , N ( ω ) d ω .
Remark 2.
It would be interesting to compare Formulas (20) and (22) with Formula (16) from [6]:
C = W log s m ( ω ) s ϵ ( ω ) d ω .
for a channel capacity of the linear system in Figure 2.
Remark 3.
Here the “cross” notation of the diagram nodes means the same as the “addition” nodes notation in Figure 1.
Here, s m ( ω ) and s ϵ are spectral densities of the “useful” signal m ( t ) and the “error” signal ϵ ( t ) correspondingly.
The integral should be calculated over the total frequency band W. In particular, if u ( t ) = 0 and m and n are non-correlated, then
C = W log s m ( ω ) + s n ( ω ) s n ( ω ) d ω .
We assume that the perturbation n ( t ) has the spectral density with an additional “white noise” spectral density: s n ( ω ) = s ˜ n ( ω ) + c 2 . We have from (24)
C = W log s m ( ω ) + s ˜ n ( ω ) + c 2 s ˜ n ( ω ) + c 2 d ω .
The analog of the the balance relations (18) and (21) reads
H ( n , m ) H ( n , ϵ ) = C = 1 W W log s m ( ω ) + s ˜ n ( ω ) + c 2 s ˜ n ( ω ) + c 2 d ω .
The authors of [5] argue that the entropy quantity H ( n , ϵ ) is a characteristic of information lost while the signal goes through the system of Figure 2, and the minimum of this loss is achieved when c 0 . When c , the entropy H ( n , ϵ ) H ( n , m ) i.e., the “error” entropy tends to the entropy of the “useful” signal and the channel capacity C goes to 0 .
Remark 4.
We should stress here that:
  • all signals are supposed to be stationary random Gaussian with zero expectation value;
  • the “optimality” solution in [5] was founded with the type “energy” constraint (3), which means in the context the minimality condition of the second momentum (“dispersion”) for the error signal s ϵ :
    W s ϵ d ω = W s m ( s n ˜ + c 2 ) s m + s n ˜ + c 2 d ω = σ ϵ 2
  • The minimal value min σ ϵ 2 is achieved when c = 0 and this condition shows that the “energy” and informational criteria give the same result (compare with the similar conclusion for the case of the system in Figure 1)
  • The independency condition (the signals m ( t ) and n ( t ) are non-correlated) can be released, and this leads to the following generalization of capacity channel formula (24):
    C = W log s m ( s m + s n ) s m s n + | s m n | 2 d ω .

3.2. Variational Problems for the Channel Capacity Functionals

Now, we need to substitute the spectral densities of all signals in Formulas (20) and (22) and, as a result, obtain the following two channel capacity functionals:
C Y 0 = 1 4 π + log s Y 0 S 0 , r , ϕ s P s N ( s Y 0 + R 0 2 S r , N , P ) d ω
and
C P , N = 1 4 π + log s P s N ( s Y 0 + R 0 2 S r , N , P ) S 0 , P , N S 0 , r , ϕ d ω
Remark 5.
It is evident that the functional (28) generalizes the similar thing for the Y 0 = 0 case
C P , N | Y 0 = 0 = 1 4 π + log s P s N R 0 2 S r , N , P S 0 , P , N S 0 , r , ϕ d ω
whose extremum under conditions (5) are
C P , N m i n | Y 0 = 0 = 1 4 π + log s P s N R 0 2 S 0 , P , N 2 d ω
Remark 6.
(One can straightforwardly check that this extremal value is a sum of the informational performances ( Q 0 , Q r ) of the object and the regulator channels with closed loop–controllers outputs under the conditions (5):
C P , N m i n , 0 | Y 0 = 0 = ( Q 0 + Q r ) | ϕ = ϕ 0 , R r = R 0 s P S N .
These remarks definitely indicate that there are more general information–performance balance relations and we start to discuss them.
Theorem 2.
The functional (27) has the maximal value
C Y 0 m a x = 1 4 π + log s Y 0 s N 1 + s N s Y 0 + R 0 2 s P d ω
and the functional (28) has the minimal value
C P , N m i n = 1 4 π + log ( s Y 0 + R 0 2 s P ) s N ( s Y 0 + S 0 , P , N ) S 0 , P , N d ω
for the same set of extremals (14):
ϕ r = ϕ r ^ : = ϕ 0 , R r = R r ^ : = R 0 s P s N [ 1 + s Y 0 R 0 2 s p ]
Proof of Theorem 2.
Both functionals are degenerated in the same sense as above. Therefore, their extremals are straightforwardly reduced to their Lagrangian densities’ usual minimax computations. Similarly, to determine a character of obtained extremums, it is enough to verify the Hessian matrices definiteness for these densities instead of computations of the second variation δ 2 ( C Y 0 ) and δ 2 ( C N , P ) . We omit these routine but tedious computations which are absolutely similar to those in Theorem 1. □

3.3. Informational Balance in the Linear Control System

The following theorem was obtained in a collaboration with Bukanov and Kublanov and was announced in [7].
Theorem 3.
There are two balance relations between the channel capacities:
  • For the channel capacity of the program Y 0 and the informational rate in the control signal U in the “defect" signal Z:
    C Y 0 m a x C Y 0 = J U , Z g e n .
  • For the channel capacity of perturbations P and N and the above informational rate:
    C P , N C m i n P , N = J U , Z g e n .
Proof. 
Both identities are easily checked by manipulations with definitions of the corresponding channel capacities and computed extrema. Let us verify the first one:
C Y 0 m a x C Y 0 = 1 4 π + log s Y 0 s N 1 + s N s Y 0 + R 0 2 s P ( ) d ω 1 4 π + log s Y 0 S 0 , r , ϕ s P s N ( s Y 0 + R 0 2 S r , N , P ) d ω
= 1 4 π + log s P ( s Y 0 + R 0 2 s P + s N ) ( s Y 0 + R 0 2 S r , N , P ) S 0 , r , ϕ ( s Y 0 + R 0 2 s P ) d ω = J U , Z g e n .
 □
Remark 7.
Using (20), one can recast Formula (17) re-writing the first balance relation (33) as
J U , Z g e n = 1 4 π + log s Y 0 ( ω ) s Z ^ ( ω ) d ω 1 4 π + log s Y 0 ( ω ) s Z ( ω ) d ω = 1 4 π + log s Z ( ω ) s Z ^ ( ω ) d ω .
On other hand, it means that
J U , Z g e n = 1 4 π + log ( 2 π e ) s Z ( ω ) d ω 1 4 π + log ( 2 π e ) s Z ^ ( ω ) d ω = H ( Z ) H ( Z ^ )
which shows a similarity with the Brillouin “negentropy" principle which says that the system loses information during a transition from a state with low entropy to a state with higher entropy. (We include a brief description of this notion in the Discussion section below).
Naively speaking, the entropy H ( Z ) of the error signal of the linear system with any regulator is bigger (“more chaos") than the error signal entropy H ( Z ^ ) for the case of the system with optimal regulator:
H ( Z ^ ) = H ( Z ) J U , Z g e n 0 .
Similarly, the same result can be obtained from the second balance relation
C P , N C P , N m i n = H ( Z ) H ( P , N ) H ( Z ^ ) + H ( P , N ) = J U , Z g e n .
One can imagine here the informational rate J U , Z g e n as an “informational difference” between Z and Z ^ or as a “Brillouin–Schrödinger negentropy” [12].
Remark 8.
We should admit that the first balance relation (33) does not exist in the absence of the program signal ( Y 0 = 0 ), while the second is a generalization of the previously described informational balance relation (8): The LHS of the identity (34) under condition Y 0 = 0 is:
( C P , N C m i n P , N ) | Y 0 = 0 = 1 4 π + log s P s N R 0 2 S r , N , P S 0 , P , N S 0 , r , ϕ d ω 1 4 π + log s P s N R 0 2 S 0 , P , N 2 d ω
= 1 4 π + log S 0 , P , N S r , N , P S 0 , r , ϕ d ω = J U Z = ( J U Z g e n ) | Y 0 = 0 .
which is exactly the RHS of (34) in the same condition.

4. Entropy Rate Functional for the Linear Stationary Control System

Now, we shall discuss the entropy rate functional (9) for the pair of Gaussian stationary error signal (Z) and the error signal with optimal regulator ( Z ^ ) in the linear control system above:
H Z , Z ^ = 1 4 π + ( s Z s Z ^ 1 log s Z s Z ^ ) d ω .
Introduce the function F ( X ) : = log ( X ) 1 + 1 X of a positive argument X > 0 and study its behavior.

4.1. Properties of the Function F ( X ) and Legendre–Fenchel Transformation

Lemma 1.
The function F ( X ) has the following (almost) evident properties:
  • F ( X ) 0 and F ( 1 ) = 0 .
  • Let X = u v such that u > 0 ; v > 0 , then
    log ( u v ) ( 1 v u ) ;
  • If u = u ( ω ) and v = v ( ω ) are integrable on ( R , d μ ( ω ) ) with some measure μ ( ω ) and
    R , μ u d μ ( ω ) = R , μ v d μ ( ω )
    then
    R , μ u log ( u v ) d μ ( ω ) 0 ;
  • Let Y = 1 X and F ( Y ) : = F ( 1 X ) ; then, the function F ( Y ) = Y 1 log ( Y ) admits the Legendre–Fenchel transformation F * ( p ) = log ( 1 p ) .
Proof. 
  • To prove the first positivity property of F ( X ) , we split the half line { X 0 } in two subsets: { 0 X < 1 } and { X 1 } . Start with the second subset: for 1 t X , we have 1 t 1 X and
    log X = 1 X d t t 1 X d t X = 1 X 1 X d t = X 1 X = 1 1 X .
    In the first subset, one has similarly 1 t 1 X if X t < 1 and
    log X = X 1 d t t X 1 d t X = 1 1 X .
  • Tautological corollary of the computations in (1).
  • Using the inequality (2), we obtain:
    R , μ u log ( u v ) d μ ( ω ) R , μ u ( 1 v u ) d μ ( ω ) = 0 ;
  • The Legendre–Fenchel transformation F * ( p ) of the function F ( Y ) exists because the function is smooth and convex for Y > 0 . The only critical point Y c r i t = 1 with the critical value F ( Y c r i t ) = 0 is obtained from 1 1 Y = 0 . By the definition (see, f.e. [13], we put p : = 1 1 Y and find Y ( p ) = 1 1 p . Then, we find the Legendre-dual function from d d p ( F * ( p ) ) = 1 1 p , F * ( p ) = log ( 1 p ) . We shall check that it satisfies the duality relation
    F * ( p ) = p Y ( p ) F ( Y ( p ) ) :
    p Y ( p ) F ( Y ( p ) ) = p 1 1 p ( 1 1 p 1 log ( 1 1 p ) ) = log ( 1 p ) = F * ( p ) .
 □
We note the famous Fenchel–Young inequality which reads in this case as
F ( Y ) + F * ( p ) Y p
which gives the evident inequality
1 + log [ Y ( 1 p ) ] Y ( 1 p ) .

4.2. Legendre Transformation Analogy for the Control System Functionals

We can observe that the Pinsker entropy rate functional (9) can be written in the form
H ξ , η = 1 4 π + ( s ξ s η 1 log s ξ s η ) d ω = 1 4 π + F ( Y ( ω ) ) d ω ,
where Y ( ω ) : = s ξ ( ω ) s η ( ω ) . Then, the Legendre–Fenchel dual functional should be
1 4 π + F * ( p ( ω ) ) d ω ,
with p ( ω ) = 1 1 Y ( ω = 1 s η ( ω ) s ξ ( ω ) . Using the part 4 of the Lemma 1, one can compute
F * ( p ( ω ) ) = log ( 1 p ( ω ) ) = log ( s η ( ω ) s ξ ( ω ) ) = log ( s ξ ( ω ) s η ( ω ) ) = log Y ( ω ) .
Thus, we have obtained the following Legendre duality relation:
1 4 π + F ( Y ( ω ) ) d ω + 1 4 π + F * ( p ( ω ) ) d ω = 1 4 π + p ( ω ) Y ( ω ) d ω .
Coming back to the Pinsker entropy rate functional Formula (9), one observes that the duality implies
H ξ , η + 1 4 π + log s ξ ( ω ) s η ( ω ) d ω = 1 4 π + s ξ ( ω ) s η ( ω ) 1 d ω .
Formula (37) expressed the duality relation between the entropy rate and the functional 1 4 π + log ( s ξ ( ω ) s η ( ω ) ) d ω for any pair of stationary Gaussian signals. Let us consider the case of the Pinsker entropy rate for our control system (36), i.e., s ξ ( ω ) = s Z ( ω ) and s η ( ω ) = s Z ^ ( ω ) . In this case, (37) and (36) give us the Legendre duality condition between the functionals:
H Z , Z ^ + 1 4 π + log s Z ( ω ) s Z ^ ( ω ) d ω = 1 4 π + s Z ( ω ) s Z ^ ( ω ) 1 d ω .
Summing up the discussion in Section 4.2, one can formulate the following theorem:
Theorem 4.
The Pinsker functional of the entropy rate of the error signal in the stationary linear control system with respect to the error signal of this control system with the optimal by the informational criterion regulator and the informational rate functional for these signals are in Legendre duality:
H Z , Z ^ + J U , Z g e n = 1 4 π + s Z ( ω ) s Z ^ ( ω ) 1 d ω .
Remark 9.
We can also check that the the duality relation (37) can be easily generalized to the multidimensional case for two n dimensional stationary Gaussian processes ξ and η such that ξ = ( ξ 1 , , ξ n ) , η = ( η 1 , , η n ) and such that the pairs ( ξ j , η j ) are pair-wise uncorrelated. Then, one can define n dimensional analogues of entropy rate and informational rate functionals:
H ξ , η = 1 4 π + j , k = 1 n s ξ j ξ k ( ω ) s η j η k 1 ( ω ) n log det s ξ j ξ k ( ω ) det s η j η k ( ω ) d ω
which can be simplified (because of the mutual independence of the signals ξ and η ) to
H ξ , η = 1 4 π + j = 1 n s ξ j ξ j ( ω ) s η j η j ( ω ) 1 log s ξ j ξ j ( ω ) s η j η j ( ω ) d ω ,
where s ξ j ξ k ( ω ) (resp. s η j η k ( ω ) ) is the mutual spectral densities matrix of signal-components of ξ (resp.η) and s η j η k 1 ( ω ) denotes the ( j k ) entry of the matrix s η j η k ( ω ) .
We see that one can apply the Legendre duality Formula (37) to each summand of (40) and re-write it as
H ξ , η + L ξ , η = 1 4 π + j = 1 n s ξ j ξ j ( ω ) s η j η j ( ω ) 1 d ω ,
where the functional
L ξ , η = 1 4 π + log det s ξ j ξ k ( ω ) det s η j η k ( ω ) d ω = 1 4 π + j = 1 n log s ξ j ξ j ( ω ) s η j η j ( ω ) d ω
plays the role of the Pinsker informational rate functional (see chapters 10 and 11 in [2]) and for n = 1 coincides with 1 4 π + log ( s ξ ( ω ) s η ( ω ) ) d ω .
We want to stress that our multidimensional considerations are purely formal. We do not discuss the conditions of existence and finiteness for all formulas above and refer to [2,14] for all necessary mathematical details. We also do not know if these multidimensional generalizations can immediately apply to an optimal automatic control system problem, but we hope that such functional might be useful in comparison with the method of multidimensional optimization where the minimum of the error expectation value square and maximum of the information mean value criteria were used [15].

5. Discussion

The relationship between entropy and information was first discovered in the seminal work of Szilard [16]. Later, in the works of Brillouin [17], the negentropy principle of information was formulated, generalizing the second law of thermodynamics. According to this principle, both entropy and information should be considered jointly and cannot be interpreted separately. Brillouin had shown [18] that the definition of “information” leads to a direct connection between information and the negative of entropy (here is the abbreviation “negentropy”). Every experiment consumes negentropy (increases entropy) and yields information. The negentropy principle of information is a generalization of Carnot’s principle in thermodynamics that Brillouin had formulated in the following form: the amount of information contained in a physical system must be considered as a negative term in the total entropy of this system. If a system is isolated, it fulfills Carnot’s principle in its generalized form. According to this principle, the total entropy of the system does not decrease. Our Theorem 4 is a good illustration of these energy-information principle.
The functional L ξ , η : = 1 4 π + log s ξ ( ω ) s η ( ω ) d ω enters in the Legendre duality relation (37) with the entropy rate functional H ξ , η and has explicit interpretation in terms of the informational rate J Z , Z ^ in the case of the control system. The Legendre balance relation predicts some thermodynamical allusions going back to Gibbs thermodynamical potential relation:
S ( E ) + F ( β ) = β E ,
which relates the Helmholtz free energy F to the entropy S, total energy E, and β = 1 T is inverse to the temperature (the Boltzman constant is supposed to be 1 here). If we shall follow the initial idea of Bukanov about a “similarity” of the informational rate functional L ξ , η with the Brillouin negentropy N ( ρ ) and the information gain or Kullback–Leibler divergence I ( p , q ) = ρ log ρ d q which is “similar (up to sign) to the differential entropy expression”:
N ( ρ ) = I ( ξ ) + I ( p , q ) = ξ log ξ d q + I ( p , q ) = log ρ ξ d q
for Gaussian ξ , which has the same mean value as ρ . (notations and the statement from p. 15 of [9]), then we should conclude that, in our case, the informational rate functional J ξ , η is rather a Legendre dual to the entropy rate H ξ , η , though it plays a “role of Hamiltonian” while H ξ , η figured here as a “Lagrangian”. Following the “mechanical and thermodynamical” allusions of [9], we could interpret the results of our Theorems 1 and 2 as a (specified) example of the “principle of minimal gain information” and the balance relations in Theorem 3 show the connection with Jaynes principle of maximum entropy. It would interesting to give further interpretations in this line and to understand what the role of the “partition function” Z is, which, in our context, should naively read
L ξ , η = log s ξ s η d ω = log Z d ω .
This question, together with more serious comprehension of appropriate Lagrangian or canonical coordinates in these application approaches, would open a door to powerful invariant methods in the theory of information control systems.
There are many other possible developments to study, but all of them look more traditional. We could almost straightforwardly generalize our results using slightly more complicated cases of the capacity channel formulas in comparison with the results of Petrov et al. for signals with correlations, multidimensional systems generalizing the system in Figure 2, etc. The “physical realization” conditions are also beyond the scope of this paper.

Funding

I was partially supported during this work by the Interdisciplinary CNRS Project “Granum” (“Graded Geometry, Numerical methods, Mechanics, and Control Theory”) and Russian Foundation for Basic Research for a partial support under the Grant RFBR 18-01-00461.

Acknowledgments

It is my pleasure to dedicate this paper to Nikolay Petrovich Bukanov, who had the courage many years ago to hire me as a young researcher and the patience to introduce me in the subject of applied Information Theory. His ideas and vision were very influential and his support was crucial at some difficult moments at the start of my scientific career. I am also very grateful to V. S. Kublanov for his collaboration on [7] and to V.V. Lychagin for numerous discussions of various aspects of geometric approach to thermodynamics and the allusions from Information theory. He encouraged me to revise my old computations and to transform them in these notes. It is also my pleasure to thank O.N. Poberejnaia for numerous useful discussions in the early stages of this work. Finally, I am very grateful to the anonymous referee No. 2 for his useful remarks.

Conflicts of Interest

The author declares no conflict of interest.

Appendix A. Proof of the Theorem 1, Second Condition

δ J U , Z g e n δ R r = 1 4 π + δ δ R r log ( s Y 0 + R 0 2 s P + R 0 2 R r 2 s N ) log S 0 , r , ϕ d ω
= 1 4 π + s N R r R 0 ( s Y 0 + R 0 2 s P + R 0 2 R r 2 s N ) cos ( ϕ 0 + ϕ r ) + R 0 R r 1 + 2 R 0 R r cos ( ϕ 0 + ϕ r ) + R 0 2 R r 2 d ω .
Using the first condition of (14), one can re-write it as
1 4 π + s N R r R 0 ( s Y 0 + R 0 2 s P + R 0 2 R r 2 s N ) 1 + R 0 R r 1 + 2 R 0 R r + R 0 2 R r 2 d ω =
1 4 π + s N R r R 0 ( s Y 0 + R 0 2 s P + R 0 2 R r 2 s N ) 1 1 + R 0 R r d ω = 1 4 π + s N R r R 0 ( s Y 0 + R 0 2 s P ) ( s Y 0 + R 0 2 s P + R 0 2 R r 2 s N ) ( 1 + R 0 R r ) d ω .
Now, using the basic lemma of variational calculus, we obtain the second condition in (14):
s N R r R 0 = s Y 0 + R 0 2 s P .

Appendix B. Formulas for Transition Functions in Frequency Variables

We shall display for completeness reasons the formulas (from [1,7]) for transition functions of the linear system in Figure 1 which we have used in computations of the spectral densities s U and s Z (12):
W Z Y 0 ( j ω ) = 1 1 + R 0 R r , W Z N ( j ω ) = R 0 R r 1 + R 0 R r W Z P ( j ω ) = R 0 1 + R 0 R r , W U Y 0 ( j ω ) = R r 1 + R 0 R r W U N ( j ω ) = R r 1 + R 0 R r , W U P ( j ω ) = R 0 R r 1 + R 0 R r .

References

  1. Bukanov, N.P. Information criterion of optimality of automatic control systems. Avtomat. Telemekh. 1972, 1, 57–62. (In Russian) [Google Scholar]
  2. Pinsker, M.S. Information and Informational Stability of Random Variables and Processes; Problemy Peredači Informacii; Izdat. Akad. Nauk SSSR: Moscow, Russia, 1960; Volume 7, 203p, (In Russian). English Translation: Pinsker, M.S., Translated and edited by Amiel Feinstein; Holden-Day, Inc.: San Francisco, CA, USA; London, UK; Amsterdam, The Netherlands, 1964; 243p. [Google Scholar]
  3. Shannon, C. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 379–423, 623–656. [Google Scholar] [CrossRef] [Green Version]
  4. Petrov, B.N.; Petrov, V.V.; Ulanov, G.M.; Zaporczets, A.W.; Uskov, A.S.; Kotchubievsky, l.L. Stochastic Process in Control and Information Systems, Survey Paper 63; IFAC: Warszawa, Poland, 1969. [Google Scholar]
  5. Petrov, V.V.; Uskov, A.S. Informational aspects of the regularization problem. Doklady AN SSSR 1970, 195, 780–782. (In Russian) [Google Scholar]
  6. Petrov, V.V.; Uskov, A.S. Channel capacities of feed-back informational systems with internal perturbations. Doklady AN SSSR 1970, 194, 1029–1031. (In Russian) [Google Scholar]
  7. Bukanov, N.P.; Kublanov, V.S.; Rubtsov, V.N. About a question on a synthesis of an automatic control system optimal with respect to an informational criterion. In Proceedings of the All-Union Conference Optimization and Efficiency of Systems and Processes in Civil Aviation, MIIGA, Moscow, Russia, March 1979; pp. 83–88. (In Russian). [Google Scholar]
  8. Petrov, V.V.; Sobolev, V.I. Entropy approach to a quality criterial of automatic control systems. Doklady AN SSSR 1986, 287, 799–801. (In Russian) [Google Scholar]
  9. Lychagin, V.V. Contact Geometry, Measurement, and Thermodynamics. In Nonlinear PDEs, Their Geometry, and Applications’; Kycia, R.A., Ulan, M., Schneider, E., Eds.; Birkhäuser: Basel, Switzerland, 2019; pp. 3–52. [Google Scholar]
  10. Gelfand, I.M.; Yaglom, A.M. Computation of the amount of information about a stochastic function contained in another such function. Uspehi Mat. Nauk 1957, 12, 3–52. (In Russian) [Google Scholar]
  11. Tarasenko, F.P. Introduction in the course of Information Theory; Tomsk State University: Tomsk, Russia, 1963; p. 239. (In Russian) [Google Scholar]
  12. Schrödinger, E. What Does Life Is from the Physics Point of View? State Edition of Foreign Literature: Moscow, Russia, 1947; p. 147. (In Russian) [Google Scholar]
  13. Arnold, V.I. Mathematical Methods of Classical Mechanics; Translated by K. Vogtmann and A. Weinstein; Graduate Texts in Mathematics, 60; Springer: New York, NY, USA, 1989; 516p. [Google Scholar]
  14. Pinsker, M.S. Entropy, entropy rate and entropy stability of Gaussian random variables and processes. Doklady AN SSSR 1960, 133, 531–534. (In Russian) [Google Scholar]
  15. Petrov, V.V.; Zaporozhets, A.V. Principle of developable systems. Doklady AN SSSR 1975, 224, 779–782. (In Russian) [Google Scholar]
  16. Szilard, L. On the reduction of entropy in a thermodynamic system by the intervention of intelligent beings. Z. Phys. 1929, 53, 840–856. (In German) [Google Scholar] [CrossRef]
  17. Brillouin, L. Science and Information Theory; State Edition of Foreign Literature: Moscow, Russia, 1960; p. 392. (In Russian) [Google Scholar]
  18. Brillouin, L. Thermodynamics, statistics, and information. Am. J. Phys. 1961, 29, 318–328. [Google Scholar] [CrossRef]
Figure 1. This is a figure from [1]. Here, K 0 denotes the operator object which transform the input function X ( t ) to the output function Y ( t ) which should be compared with the given function Y 0 (the program function). The “defect” signal Z together with a measurement “noise” N forms the input of a regulator given by an operator K r . The resulting control signal U with an external perturbation P goes back to the input of the object operator K 0 .
Figure 1. This is a figure from [1]. Here, K 0 denotes the operator object which transform the input function X ( t ) to the output function Y ( t ) which should be compared with the given function Y 0 (the program function). The “defect” signal Z together with a measurement “noise” N forms the input of a regulator given by an operator K r . The resulting control signal U with an external perturbation P goes back to the input of the object operator K 0 .
Entropy 22 01300 g001
Figure 2. This is a figure from [5]. Here, W ( p ) denotes the transfer function of the operator object. The signal m ( t ) with the perturbation signal n ( t ) form a function ϕ ( t ) on the input of a regulator given by an operator W k ( p ) . The system should reproduce the “useful” signal m ( t ) as an output.
Figure 2. This is a figure from [5]. Here, W ( p ) denotes the transfer function of the operator object. The signal m ( t ) with the perturbation signal n ( t ) form a function ϕ ( t ) on the input of a regulator given by an operator W k ( p ) . The system should reproduce the “useful” signal m ( t ) as an output.
Entropy 22 01300 g002
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Rubtsov, V. On Energy–Information Balance in Automatic Control Systems Revisited. Entropy 2020, 22, 1300. https://doi.org/10.3390/e22111300

AMA Style

Rubtsov V. On Energy–Information Balance in Automatic Control Systems Revisited. Entropy. 2020; 22(11):1300. https://doi.org/10.3390/e22111300

Chicago/Turabian Style

Rubtsov, Vladimir. 2020. "On Energy–Information Balance in Automatic Control Systems Revisited" Entropy 22, no. 11: 1300. https://doi.org/10.3390/e22111300

APA Style

Rubtsov, V. (2020). On Energy–Information Balance in Automatic Control Systems Revisited. Entropy, 22(11), 1300. https://doi.org/10.3390/e22111300

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop