Next Article in Journal
Distributing Secret Keys with Quantum Continuous Variables: Principle, Security and Implementations
Next Article in Special Issue
Quantum Secure Direct Communication Based on Dense Coding and Detecting Eavesdropping with Four-Particle Genuine Entangled State
Previous Article in Journal
The Effect of a Long-Range Correlated-Hopping Interaction on Bariev Spin Chains
Previous Article in Special Issue
Minimal Rényi–Ingarden–Urbanik Entropy of Multipartite Quantum States
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Self-Similar Solutions of Rényi’s Entropy and the Concavity of Its Entropy Power

by
Agapitos N. Hatzinikitas
1,2
1
Department of Mathematics, School of Sciences, University of Aegean, Karlovasi, Samos 83200, Greece
2
School of Physics and Astronomy, University of Leeds, Leeds LS2 9JT, UK 
Entropy 2015, 17(9), 6056-6071; https://doi.org/10.3390/e17096056
Submission received: 8 July 2015 / Revised: 26 August 2015 / Accepted: 27 August 2015 / Published: 31 August 2015
(This article belongs to the Special Issue Quantum Computation and Information: Multi-Particle Aspects)

Abstract

:
We study the class of self-similar probability density functions with finite mean and variance, which maximize Rényi’s entropy. The investigation is restricted in the Schwartz space S ( R d ) and in the space of l-differentiable compactly supported functions C c l ( R d ) . Interestingly, the solutions of this optimization problem do not coincide with the solutions of the usual porous medium equation with a Dirac point source, as occurs in the optimization of Shannon’s entropy. We also study the concavity of the entropy power in R d with respect to time using two different methods. The first one takes advantage of the solutions determined earlier, while the second one is based on a setting that could be used for Riemannian manifolds.

1. Introduction

The last two decades have witnessed an enormous growing interest in using information concepts in diverse fields of science. Although Rényi entropy was introduced as early as 1961, only recently, a wide range of applications has emerged as in the analysis of quantum entanglement [1], quantum correlations [2], computer vision [3], clustering [4], quantum cryptography [5] and pattern recognition [6,7].
Our contribution in this article is three-fold: First, we rigorously establish the solutions of two maximization problems ((12) and (33)) using self-similar probability density functions (pdfs) in the Schwartz space and the space of compactly-supported continuous functions on R d , by successfully applying the method of calculus of variations. Non-trivial mathematical technicalities, not addressed in previous maximization problems, are resolved, providing in this way a crystal clear presentation of the method, which was lacking from the literature. In particular, the perturbed pdfs we consider have the form g ϵ = f + ϵ h , where | ϵ | < ϵ 0 < 1 and h are functions, which are chosen in such a way that g ϵ is a pdf and has the same variance as f. The vanishing of the first variation of the Lagrange functional provides us with an integral equation, which determines the critical points. It turns out that the solutions are unique. The non-negativity of the second variation leads to an integral inequality, which is preserved by the admissible perturbations we consider. Therefore, the critical point is a local maximum of the functional. To prove its global nature, we use the concept of the relative Rényi entropy and examine its positivity at the critical point. This procedure can be generalized on R d straightforwardly. The solutions of the first optimization problem (12) have been derived in the past by [8], but following a completely different route. The second optimization problem (33), with the additional feature of non-zero mean, is proven to be equivalent to the first one by performing a suitable transformation to the random variable. Its solutions are new.
Second, we compare the specified solutions with those of Zel’dovich, Kompaneets and Barenblatt (ZKB) [9,10,11]. The difference is in the diffusion coefficient, which depends not only on the shape and the size of a molecule, but also on the order α of the Rényi entropy and the dimension of the space. We plot both our solution and the Barenblatt’s solutions and observe that their L norms obey u f , d d + 2 < α α th . ( d ) , while for values of α greater than the threshold α th . ( d ) , the inequality changes direction.
Third, the problem of concavity of the entropy power in R d is confronted by utilising two different methods. In the first method, the solutions of the first problem guarantee concavity on the condition that the second time derivative of Rényi’s entropy fulfils Inequality (49). The second method is closer to the spirit analysed in [12,13], where the α = 1 and α 1 cases were studied, respectively. The advantage of our approach stems from the simplicity of the arguments used in the proof and the discovery of the non-trivial Constraint (61), which is justified by self-consistency.
The paper is organized as follows. Section 2 reviews and proves some fundamental properties of the entropy. Section 3 and Section 4 determine the solutions of the two maximization problems using the method of calculus of variations and examines their global validity using the concept of relative Rényi entropy. Section 5 constructs the nonlinear diffusion equation corresponding to the solutions of the two maximization problems and compares it to the usual one. Section 6 proves the concavity of the Rényi entropy with respect to time following two different methods. Section 7 concludes the work and comments on more general constraints one could have considered.

2. Preliminaries

In this section, we briefly review some properties of the Rényi entropy, and for the sake of clarity, we present the corresponding proofs.
Definition 2.1. Let ( Ω , A , μ ) be a probability space and an A -measurable function f : Ω R + be a probability density function (pdf). The differential Rényi entropy of order α, α R + , is the nonlinear functional H α : f R defined by:
H α [ f ] : = 1 1 - α ln Ω d F d μ α - 1 d F
where F is the probability measure induced by f, namely, F ( E ) = E f ( x ) d μ ( x ) , E A .
Other equivalent ways of defining the Rényi entropy are: H α [ f ] = 1 1 - α ln E X ( f α - 1 ( X ) ) , α or H α [ f ] = 1 1 - α ln f L α α , α > 1 .
Properties:
(a)
H α [ f ] is a continuous ( α 1 ) and strictly decreasing function in α unless f is the uniform density, in which case it is constant.
Proof: By Hölder’s inequality, there is a family of relations:
Ω f ( 1 - θ ) p + θ q d μ Ω f p d μ 1 - θ Ω f q d μ θ
holding whenever p , q 0 and θ [ 0 , 1 ] . Taking p = 1 and assuming f to be a pdf, we have:
Ω f 1 - θ + θ q d μ Ω f q d μ θ .
Let 0 < α < β < 1 and q = α . Then, for θ = ( 1 - β ) / ( 1 - α ) < 1 , the previous inequality becomes:
Ω f β d μ Ω f α d μ 1 - β 1 - α
which, using that the logarithmic function is an increasing function, implies that H β [ f ] H α [ f ] . The same proof holds for α , β > 1 .
(b)
H α ( f ) as a function of α converges to the following limits:
lim α 0 + H α [ f ] = ln ( μ ( Ω ) )
lim α 1 H α ( f ) = H 1 [ f ] = - Ω f ln f d μ
lim α H α [ f ] = f = inf { 0 < λ < : f λ a . e . ω Ω } ,
where H 1 [ f ] is the Shannon entropy of f.
(c)
If the L 1 ( Ω , R + ) norm is invariant under the homogeneous dilations:
x x ˜ = λ γ x , λ > 0 , x Ω R d
f ( x ) f λ ( x ) = λ δ f ( x ˜ )
then H α , for α < 1 , scales as:
H α [ f λ ] = H α [ f ] - δ ln λ , δ = d γ .
Proof 
The L 1 norm invariance of f, namely f L 1 ( Ω ) = f L 1 ( Ω ˜ ) = M , implies the condition δ = d γ , which, combined with the definition of Rényi’s entropy, produces the desired result.

3. Formulation of the First Problem and Its Solutions

In what follows, we restrict on the probability space ( R , B , d x ) , where B = σ ( O ) is the sigma algebra on open sets and d x the Lebesgue measure on R . The domain of the Rényi functional, D ( H α ) , is defined to be:
D ( H α < 1 ) = f S ( R ) : R f ( x ) d x = 1
D ( H α > 1 ) = f + C c k ( R ) , k Z + and k 1 α - 1 : R f + ( x ) d x = 1 ,
where f is a positive and integrable real-valued function, f + ( x ) = f ( x ) χ B ( x ) with χ the indicator function of the set B, and [ · ] is the integer part of the number.
The first entropy maximization problem with vanishing mean and finite variance is formulated as:
m a x . H α [ f ] f D ( H α < 1 ) subjected to the constraint E ( X 2 ) = R x 2 f ( x ) d x = μ 2 2 < .
Using the method of Lagrange multipliers, we construct the functional:
F ( f ; λ 0 , λ 2 ) = H α [ f ] - k { 0 , 2 } λ k E ( X k ) - μ k 2 , μ 0 = 1
and impose appropriate conditions on the perturbations in order to calculate its first and second variation.
Definition 3.1. A perturbation h S ( R ) is called admissible if it satisfies the following conditions:
| h ( x ) | < c | f ( x ) | , 0 < c < 1 , x R , R h ( x ) d x = 0 and R x 2 h ( x ) d x = 0 .
If we introduce the usual inner product in S ( R ) , the previous integral conditions imply that we search for a class of functions, which are orthogonal to the unity and x 2 . Odd functions in the Schwartz space, such as Q ( x ) e - b | x | μ , b , μ > 0 , with Q ( x ) a polynomial of odd powers of x, satisfy these criteria.
Expanding the Lagrange functional in a Taylor series up to second order in ϵ, we obtain:
F ( f + ϵ h ; λ 0 , λ 2 ) = F ( f ; λ 0 , λ 2 ) + δ F f ( h ) ϵ + δ 2 F f ( h ) ϵ 2 + o ( ϵ 2 ) .
The first-order necessary condition for optimality requires [14]:
δ F f ^ ( h ) = d d ϵ F ( f + ϵ h ; λ 0 , λ 2 ) | ϵ = 0 = 0 , | ϵ | < ϵ 0 < 1
or, equivalently,
R f α - 1 ( x ) - k { 0 , 2 } λ ˜ k x k h ( x ) d x = 0 , where λ ˜ k = λ k g ( α ) = λ k ( 1 - α ) α R f α ( x ) d x .
The function f ( x ) is determined by using the following lemma.
Lemma 3.2.
If f S ( R ) and if:
R f ( x ) g ( x ) d x = 0 , g S ( R ) such that lim | x | g ( x ) = 0
then f ( x ) = 0 , x R .
Proof 
Suppose that f ( x ) 0 , then there exists ξ R , such that f ( ξ ) = c > 0 (assuming that the constant is positive). Since f S ( R ) , there exists a neighbourhood ( a , b ) of ξ in which f ( x ) > 0 , x ( a , b ) . Define the function:
g ( x ) = ( x - a ) ( b - x ) e - x 2 χ ( a , b ) ( x ) .
The function g ( x ) is positive in ( a , b ) with g ( a ) = g ( b ) = 0 . However:
R f ( x ) g ( x ) d x = a b f ( x ) ( x - a ) ( b - x ) e - x 2 d x > 0
since the integrand is positive (except at a and b). This contradiction proves the lemma.
Therefore, f ( x ) is given by:
f ^ ( x ) = k { 0 , 2 } λ ˜ k x k - 1 1 - α = f ^ 1 + λ ˜ 2 λ ˜ 0 x 2 - 1 1 - α ,
where the L norm is with respect to x and λ ˜ k > 0 , since f ^ D α < 1 . In order for f ^ to be a local maximum, the following second-order necessary condition for optimality should also hold:
δ 2 F | f ^ ( h ) 0 or d 2 d ϵ 2 F ( f + ϵ h ; λ 0 , λ 2 ) | ϵ = 0 , f = f ^ 0 .
In other words, the second variation of F at f ^ should be positive semidefinite on the space of admissible perturbations h. The previous inequality is translated into:
α α - 1 R f ^ α - 1 ( x ) h ( x ) d x R f ^ α ( x ) d x R f ^ α - 2 ( x ) h 2 ( x ) d x .
It is easily checked that the solution f ^ together with the admissible perturbations satisfy the strict inequality; therefore, f ^ is a strict one-parameter family of local maxima with increasing Rényi entropy, as Property (a) of Section 2 guarantees.
Remark If 0 < R f α ( x ) d x 2 , then Taylor expanding the logarithmic function around unity, the Rényi entropy reduces to the Havrda-Charvát entropy [15], also called Tsallis entropy [16],
S α ( f ) = 1 α - 1 1 - R f α ( x ) d x .
The solution in this case is the previous one, but with the substitution λ ˜ k = λ k ( 1 - α ) α .
Depending on the space of functions, we distinguish the following two types of solutions.
(3α)
The S ( R ) solution ( α < 1 ). The requirement that f ^ is a pdf and the second moment constraint lead to the relation:
λ ˜ 0 λ ˜ 2 B 3 2 B 1 2 = λ 0 λ 2 1 - α 3 α - 1 = μ 2 2 for 1 3 < α < 1 .
We adopt the abbreviation B s / 2 B ( s / 2 , 1 / ( 1 - α ) - s / 2 ) for Euler’s beta function defined by B ( x , y ) = 0 1 t x - 1 ( 1 - t ) y - 1 d t for R e ( x ) , R e ( y ) > 0 . Using (25), the exact solution can be expressed as:
f ^ ( x ) = 1 - α 3 α - 1 1 2 1 μ 2 B ( 1 2 , 1 1 - α - 1 2 ) 1 + 1 - α 3 α - 1 x 2 μ 2 2 - 1 1 - α , 1 3 < α < 1 .
This one-parameter family of local maxima of H α is unique, and it remains to prove that it is actually also a one-parameter family of global maxima in D ( H α < 1 ) . For this, we use the notion of the relative α-Rényi entropy of two densities f ^ and g, defined in [17] by:
D α ( g | | f ^ ) = 1 1 - α ln R f ^ α - 1 ( x ) g ( x ) d x + ( 1 - α ) α H α [ f ^ ] - 1 α H α [ g ] ,
where g satisfies the same second moment constraint as f ^ . The first term in the right-hand side of (27) equals H α [ f ^ ] , as one may check, since R f ^ α - 1 ( x ) g ( x ) d x = R f ^ α ( x ) d x . Therefore:
D α ( g | | f ^ ) = 1 α ( H α [ f ^ ] - H α [ g ] ) 0
by applying Hölder’s inequality to the functions ( f ^ α - 1 g ) α and f ^ α ( α - 1 ) . The same result holds in the α > 1 case.
Proposition 3.3.
The optimization problem for α < 1 has a one-parameter family of global maxima f ^ , which in the α 1 - limit converges to the global maximum of the Shannon entropy, which is the normal distribution N ( 0 , μ 2 ) .
Proof 
Relation (26) in the α 1 - limit gives:
lim α 1 - λ ˜ 0 - 1 1 - α ( α ) 1 + 1 - α 3 α - 1 x 2 μ 2 2 - 1 1 - α = 1 2 π μ 2 2 lim ρ 0 + 1 + x 2 2 μ 2 2 ρ - 1 ρ = 1 2 π μ 2 2 e - x 2 2 μ 2 2 ,
where the change of variable ρ = ( 3 α - 1 ) / 2 ( 1 - α ) has been performed.
(3β)
The l-differentiable compactly-supported solution ( α > 1 ). In this case, following steps similar to the previous one, the solution turns out to be:
f ^ ( x ) = λ ˜ 0 1 α - 1 1 - α - 1 3 α - 1 x 2 μ 2 2 + 1 α - 1 , α > 1 and x + m a x { 0 , x } x θ ( x ) ,
where θ ( x ) is the unit-step function and the arbitrary constant λ ˜ 0 is specified as usual by imposing the requirement that f ^ be a pdf:
λ 0 ˜ ( α ) = α - 1 3 α - 1 1 2 1 μ 2 B ( 1 2 , 1 α - 1 + 1 ) α - 1 .
Remarks The derivations of ( 3 α ) and ( 3 β ) are based on the integral Formulas (3.241.4) and (3.251.1) of [18], which are connected through the transformation u = x 2 / ( x 2 + 1 ) with u ( 0 , 1 ) ; see (A.2). The solution ( 3 α ) can also be derived by integrating Equation (17), in which case, we obtain the relation between the Lagrange multipliers:
λ 0 = α 1 - α - μ 2 2 λ 2 .
Eliminating λ 0 from the solution and applying the constraints, we arrive at (26).
The previous setup can be generalized to include the case of a covariance matrix C-constraint. The new solution can be derived from the old one by replacing x 2 by x T C - 1 x ; see [8].

4. Formulation of the Second Problem and Its Solutions

The second entropy maximization problem, with non-vanishing mean and finite variance, can equivalently be restated as:
m a x . H α [ f Y ] f Y D ( H α < 1 ) subjected   to   the   new   constraints E ( Y ) = R y f Y ( y ) d y = 0 , Var ( Y ) = E ( Y 2 ) = 1 .
The random variables X , Y are related through the transformation X = μ 2 Y + μ 1 , and their corresponding pdfs f X , f Y satisfy f X ( x ) = f Y ( ( x - μ 1 ) / μ 2 ) / μ 2 . As a consequence, the entropies are connected through:
H α [ f X ] = H α [ f Y ] + ln μ 2 .
The solutions of the second problem (33) are therefore given by the solutions of the first problem (12), with the substitution ( x - μ 1 ) / μ 2 x . The first-order necessary optimization condition dictates the solution:
f ^ α - 1 ( x ) = k = 0 2 λ ˜ k x k ,
which can also be proven to be a global maximum.
We distinguish the following two classes of solutions.
(4α)
The S ( R ) solution ( α < 1 ). The positivity of the solution f ^ α - 1 ( x ) , x R requires λ ˜ 2 > 0 and λ ˜ 2 λ ˜ 0 - ( λ ˜ 1 ) 2 > 0 . The pdf, the mean value and the variance constraints lead to the conditions:
μ 1 = - λ 1 λ 2 , λ 0 λ 2 = ( 2 n - 3 ) ! ! ( 2 n - 5 ) ! ! μ 2 2 + μ 1 2 , n = 1 1 - α Z + .
Finally, the solution is written as:
f ^ ( x ) = ( 2 n - 2 ) ! ! ( 2 n - 3 ) ! ! ( 2 n - 5 ) ! ! ( 2 n - 3 ) ! ! 1 μ 2 1 + ( 2 n - 3 ) ! ! ( 2 n - 5 ) ! ! x - μ 1 μ 2 2 - n ,
which is identical to the solution derived from the equivalent problem.
(4α)
The l-differentiable compactly-supported solution ( α > 1 ). In this case, the polynomial k λ ˜ k x k should be positive between its real roots. This occurs provided that λ ˜ 2 < 0 and ( λ ˜ 1 ) 2 - λ ˜ 2 λ ˜ 0 > 0 . Using the indicator function χ ( x 1 , x 2 ) , with x 1 , x 2 the roots of the polynomial, we find the previous solution with a relative minus sign between the terms inside the parentheses, while the power is now positive.

5. Comparison with the FDE and PME Solutions

The pdf, f ^ , which maximizes the Shannon entropy under the finiteness of the second moment turns out to be identical to the fundamental solution of the diffusion equation with a Dirac point source. It is worth noting that f ^ is actually a global maximum of H 1 . This observation is accidental, as one may justify from the study of the corresponding optimization problem for the Rényi entropy. In particular, the nonlinear, initial-valued problem:
t u ( x , t ) = Δ u α ( x , t ) , ( x , t ) R d × ( 0 , )
lim t 0 + R d u ( x , t ) d x = δ 0 ( x )
has the following self-similar solutions [19]:
FDE solution : u ( x , t ) α < 1 = t - δ C α < 1 + ( 1 - α ) δ 2 α d x t γ 2 - 1 1 - α ,
PME solution : u ( x , t ) α > 1 = t - δ C α > 1 - ( α - 1 ) δ 2 α d x t γ 2 + 1 α - 1 ,
where FDE, PME stand for fast diffusion equation, porous medium equation, respectively,
δ = d γ , γ = 1 ( 2 + d ( α - 1 ) ) , | κ | = | 1 - α | 2 α d δ ,
C α < 1 = B ( d 2 , 1 1 - α - d 2 ) | S d - 1 | 2 | κ | d 2 2 γ ( 1 - α ) and C α > 1 = 2 | κ | d 2 B ( d 2 , α α - 1 ) | S d - 1 | 2 γ ( α - 1 ) .
The d-dimensional time-dependent functions:
f ( x , t ) α < 1 = 1 μ 2 d ( t ) A α < 1 1 + β x 2 μ 2 2 ( t ) - 1 1 - α where β β ( α , d ) = | 1 - α | 2 α - d ( 1 - α ) ,
A α < 1 = β d 2 2 | S d - 1 | B ( d 2 , 1 1 - α - d 2 ) if d d + 2 < α < 1 , and
f ( x , t ) α > 1 = 1 μ 2 d ( t ) A α > 1 1 - β x 2 μ 2 2 ( t ) + 1 α - 1
where A α > 1 = β d 2 2 | S d - 1 | B ( d 2 , α α - 1 ) if α > 1
derived from the optimization of Rényi’s entropy, can be shown to satisfy the following initial value problem:
K α t f ( x , t ) = Δ f α ( x , t ) , ( x , t ) R d × ( 0 , )
lim t 0 + R d f ( x , t ) d x = 1
provided that γ γ ( α , d ) = 1 / ( 2 + d ( α - 1 ) ) , and the coefficient is given by:
K α 1 = 2 α ( 2 + d ( α - 1 ) ) ( 2 α + d ( α - 1 ) ) A α 1 α - 1 ,
where μ 2 ( t ) = t γ . The presence of the ratio x / t γ is implied by the self-similar property of the solution, which requires the function in the parentheses to remain invariant under the rescalings: x x ˜ = λ γ x and t t ˜ = λ t . Therefore, in general, the pdf maximizing the Rényi entropy is a solution of an appropriately-constructed diffusion equation problem.
In Figure 1 and Figure 2, we plot the FDE and PME solutions of (38) versus the solutions of (45), respectively.
Figure 1. A snapshot at t = 1 of the FDE solution of (38) (red line) and (45) (blue line) initial-value problems, with parameters: d = 1 , α = 3 / 4 .
Figure 1. A snapshot at t = 1 of the FDE solution of (38) (red line) and (45) (blue line) initial-value problems, with parameters: d = 1 , α = 3 / 4 .
Entropy 17 06056 g001
In the fast diffusion case u < f , since C α < 1 < A α < 1 , d d + 2 < α < 1 , as one may prove using Proposition (7.1) of the Appendix.
Figure 2. A snapshot at t = 1 of the PME solution of (38) (red line) and (45) (blue line) initial-value problems, with parameters: d = 1 , α = 2 . 2 .
Figure 2. A snapshot at t = 1 of the PME solution of (38) (red line) and (45) (blue line) initial-value problems, with parameters: d = 1 , α = 2 . 2 .
Entropy 17 06056 g002
In the porous medium regime u < f up to the threshold value α th . ( d = 1 ) = 1 . 8268 , for which C α > 1 = A α > 1 and then u > f , α > α th . . The threshold value α th . , which depends on the dimension d, is determined arithmetically.

6. The Concavity of Rényi’s Entropy Power

Definition 6.1. Let f : Ω × ( 0 , ) R d × ( 0 , ) R + d × ( 0 , ) be a pdf in D α . The α-weighted Fisher information of f is defined by:
I α [ f ] ( t ) = 1 Ω f α d μ Ω | f α | 2 f d μ ,
while the entropy power of f associated with the Rényi entropy H α is defined by:
N α [ f ] ( t ) = e 2 d + ( α - 1 ) H α [ f ] ( t ) = Ω f α d μ - 1 Ω f α d μ 2 d ( 1 - α ) , 0 < α < , α 1 e 2 d H 1 [ f ] ( t ) , α = 1 μ ( { x : f ( x , t ) > 0 } ) 2 d - 1 , α = 0 .
Proposition 6.2.
Let Ω = E d be the d-dimensional Euclidean space. The entropy power N α is a concave function of t , t ( 0 , ) provided that:
d 2 H α [ f ] ( t ) d t 2 - d ( 2 + d ( α - 1 ) ) t - 2 if d d + 2 < α < 1 - d ( 2 α - 1 ) 2 ( 2 + d ( α - 1 ) ) t - 2 if α > 1 ,
where the right-hand side of the inequality represents the contributions from the global maximum of H α .
Proof 
Using the nonlinear diffusion equation, a straightforward calculation reveals the connection between d H α / d t and I α expressed by the relation:
d H α [ f ] ( t ) d t = K α - 1 I α [ f ] ( t ) .
The entropy power is a concave function of time iff:
d 2 N α [ f ] ( t ) d t 2 0 d 2 H α [ f ] ( t ) d t 2 = K α - 1 d I α [ f ] ( t ) ) d t - 2 d + ( α - 1 ) d H α [ f ] ( t ) d t 2 .
The first step is to establish the identity:
d H α [ f ] ( t ) d t = α 2 K α - 1 ( 1 - 2 α ) R d f α d x R d f 2 α - 1 Δ ln f d x .
For this reason, we rewrite the integral as:
R d | f α | 2 f d x = α ( 1 - α ) ( 2 α - 1 ) R d f 2 α - 1 · ( ln f ) d x - α R d f 2 α - 1 Δ ( ln f ) d x
= α 2 ( 1 - 2 α ) R d f 2 α - 1 Δ ( ln f ) d x .
The next step is to prove that the first time derivative of the Rényi entropy satisfies the following upper bounds:
d H α [ f ] ( t ) d t d ( 2 + d ( α - 1 ) ) t - 1 if d d + 2 < α < 1 d ( 2 α - 1 ) ( 2 + d ( α - 1 ) ) t - 1 if α > 1 .
Note that the contribution of the integral involving the term Δ ( ln f ^ ) is given by:
R d f ^ 2 α - 1 Δ ( ln f ^ ) d x = - 1 | 1 - α | 2 β 1 - d 2 | S d - 1 | t 2 γ ( 1 + d ( α - 1 ) )
× A α < 1 2 α - 1 d B ( d 2 , 1 1 - α - d 2 ) + ( d - 2 ) B ( 1 + d 2 , α 1 - α - d 2 ) if d d + 2 < α < 1 A α > 1 2 α - 1 d B ( d 2 , α α - 1 ) - ( d + 2 ) B ( 1 + d 2 , α α - 1 + 1 ) if α > 1 ,
while:
α 2 ( 1 - 2 α ) K α - 1 R d f α ( x , t ) d x = α ( 2 α + d ( α - 1 ) ) β d 2 t d γ ( α - 1 ) 2 ( 1 - 2 α ) ( 2 + d ( α - 1 ) ) | S d - 1 | A α < 1 1 - 2 α B ( d 2 , α 1 - α - d 2 ) if d d + 2 < α < 1 A α > 1 1 - 2 α B ( d 2 , α α - 1 + 1 ) if α > 1 .
Substituting (54) into (51), we recover the expected result. Note that in the α 1 - limit, we reproduce the well-known result valid for Shannon’s entropy.
Next, we prove the concavity of Rény’s power entropy on a different setting. This problem was also studied in [13], but our approach leads to a condition not predicted before. We will need the following lemma.
Lemma 6.3.
Let G be defined as:
G α [ f ] ( t ) = R d | f α | 2 f d x = R d f | v | 2 d x w i t h v = α α - 1 f α - 1
then:
d G α [ f ] ( t ) d t = - 2 K α - 1 R d f α v 2 + ( α - 1 ) ( Δ v ) 2 d x a n d
d I α [ f ] ( t ) d t = ( 1 - α ) K α - 1 I α 2 + 1 R d f α d x d G α [ f ] ( t ) d t .
Proof 
Using the porous medium equations for f , v , as well as the relation f α = f v , we have:
d G [ f ] ( t ) d t = R d f t | v | 2 + 2 f t v ( x , t ) · v t d x , v t = v t
= K α - 1 R d Δ f α | v | 2 + 2 f v · ( α - 1 ) v v + | v | 2 d x
= K α - 1 R d Δ f α | v | 2 + 2 ( α - 1 ) f | v | 2 Δ v + f v v · Δ v + 2 f v · | v | 2 d x
= K α - 1 R d Δ f α | v | 2 + 2 ( α - 1 ) f α · v Δ v + 2 α f α v · Δ v + 2 f α · | v | 2 d x
= K α - 1 R d Δ f α - 2 Δ f α | v | 2 - 2 ( α - 1 ) f α ( Δ v ) 2 + v · Δ v + 2 α f α v · Δ v d x
= K α - 1 R d f α 2 v · Δ v - Δ ( | v | 2 ) - 2 ( α - 1 ) f α ( Δ v ) 2 d x
= - 2 K α - 1 R d f α v 2 + ( α - 1 ) ( Δ v ) 2 d x ,
where partial integrations in the fourth and fifth equalities have been used. Furthermore, in the last step, Bochner’s formula in Euclidean space has been applied. Relation (59) is proven using (58).
Theorem 6.4.
The Rényi entropy power, for self-similar solutions, is concave in t, provided that the following inequality is satisfied:
R d f α ( ξ ) - C f 2 α - 1 ( ξ ) d ξ 0 , ξ = x t γ , C = 1 α 2 ( 1 + d ( α - 1 ) ) ,
where the equality holds for α = 1 .
Proof 
The Rényi entropy power is concave in t iff Condition (51) holds, which can be written equivalently as:
I α 2 [ f ] ( t ) d R d f α d x - 1 R d f α v 2 + ( α - 1 ) ( Δ v ) 2 d x
d ( 1 + d ( α - 1 ) ) R d f α d x - 1 R d f α v 2 d x
d ( 1 + d ( α - 1 ) ) J α [ f ] ( t ) ,
where v 2 = Hess v 2 = i , j = 1 d ( i j v ) 2 , and we used the identity Hess v 2 = Tr ( Hess v ) 2 Tr ( Hess v ) 2 / d . Expanding · 2 and writing it in terms of the logarithm of f, the last integral can be cast into the form:
R d f α v 2 d x = α 2 R d f 2 α - 1 i , j = 1 d ( α - 1 ) ( i ln f ) ( j ln f ) + i j 2 ln f 2 d x ,
which by applying the Cauchy-Schwarz inequality, we get that:
R d f α v 2 d x α 2 d R d f 2 α - 1 d x - 1 R d f 2 α - 1 ( α - 1 ) | ln f | 2 + Δ ln f d x 2 .
Using the identity Δ f α f α = | ln f α | 2 + Δ ln f α , we eliminate the | ln f | 2 term from the previous relation and obtain:
J α [ f ] ( t ) α 2 d R d f α R d f 2 α - 1 I α 2 ( f ) 1 d ( 1 + d ( α - 1 ) ) I α 2 ( f ) .
In 1979, Aronson and Bénilan obtained a second-order differential inequality of the form [20]:
i = 1 d x i α f α - 2 f x i = Δ α α - 1 f α - 1 - d d ( α - 1 ) + 2 t , α > α c : = 1 - 2 d ,
which applies to all positive smooth solutions of the porous medium equation defined on the whole Euclidean space. For general solutions and d = 1 the restriction is α > 0 . In 1986, Li-Yau studied a heat type flow [21] on complete Riemannian manifolds ( M , g ) with a nonnegative Ricci scalar, and for any positive function f on M and any t > 0 , they arrived at the following sharp lower bound:
Δ ( ln f ) - d 2 t ,
pointwisely, where Δ is the Laplace–Beltrami operator on M. An extension of the Aronson and Bénilan estimate to the PME flow for all α > 1 and the FDE flow for α ( α c , 1 ) on complete Riemannian manifolds with Ricci scalar bounded from below was given in [22].
In our case, these bounds change, because our pdf differs from the solution of the PME derived from the gradient flow of the functional [23]:
E [ f ] ( t ) = 1 1 - α R d f α ( x , t ) d x .

7. Conclusion

In this article, we have proven that the pdfs that maximize Rényi’s entropy under the conditions of finite variance and of zero or non-zero mean are given by a one-parameter family of functions, which belong to S ( R d ) for α < 1 and to C c l ( R d ) for α > 1 . This one-parameter family of functions is a global maximum of the entropy and satisfies the non-linear diffusion Equation (45). The L norms of these solutions when compared to the corresponding ones derived from the fast and porous medium diffusion equation initial value problem (38) appear to behave differently, as seen in Figure 1 and Figure 2.
If one considers finite even moments of the random variable X, as constraints, and tries to solve the corresponding maximization problem, then the S ( R ) solution exists whenever f ^ α - 1 is a complete polynomial of even degree or, equivalently, when all of the coefficients λ ˜ 2 k + 1 vanish. In those cases, there exists the possibility not to have interception points with the x-axis, since the roots come into conjugate complex pairs. The compactly-supported solution, under certain conditions, can always be determined.
The concavity of the entropy power holds whenever the second time derivative of the entropy varies according to (49) or the function f α - C f 2 α - 1 belongs to L 1 ( R d ) . It would be more appealing to have a deeper understanding of the origin of the later constraint, which, at this stage, seems to be a requirement for consistency.

Acknowledgements

The author would like to thank Jiannis K. Pachos for the fruitful discussions related to this project, the Department of Physics and Astronomy of the University of Leeds for its hospitality during my visit (May–July 2015) and the anonymous reviewers for their constructive comments and suggestions, which improved the quality of this article.

Conflicts of Interest

The author declares no conflict of interest.

Appendix A

Using the integral Formulas (3.241.4) and (3.251.1) of [18], one can prove the following ones, used in Section 3,
0 x μ - 1 ( 1 + x 2 ) ν - 1 d x = 1 2 0 1 u μ 2 - 1 ( 1 - u ) - ν - μ 2 = 1 2 B ( μ 2 , 1 - ν - μ 2 ) ,
where we made the change of variable u = x 2 / ( 1 + x 2 ) .
S d - 1 0 r d - 1 ( C ( α ) + g ( α ) r 2 ) λ ( α ) d r d S = | S d - 1 | C d 2 - λ 2 g d 2 B ( d 2 , λ - d 2 ) where C , g , λ > 0 , d < 2 λ
and | S d - 1 | = 2 π d 2 Γ d 2 is the volume of the unit ball and B ( x , y ) = Γ ( x ) Γ ( y ) / Γ ( x + y ) is the Euler’s beta function. The constraint d < 2 λ for λ = 1 / ( 1 - α ) allows only values:
α c 0 = 1 - 2 d < α < 1 for d 3 ,
while α c 0 = 0 for d = 1 , 2 .
S d - 1 0 r d + 1 ( C ( α ) + g ( α ) r 2 ) λ ( α ) d r d S = | S d - 1 | C d 2 + 1 - λ 2 g 1 + d 2 B ( d 2 + 1 , λ - d 2 - 1 ) where d + 2 < 2 λ .
The constraint between d , λ becomes: α c 2 = 1 - 2 d + 2 < α < 1 for d 1 .
The last formula is:
S d - 1 0 r d - 1 ( C ( α ) - g ( α ) r 2 ) + k ( α ) d r d S = | S d - 1 | 1 2 g ( α ) - d 2 C ( α ) d 2 + k ( α ) B ( d 2 , k ( α ) + 1 ) if k ( α ) > 0 .
Formulas (A.2) and (A.3) are direct consequences of (3.241.4), (3.251.1) and (A.1). Another integral formula used in Section 4 is (3.252.6) of [18].
Identities involving the Euler’s beta function:
lim ρ ρ s 2 B ( s 2 , ρ - s 2 ) = Γ s 2
Proof of (A.6): The large asymptotic expansion of the ratio of gamma functions [24] is given by:
Γ ( ρ + a ) Γ ( ρ + b ) = ρ a - b 1 + ( a - b ) ( a + b - 1 ) ρ + O ( ρ - 2 ) .
Substituting the values a = - s / 2 , b = 0 , in the previous expression and taking the limit ρ , we recover the desired result.
d B ( d 2 , 1 1 - α - d 2 ) B ( d 2 , α 1 - α - d 2 ) + ( d - 2 ) B ( 1 + d 2 , α 1 - α - d 2 ) B ( d 2 , α 1 - α - d 2 ) = d α ( 2 α - 1 )
d B ( d 2 , α α - 1 ) B ( d 2 , α α - 1 + 1 ) - ( d + 2 ) B ( 1 + d 2 , 1 α - 1 + 1 ) B ( d 2 , α α - 1 + 1 ) = d α .
Proof of (A.8): These are direct consequences of the definition of the Euler’s beta function and the gamma’s function property Γ ( x + 1 ) = x Γ ( x ) .
B ( k + 1 2 , n - k - 1 2 ) = ( 2 k - 1 ) ! ! ( 2 n - 2 k - 3 ) ! ! ( 2 n - 2 ) ! ! π , k Z +
B ( 1 2 , n + 1 ) = 2 2 n + 1 B ( n + 1 , n + 1 ) .
Proof of (A.9): Use the doubling formula for gamma functions and Γ ( n ) in terms of the double factorial:
Γ n + 1 2 = π 2 n ( 2 n - 1 ) ! ! , Γ ( n ) = ( 2 n - 2 ) ! ! 2 n - 1 .
Proposition 7.1.
The function Γ ( x ) / Γ ( x + c ) with x , c R + is decreasing in x.
Proof of (7.1): It is enough to show that:
d d x ln Γ ( x ) Γ ( x + c ) = ψ ( x ) - ψ ( x + c ) < 0 .
A straightforward computation gives:
- ln 1 + c x - c x ( x + c ) ψ ( x ) - ψ ( x + c ) - ln 1 + c x - c 2 x ( x + c ) < 0
by using the inequality: ln x - 1 x ψ ( x ) ln x - 1 2 x .

References

  1. Horodecki, R.; Horodecki, P.; Horodecki, M.; Horodecki, K. Quantum entanglement. Rev. Mod. Phys. 2009, 81, 865–942. [Google Scholar] [CrossRef]
  2. Evangelisti, S. Quantum Correlations in Field Theory and Integrable Systems; Minkowski Institute Press: Montreal, QC, Canada, 2013. [Google Scholar]
  3. Geman, D.; Jedynak, B. An active testing model for tracking roads in satellite images. IEEE Trans. Pattern Anal. Machine Intell. 1996, 1, 10–17. [Google Scholar] [CrossRef]
  4. Jenssen, R.; Hild, K.E.; Erdogmus, D.; Principe, J.C.; Eltoft, T. Clustering using Renyi’s Entropy. In Proceedings of the International Joint Conference on Neural Networks, Portland, OR, USA, 20–24 July 2003; pp. 523–528.
  5. Bennett, C.H.; Brassard, G.; Crépeau, C.; Maurer, U.M. Generalized privacy amplification. IEEE Trans. Inf. Theory 1995, 41, 1915–1923. [Google Scholar] [CrossRef]
  6. Sahoo, P.; Wilkins, V.; Yeager, J. Threshold selection using Reny’is entropy. Pattern Recognit. 1997, 30, 71–84. [Google Scholar] [CrossRef]
  7. Sankur, B.; Sezgin, M. Image thresholding techniques: A survey over categories. Pattern Recognit. 2001, 34, 1573–1607. [Google Scholar]
  8. Johnson, O.; Vignat, C. Some results concerning maximum Rényi entropy distributions. Annales de l’Institut Henri Poincare (B) 2007, 43, 339–351. [Google Scholar] [CrossRef]
  9. Barenblatt, G.I. On some unsteady motions of a liquid and gas in a porous medium. Prikl. Mat. Mec. 1952, 16, 67–78. [Google Scholar]
  10. Barenblatt, G.I. Scaling, Self-Similarity and Intermediate Asymptotics; Cambridge University Press: Cambridge, UK, 1996. [Google Scholar]
  11. Zeld’dovich, Y.B.; Kompaneets, A.S. Towards a theory of heat conduction with thermal conductivity depending on the temperature. In Collection of Papers Dedicated to 70th Birthday of Academician A. F. Ioffe; Izd. Akad. Nauk SSSR: Moscow, Russia, 1950; pp. 61–71. [Google Scholar]
  12. Villani, C. A short proof of the “concavity of entropy power”. IEEE Trans. Inf. Theory 2000, 46, 1695–1696. [Google Scholar] [CrossRef]
  13. Savaré, G.; Toscani, G. The concavity of Rényi entropy power. IEEE Trans. Inf. Theory 2014, 60, 2687–2693. [Google Scholar] [CrossRef]
  14. Liberzon, D. Calculus of Variations and Optimal Control Theory; Princeton University Press: Princeton, NJ, USA, 2012. [Google Scholar]
  15. Havrda, M.; Charvát, F. Quantification method of classification processes: Concept of structural α-entropy. Kybernetika 1967, 3, 30–35. [Google Scholar]
  16. Tsallis, C. Possible generalizations of Boltzmann and Gibbs statistics. J. Stat. Phys. 1988, 52, 479–487. [Google Scholar] [CrossRef]
  17. Lutwak, E.; Yang, D.; Zhang, G. Cramér-Rao and moment-entropy inequalities for Rényi entropy and generalized Fisher information. IEEE Trans. Inf. Theory 2005, 51, 473–478. [Google Scholar] [CrossRef]
  18. Gradshteyn, I.S.; Ryzhik, I.M. Table of Integrals, Series, and Products; Academic Press: Boston, MA, USA, 1994. [Google Scholar]
  19. Vázquez, J.L. The Porous Medium Equation: Mathematical Theory; Oxford University Press: Oxford, UK, 2007. [Google Scholar]
  20. Aronson, D.G.; Bénilan, P. Régularité des solutions de l’équation des milieux poreux dans ℝn. C. R. Acad. Sci. Paris. Sér. 1979, A-B 288, 103–105. [Google Scholar]
  21. Li, P.; Yau, S.T. On the parabolic kernel of the Schrödinger operator. Acta Math. 1986, 56, 153–201. [Google Scholar] [CrossRef]
  22. Lu, P.; Ni, N.; Vázquez, L.J.; Villani, C. Local Aronson-Bénilan estimates and entropy formulae for porous medium and fast diffusion equations on manifolds. J. Math. Pures Appl. 2009, 91, 1–19. [Google Scholar] [CrossRef]
  23. Otto, F. The geometry of dissipative evolution equations: The porous medium equation. Commun. Partial Differ. Equ. 2001, 26, 101–174. [Google Scholar] [CrossRef]
  24. Tricomi, F.G.; Erdélyi, A. The asymptotic expansion of a ratio of gamma functions. Pacific J. Math. 1951, 1, 133–142. [Google Scholar] [CrossRef]

Share and Cite

MDPI and ACS Style

Hatzinikitas, A.N. Self-Similar Solutions of Rényi’s Entropy and the Concavity of Its Entropy Power. Entropy 2015, 17, 6056-6071. https://doi.org/10.3390/e17096056

AMA Style

Hatzinikitas AN. Self-Similar Solutions of Rényi’s Entropy and the Concavity of Its Entropy Power. Entropy. 2015; 17(9):6056-6071. https://doi.org/10.3390/e17096056

Chicago/Turabian Style

Hatzinikitas, Agapitos N. 2015. "Self-Similar Solutions of Rényi’s Entropy and the Concavity of Its Entropy Power" Entropy 17, no. 9: 6056-6071. https://doi.org/10.3390/e17096056

APA Style

Hatzinikitas, A. N. (2015). Self-Similar Solutions of Rényi’s Entropy and the Concavity of Its Entropy Power. Entropy, 17(9), 6056-6071. https://doi.org/10.3390/e17096056

Article Metrics

Back to TopTop