Next Article in Journal
How Can We Fully Use Noiseless Feedback to Enhance the Security of the Broadcast Channel with Confidential Messages
Next Article in Special Issue
An Investigation into the Relationship among Psychiatric, Demographic and Socio-Economic Variables with Bayesian Network Modeling
Previous Article in Journal
Coarse-Graining and the Blackwell Order
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Generalized Skew-Normal Negentropy and Its Application to Fish Condition Factor Time Series

by
Reinaldo B. Arellano-Valle
1,
Javier E. Contreras-Reyes
2,3,4,* and
Milan Stehlík
4,5,6
1
Departamento de Estadística, Pontificia Universidad Católica de Chile, Santiago 7820436, Chile
2
Departamento de Matemática, Universidad Técnica Federico Santa María, Valparaíso 2390123, Chile
3
Departamento de Evaluación de Recursos, División de Investigación Pesquera, Instituto de Fomento Pesquero (IFOP), Valparaíso 2361827, Chile
4
Instituto de Estadística, Universidad de Valparaíso, Valparaíso 2360102, Chile
5
Institute of Applied Statistics, Johannes Kepler University Linz, Linz 4040, Austria
6
Linz Institute of Technology, Johannes Kepler University Linz, Linz 4040, Austria
*
Author to whom correspondence should be addressed.
Entropy 2017, 19(10), 528; https://doi.org/10.3390/e19100528
Submission received: 7 September 2017 / Revised: 28 September 2017 / Accepted: 30 September 2017 / Published: 6 October 2017
(This article belongs to the Special Issue Foundations of Statistics)

Abstract

:
The problem of measuring the disparity of a particular probability density function from a normal one has been addressed in several recent studies. The most used technique to deal with the problem has been exact expressions using information measures over particular distributions. In this paper, we consider a class of asymmetric distributions with a normal kernel, called Generalized Skew-Normal (GSN) distributions. We measure the degrees of disparity of these distributions from the normal distribution by using exact expressions for the GSN negentropy in terms of cumulants. Specifically, we focus on skew-normal and modified skew-normal distributions. Then, we establish the Kullback–Leibler divergences between each GSN distribution and the normal one in terms of their negentropies to develop hypothesis testing for normality. Finally, we apply this result to condition factor time series of anchovies off northern Chile.

1. Introduction

Recent studies deal with the problem of measuring the disparity of a particular probability density function (pdf) from the normal one [1]. A typical technique to deal with the problem has been exact expressions using information measures over particular distributions. For example, Vidal et al. [2] measure the sensitivity of the skewness parameter using the L 1 distance between symmetric and asymmetric distributions. Stehlík [3] proved results on the decomposition of Kullback–Leibler (KL) divergences [4] in the gamma and normal family for divergence between the Maximum Likelihood Estimator (MLE) of the canonical parameter and the canonical parameter of the regular exponential family [5]. Contreras-Reyes and Arellano-Valle [6] considered Jeffrey’s (J) divergence [7] to compare the multivariate Skew-Normal (SN) from the normal distribution, and Gómez-Villegas et al. [8] assessed the effect of kurtosis deviations from normality on conditional distributions, such as the multivariate exponential power family. Main et al. [9] evaluated the local effect of asymmetry deviations from normality using the KL divergence measure of the SN distribution and then compared the local sensitivity with Mardia’s and Malkovich–Afifi’s skewness indexes. They also agree on the use of the SN model to regulate the asymmetry of an empirical distribution because it reflects the deviation in a tractable way. Dette et al. [10] characterizes the “disparity” between the skew-symmetric models and their symmetric counterparts in terms of the total variation distance, which is later used to construct priors. The paper provides additional insights, to those provided in Vidal et al. [2], on the interpretation of this distance and also discusses the usage of the KL divergence among several other distances.
Some recent applications of measuring the disparity of a particular pdf from the normal one using negentropy include those by Gao and Zhang [11] and Wang et al. [12], where the negentropy method has been successfully applied to seismic wavelet estimation. Pires and Ribeiro [13] considered the negentropy to measure the distance of non-Gaussian information from the normal one in independent components, with application to Northern Hemispheric winter monthly variability of a high-dimensional quasi-geostrophic atmospheric model. Furthermore, Pires and Hannachi [14] used a tensorial invariant approximation of the multivariate negentropy in terms of a linear combination of squared coskewness and cokurtosis. Then, the method was applied to global sea surface temperature anomalies, after data anomalies were tested through a non-Gaussian distribution.
In this paper, we develop a procedure, based on KL divergences, to test the significance of the skewness parameter in the Generalized Skew-Normal (GSN) distributions, a flexible class of distributions that includes the SN and normal ones as particular cases. We consider asymptotic expansions of moments and cumulants for the negentropy of two particular cases: the SN and Modified Skew-Normal (MSN) distributions. Given that SN distributions do not accomplish the regularity condition of Fisher Information Matrix (FIM) at η = 0 , normality is tested based on the MSN distribution [15]. This allows one to implement an asymptotic normality test for testing the significance of the skewness parameter. Numerical results are studied by: (a) comparing numerical integration methods with proposed asymptotic expansions; (b) comparing the asymptotic test with the likelihood ratio test and the asymptotic normality test given by Arrué et al. [15] ; and (c) applying the proposed test to condition factor time series of anchovy (Engraulis ringens).
This paper is organized as follows: information theoretic measures are described in Section 2. In Section 3, we provide an asymptotic expansion in terms of the corresponding cumulants for the GSN, SN and MSN negentropies. We also express the KL and J divergences among each GSN distribution and the normal one in terms of negentropies (as cumulants’ expansion series) to develop the hypothesis test about the significance of the skewness parameter together with a simulation study (Section 4). A simulation study is given in Section 5. In Section 6, the real data of the condition factor time series of anchovies off northern Chile illustrate the usefulness of the developed methodology. The discussion concludes the paper.

2. Shannon Entropy and Related Measures

The Shannon Entropy (SE) of a random variable Z with pdf f is given by:
H ( Z ) = E { log f ( Z ) } .
The SE of a localization-scale random variable X = μ + σ Z does not depend on μ and is such that H ( X ) = log σ + H ( Z ) (see, e.g., [16]). The SE could serve to define a measure of disparity from normality, the so-called negentropy [17], which is zero for a Gaussian variable and positive for any distribution. It is defined by:
N ( Z ) = H ( Z 0 ) H ( Z ) ,
where Z 0 is a normal random variable with the same mean and variance as those of Z. Equation (2) expresses the negentropy in terms of the standardized version of Z, say Z * , as N ( Z ) = H ( Z 0 ) H ( Z * ) = N ( Z * ) ; here, Z * has zero mean and unit variance. Thus, negentropy measures essentially the amount of information that departs from the normal entropy. Furthermore, clearly, the negentropy becomes the KL divergence (see Equation (3) below) between Z * and Z 0 .
Given that the calculus of negentropy presents a computational challenge, where the integral involves the pdf of Z [16,18], different approximations of negentropy are used, such as cumulants’ expansion series [17,19]. Withers and Nadarajah [19] provided exact and explicit series expansions for the SE and negentropy of a standardized pdf f on R , in terms of cumulants. Yet, they did not perform numerical studies that allow evaluation and comparison with other procedures in some specific families of distributions.
Other measures related to the SE are KL and J divergences. They measure the degree of divergence between the distributions of two random variables Z 1 and Z 2 with pdfs f 1 and f 2 , respectively. The KL divergence of the pdf for Z 1 from the pdf for Z 2 is defined as:
K ( Z 1 , Z 2 ) = E log f 1 ( Z 1 ) f 2 ( Z 1 ) .
as indicated in the notation, the expectation is defined with respect to the pdf f 1 for Z 1 . Since in general K ( Z 1 , Z 2 ) differs from K ( Z 2 , Z 1 ) , the J divergence is considered as a symmetric version of the KL divergence, which is defined by:
J ( Z 1 , Z 2 ) = K ( Z 1 , Z 2 ) + K ( Z 2 , Z 1 ) .

3. Generalized Skew-Normal Distributions

An attractive class of Skew-Symmetric (SS) distributions defined in terms of the pdf appears in Azzalini [20], Azzalini and Capitanio [21] and Gupta et al. [22]:
f ( z ; η ) = 2 f ( z ) G { w ( z ; η ) } , z R ,
where η R represents a skewness/shape parameter, f and G are the respective pdf and cumulative distribution function (cdf) of symmetrical continuous distributions and w ( z ; η ) is an odd function of z, with w ( 0 ; η ) = 0 for any fixed value of η . Furthermore, we assume that w ( z ; η 0 ) = 0 for all z and some value η 0 of η (typically η 0 = 0 ), so that f ( z ; η 0 ) = f ( z ) , thus recovering symmetry.
The notation Z S S ( η ; f , G , w ) expresses that random variable Z has a distribution with the pdf given by (5). If f ( z ) = ϕ ( z ) represents the pdf of the standardized normal distribution, denoted by N ( 0 , 1 ) , then (5) becomes a family of skew-symmetric distributions generated by the normal kernel, the GSN family. In this case, Z G S N ( η ; G , w ) emerges. An important property of the GSN random variable Z is that all its moments are finite. In particular, it possesses the same even moments of Z 0 N ( 0 , 1 ) . For instance, E ( Z 2 ) = 1 , and so, V a r ( Z ) = 1 μ z 2 , where μ z = E ( Z ) . The most popular GSN distribution is Skew-Normal (SN) [23], for which w ( z ; η ) = η z and G ( z ) = Φ ( z ) is the cdf of the standardized normal distribution. Therefore, Z S N ( η ) expresses that Z follows an SN distribution. The location-scale extension of the SS pdf in (5) follows by applying the Jacobian method to the linear random variable X = μ + σ Z , where μ R and σ > 0 . In this case, we state that X follows an SS distribution with location parameter μ , scale parameter σ and shape/skewness parameter η and obtains X S S ( μ , σ 2 , η ; f , G , w ) . Furthermore, we write X G S N ( μ , σ 2 , η ; G , w ) if f = ϕ , X G S N ( μ , σ 2 , η ; G ) if f = ϕ and w ( z ; η ) = η z ; and X S N ( μ , σ 2 , η ) if f = ϕ , w ( z ; η ) = η z and G = Φ .
Two other members of the GSN family that have been studied recently are the Skew-Normal-Cauchy (SNC) distribution [24,25], which follows from (5) by taking f ( z ) = ϕ ( z ) , w ( z ; η ) = η z and G ( z ) = 1 / 2 + ( 1 / π ) arctan ( z ) , and the Modified Skew-Normal (MSN) distribution [15], for which f ( z ) = ϕ ( z ) , w ( z ; η ) = η z / 1 + z 2 and G ( z ) = Φ ( z ) . Nadarajah and Kotz [24] recall that the SNC distribution appears to attain a higher degree of sharpness than the normal distribution, i.e., disparity exists from the common normal distribution produced by the skewness parameter η . A random variable Z with the SNC or MSN distribution is denoted, respectively, by Z S N C ( η ) or Z M S N ( η ) and by X S N C ( μ , σ 2 , η ; G ) or X M S N ( μ , σ 2 , η ; G ) for their respective location-scale extensions.
We consider the SE for the GSN subclass, i.e., the SE of Z G S N ( η ; G , w ) . Thus, assuming a normal kernel in (5), we get the GSN-SE given by:
H ( Z ) = H ( Z 0 ) E log [ 2 G { w ( Z ; η ) } ] ,
where H ( Z 0 ) = ( 1 / 2 ) log ( 2 π e ) is the SE of Z 0 . It is assumed that a specific skewness value η 0 exists so that w ( z ; η 0 ) = 0 and so that G { w ( z ; η 0 ) } = 1 / 2 , thus recovering symmetry at η = η 0 . Therefore, at η = η 0 , Z and Z 0 have the same distribution and thus the same SE.
Let μ z = E ( Z ) and σ z 2 = V a r ( Z ) be the mean and variance of Z G S N ( η ; G , w ) , respectively, which must constitute functions of the skewness parameter η . Since σ z 2 = 1 μ z 2 and H ( Z 0 ) = log σ z + H ( Z 0 ) , we get from (2) that the negentropy of Z becomes:
N ( Z ) = 1 2 log ( 1 μ z 2 ) + E log [ 2 G { w ( Z ; η ) } ] .
Since at η = η 0 , we have by symmetry μ z = 0 and w ( Z ; η 0 ) = 0 , so the negentropy in this case is null, as expected. Clearly, SE and negentropy depend on the choice of the functions G ( · ) and w ( · ; η ) . In this paper, we consider both families of GSN distributions for which η 0 = 0 , with w ( z ; 0 ) = w ( 0 ; η ) = 0 and w ( z ; η ) = w ( z ; η ) , thus following that Z G S N ( η ; G , w ) and recovering the normality at η = 0 . Examples of this type of functions are w ( z ; η ) = η u ( z ) and w ( z ; η ) = u ( η z ) for some odd function u ( z ) , with u ( 0 ) = 0 . In this case, recalling that η Z = d τ Z τ , where Z τ G S N ( τ ; G , w ) , τ = | η | , and “ = d ” denotes equality in the distribution, we obtain:
E log [ 2 G { w ( Z ; η ) } ] = E log [ 2 G { w ( Z τ ; τ ) } ] ,
thus H ( Z ) = H ( Z τ ) and N ( Z ) = N ( Z τ ) . That is, the entropy and negentropy of Z S N ( η ) depend on the skewness parameter η only through its absolute value τ = | η | .
We have interest in both KL and J divergences for a GSN distribution with respect to the normal distribution. that is, assuming in (3) and (4) that Z 1 = Z G S N ( η ; G , w ) and Z 2 = Z 0 . In this case, remembering that η Z = d τ Z τ , where Z τ G S N ( τ ; G , w ) and τ = | η | , we have K ( Z , Z 0 ) = K ( Z τ , Z 0 ) and K ( Z 0 , Z ) = K ( Z 0 , Z τ ) , with:
K ( Z τ , Z 0 ) = E log 2 G { w ( Z τ ; τ ) } and K ( Z 0 , Z τ ) = E log 2 G { w ( Z 0 ; τ ) } .
Therefore, J ( Z , Z 0 ) = J ( Z τ , Z 0 ) , with:
J ( Z τ , Z 0 ) = E log 2 G { w ( Z τ ; τ ) } E log 2 G { w ( Z 0 ; τ ) } .
We also develop asymptotic expansions of the J divergence for the SN and MSN distributions from the normal distribution. To do this, we consider the following preliminary result, the proof of which stems from (9) and (10) by using the Taylor expansion of ζ ( z ; τ ) = log [ 2 G { w ( z ; τ ) } ] at z = 0 and also because of the facts that (a) all moments of Z τ G S N ( τ ; G , w ) are finite and (b) Z τ and Z 0 contain the same even moments.
Lemma 1.
Consider the composite function ζ ( z ; τ ) = log [ 2 G { w ( z ; τ ) } ] , τ > 0 , by assuming that both functions G ( z ) and w ( z ; τ ) are infinitely differentiable; hence, also ζ ( z ; τ ) is infinitely differentiable at z = 0 . If Z τ G S N ( τ ; G , u ) , then:
K ( Z τ , Z 0 ) = k = 1 ζ ( k ) ( 0 ; τ ) k ! E ( Z τ k ) , J ( Z τ , Z 0 ) = k = 1 ζ ( 2 k 1 ) ( 0 ; τ ) ( 2 k 1 ) ! E ( Z τ 2 k 1 ) ,
where ζ ( k ) ( z ; τ ) is the k-th derivative of ζ ( z ; τ ) . Moreover, from (11), the expressions (6) and (7) for the SE and negentropy of the GSN distributions have the forms:
H ( Z τ ) = 1 2 log ( 2 π e ) k = 1 ζ ( k ) ( 0 ; τ ) k ! E ( Z τ k ) , N ( Z τ ) = 1 2 log ( 1 μ τ 2 ) + k = 1 ζ ( k ) ( 0 ; τ ) k ! E ( Z τ k ) ,
respectively, where μ τ = E ( Z τ ) .
Notice in Lemma 1 that the coefficient ζ ( k ) ( 0 ; τ ) depends on the derivatives of G ( z ) and w ( z ; τ ) at z = 0 , which change for different GSN distributions. Moreover, since the expansion of ζ ( z ; τ ) emerges around z = 0 by assuming a fixed τ , the approximations may not be reasonable for some values of τ .

3.1. Skew-Normal Distribution

If Z S N ( η ) or Z S N ( 0 , 1 , η ) represents an SN random variable, then its pdf is:
f ( z ; η ) = 2 ϕ ( z ) Φ ( η z ) , z R .
Clearly, if η = η 0 = 0 , then (12) reduces to the N ( 0 , 1 ) -pdf. The SN random variable Z can be conveniently represented as a linear combination of half-normal and normal variables through the following stochastic representation [26]:
Z = d δ | U 0 | + 1 δ 2 U ,
where δ = η / 1 + η 2 , U 0 and U are independent and identically distributed with a unit normal distribution. In particular, since the half-normal random variable | U 0 | has mean b = 2 / π and variance one, it follows from (12) that the mean and variance of Z τ , τ = | η | , are given by:
E ( Z τ ) = b δ τ and Var ( Z τ ) = 1 ( b δ τ ) 2 ,
where δ τ = τ / 1 + τ 2 .
In the SN case, G ( z ) = Φ ( z ) and w ( z ; η ) = η z , which are both infinitely differentiable functions at z = 0 . Consequently, the function ζ ( z ; τ ) = ζ 0 ( τ z ) = log { 2 Φ ( τ z ) } is also infinitely differentiable at z = 0 , thus admitting a Taylor expansion about zero. Therefore, since ζ ( k ) ( z ; τ ) = τ k ζ k ( τ z ) , where ζ k ( x ) is the k-th derivative of ζ 0 ( x ) = log { 2 Φ ( x ) } , the expansion (11) in Lemma 1 of E { ζ ( Z τ ; τ ) } = E { ζ 0 ( τ Z τ ) } , Z τ S N ( τ ) , becomes:
E { ζ 0 ( τ Z τ ) } = k = 1 τ k κ k μ k k ! ,
where κ k = ζ k ( 0 ) and μ k = E ( Z τ k ) (see Appendix A).
In summary, since the even moments of Z τ S N ( τ ) are also the even moments of Z 0 , Equation (14) can be rewritten as:
E { ζ 0 ( τ Z τ ) } = k = 1 τ 2 k κ 2 k ( 2 k ) ! μ 2 k E { ζ 0 ( τ Z 0 ) } + k = 1 τ 2 k 1 κ 2 k 1 ( 2 k 1 ) ! μ 2 k 1 J ( Z τ , Z 0 ) .
Hence, considering also Equation (13), we can compute for the SN case the results for the KL and J divergences, SE and negentropy given in Lemma 1 using the following Proposition 1.
Proposition 1.
Let Z τ S N ( τ ) and Z 0 N ( 0 , 1 ) . Then:
K ( Z τ , Z 0 ) = k = 1 τ k κ k μ k k ! , J ( Z τ , Z 0 ) = b k = 1 κ 2 k 1 δ τ 2 k 1 ( 2 k 1 ) ! m = 1 k a k ( m ) τ 2 m 1 , H ( Z τ ) = 1 2 log ( 2 π e ) K ( Z τ , Z 0 ) , N ( Z τ ) = 1 2 log { 1 ( b δ τ ) 2 } + K ( Z τ , Z 0 ) ,
where the coefficients a k ( m ) , k = 1 , 2 , , are given in the Appendix A.
To gain a more complete analysis of the behavior of these series, we need appropriate forms for the calculation of the coefficients κ k = ζ k ( 0 ) , k = 1 , 2 , (see Appendix A).

3.2. Modified Skew-Normal Distribution

The pdf for a random variable Z with MSN distribution, denoted by Z M S N ( η ) , is given by:
f ( z ) = 2 ϕ ( z ) Φ η u ( z ) , z R ,
where u ( z ) = z / 1 + z 2 . Similarly to the SN case, the MSN random variable Z τ M S N ( τ ) , τ = | η | , has even moments equal to the corresponding even moments of the standardized normal random variable Z 0 [15], i.e., μ 2 k = ( 2 k 1 ) ! ! = ( 2 k ) ! / ( 2 k k ! ) (for odd moments μ 2 k 1 ( τ ) , k = 1 , 2 , , see Appendix A).
In the MSN case, G ( w ) = Φ ( w ) and w ( z ; τ ) = τ u ( z ) = τ z / 1 + z 2 , both of which are infinitely differentiable at z = 0 . Thus, in Lemma 1, we have ζ ( z ; τ ) = ζ 0 { τ u ( z ) } , where ζ 0 ( x ) = log { 2 Φ ( x ) } is also infinitely differentiable at z = 0 . Thus, the series expansion of E { ζ ( Z τ ; τ ) } = E [ ζ 0 { τ u ( Z τ ) } ] , Z τ M S N ( τ ) , can be obtained from (11) for which we need the derivatives of the composite function ζ 0 { τ u ( z ) } = log [ 2 Φ { τ u ( z ) } ] . Another way to obtain these derivatives is to define random variable Z τ * = u ( Z τ ) = Z τ / 1 + Z τ 2 and using (14) with Z τ and μ k = E ( Z τ ) replaced by Z τ * and μ k * = E { ( Z τ * ) k } , respectively. Thus, we obtain the series expansion:
E { ζ 0 ( τ Z τ * ) } = k = 1 τ k κ k k ! μ k * .
From Lemma 1, the KL and J divergences, SE and negentropy for the MSN case can be computed using the following Proposition 2.
Proposition 2.
Let Z τ M S N ( τ ) and Z 0 N ( 0 , 1 ) . Then:
K ( Z τ , Z 0 ) = k = 1 τ k κ k k ! μ k * , J ( Z τ , Z 0 ) = k = 1 τ 2 k 1 κ 2 k 1 ( 2 k 1 ) ! μ 2 k 1 * , H ( Z τ ) = 1 2 log ( 2 π e ) K ( Z τ , Z 0 ) , N ( Z τ ) = 1 2 log ( 1 [ b { 2 ξ 1 ( τ ) 1 } ] 2 ) + K ( Z τ , Z 0 ) .
In order to compute the quantities given by Proposition 2, we need to calculate the new moments μ k * = E { ( Z τ * ) k } , k = 1 , 2 , . . . . Since Z τ * = Z τ / 1 + Z τ 2 is a random variable limited to the interval ( 1 , 1 ) , all its moments are finite. In particular, Z τ * clearly has the same even moments as Z 0 * = Z 0 / 1 + Z 0 2 Moreover, from the Jacobian method, the pdf of Z τ * becomes:
f * ( u ) = 2 ( 1 u 2 ) 3 / 2 ϕ u 1 u 2 Φ τ u , u ( 1 , 1 ) .
Hence, the k-th moment of Z τ * is:
μ k * = 1 1 u k f * ( u ) d u , k = 1 , 2 , ,
which must be computed numerically.

3.3. J Divergence between SN and MSN Distributions

In the previous sections, SN and MSN distributions were compared with the normal distribution by means of the J divergence measure. As a byproduct, we were also computing the J divergence between the SN and MSN distributions, both with the same skewness parameter. This allows measuring the distance between these distributions with different w ( z ; η ) ’s. For this, we consider in Equation (4) that Z 1 S N ( τ ) and Z 2 M S N ( τ ) and define the random variables Z i * = Z i / 1 + Z i 2 for i = 0 , 1 , 2 , where Z 0 N ( 0 , 1 ) . Let μ i , k = E ( Z i k ) and μ i , k * = E { ( Z i * ) k } , i = 0 , 1 , 2 . Recall that μ i , 2 k = μ 0 , 2 k and μ i , 2 k * = μ 0 , 2 k * for all k = 1 , 2 , . Thus, using (4) and then the Taylor expansion of ζ 0 ( x ) = log { Φ ( x ) } around x = 0 , Proposition 3 is obtained:
Proposition 3.
Let Z 0 N ( 0 , 1 ) , Z 1 S N ( τ ) and Z 2 M S N ( τ ) . Define the random variables Z i * = Z i / 1 + Z i 2 , i = 1 , 2 . Then:
J ( Z 1 , Z 2 ) = E { ζ 0 ( τ Z 1 ) } E { ζ 0 ( τ Z 1 * ) } + E { ζ 0 ( τ Z 2 * ) } E { ζ 0 ( τ Z 2 ) } = J ( Z 1 , Z 0 ) J ( Z 1 * , Z 0 ) + J ( Z 2 * , Z 0 ) J ( Z 2 , Z 0 ) ,
where as before:
J ( Z i , Z 0 ) = k = 1 τ 2 k 1 μ i , 2 k 1 κ 2 k 1 ( 2 k 1 ) ! , J ( Z i * , Z 0 ) = k = 1 τ 2 k 1 μ i , 2 k 1 * κ 2 k 1 ( 2 k 1 ) ! , i = 1 , 2 .
Proposition 3 indicates that J divergence between SN and MSN distributions is decomposed to the divergences of the normal distribution with each of these distributions, which depends only on their odd moments and cumulants.

4. Asymptotic Tests

Let f ( x ; θ ) , x X , θ Θ , be the pdf of a regular parametric class of distributions, i.e., for which the sample space X does not depend on θ , the parametric space Θ is an open subset of R p , and the regularity conditions (i)–(iii) stated in Salicrú et al. [27] are satisfied. As in Salicrú et al. [27], we denote the KL divergence between f ( x ; θ ) and f ( x ; θ ) , θ , θ Θ , by:
K ( θ , θ ) = X f ( x ; θ ) log f ( x ; θ ) f ( x ; θ ) d x .
Consider the partition θ = ( θ 1 , θ 2 ) , where θ 1 Θ 1 R r and θ 2 Θ 2 = Θ Θ 1 c R p r . Let θ = ( θ 1 , θ 2 0 ) and consider the null hypothesis H 0 : θ 2 = θ 2 0 for a known θ 2 0 Θ Θ 1 c . Let θ ^ = ( θ ^ 1 , θ ^ 2 ) and θ ^ = ( θ ^ 1 , θ 2 0 ) be the (unrestricted) MLE of θ and θ , respectively, both based on a random sample of size n from X with pdf f ( x ; θ ) . Under these conditions, we have from Part (b) of Theorem 2 presented in Salicrú et al. [27] that:
2 n K ( θ ^ , θ ^ ) = 2 n X f ( x ; θ ^ ) log f ( x ; θ ^ ) f ( x ; θ ^ ) d x d n χ p r 2 , θ Θ ,
where “ d ” denotes convergence in distribution and χ s 2 denotes the chi-squared distribution function with s degrees of freedom. From (17), the above null hypothesis can be tested by the statistic 2 n K ( θ ^ , θ ^ ) , which is asymptotically chi-squared distributed with p r degrees of freedom. Specifically, for large values of n, if we observe K ( θ ^ , θ ^ ) = K 0 , then H 0 is rejected at level α if P ( χ p r 2 > 2 n K 0 ) α .

4.1. One-Sample Case: Test for Normality

The result in (17) can be applied for example to construct a normality test from the KL divergence between a regular GSN distribution and the normal distribution. Specifically, consider a random sample X 1 , , X n from X G S N ( μ , σ 2 , η , G , w ) and the null hypothesis H 0 : η = η 0 under which G { w ( z ; η 0 ) } = G ( 0 ) = 1 / 2 ; thus, the GSN random variable X becomes a N ( μ , σ 2 ) random variable. Let θ ^ = ( μ ^ , σ 2 ^ , η ^ ) be the MLE of θ = ( μ , σ 2 , η ) and θ ^ = ( μ ^ , σ 2 ^ , η 0 ) . Therefore, under H 0 : η = η 0 , we have:
2 n K ( θ ^ , θ ^ ) = 2 n K ( Z τ ^ , Z 0 ) d n χ 1 2 ,
where K ( Z τ ^ , Z 0 ) is the MLE of K ( Z τ , Z 0 ) , which is defined in Equation (11) of Lemma 1 and depends only on τ ^ = | η ^ | . As stated in the Introduction, normality is typically obtained from the GSN class at η 0 = 0 or equivalently τ 0 = | η 0 | = 0 .
Azzalini [20], Arellano-Valle and Azzalini [28] and Azzalini and Capitanio [23] recall the singularity of SN FIM at η = 0 , preventing the asymptotic distribution of the above statistic tests. As suggested by Azzalini [20], a solution to recover the non-singularity of the information matrix under the symmetry hypothesis comes from the use of the so-called centered parametrization defined in terms of the mean, variance and the skewness parameters of the SN distribution (see also [28,29]). Otherwise, the FIM of the MSN model is non-singular at η = 0 [15]. Thus, this model satisfies all the standard regularity conditions of Salicrú et al. [27], leading to consistence and asymptotic normality of the MLEs under the null hypothesis of normality. Therefore, the MSN model serves to test the null hypothesis of normality using (18). Hence, the symmetry null hypothesis H 0 : τ = 0 is rejected at level α if P ( χ 1 2 > 2 n K 0 ) α , with K 0 = K ( Z τ ^ , Z 0 ) .

4.2. Two-Sample Case

Consider two independent samples of sizes n 1 and n 2 from X 1 and X 2 , respectively; where θ , θ Θ R p , and X 1 and X 2 have pdf’s f ( x ; θ 1 ) and f ( x ; θ 2 ) , respectively. Suppose partition θ i = ( θ i 1 , θ i 2 ) , i = 1 , 2 , and assume θ 21 = θ 11 Θ 1 R r , so that θ i 2 Θ Θ 1 c R p r , i = 1 , 2 . Let θ ^ i = ( θ ^ 11 , θ ^ i 2 ) be the MLE of θ i = ( θ 11 , θ i 2 ) , i = 1 , 2 , which correspond to the MLE of the full model parameters ( θ 1 , θ 2 ) under null hypothesis H 0 : θ 21 = θ 11 . Thus, Part (b) of Corollary 1 in Salicrú et al. [27] establishes that if the null hypothesis H 0 : θ 22 = θ 12 holds and n 1 n 1 + n 2 n 1 , n 2 λ , with 0 < λ < 1 , then:
2 n 1 n 2 n 1 + n 2 K ( θ ^ 1 , θ ^ 2 ) = 2 n 1 n 2 n 1 + n 2 f ( x ; θ ^ 1 ) log f ( x ; θ ^ 1 ) f ( x ; θ ^ 2 ) d x d n 1 , n 2 χ p r 2 .
Thus, a test of level α for the above homogeneity null hypothesis consists of rejecting H 0 if:
2 n 1 n 2 n 1 + n 2 K ( θ ^ 1 , θ ^ 2 ) > χ p r , 1 α 2 ,
where χ p r , α 2 is the α -th percentile of the χ p r 2 -distribution.
Contreras-Reyes and Arellano-Valle [6] considered the result of Kupperman [30] to develop an asymptotic test of complete homogeneity in terms of the J divergence between two SN distributions. The SN distribution satisfies all the aforementioned regularity conditions when skewness parameter η 0 . Thus, considering this condition, we can also apply (17) and (19) to obtain, respectively, asymptotic tests with one or two samples of other hypotheses not covered by Kupperman’s test.

5. Simulations

In this section, we study the behavior of the series expansions of the SE and negentropy for the SN and MSN distributions. In both cases, we compare the SE and negentropies obtained from their series expansions with their corresponding “exact” versions computed from the Quadpack numerical integration method of Piessens et al. [31]. More precisely, the “exact” expected values E { ζ 0 ( τ Z τ ) } and E { ζ 0 ( τ Z τ * ) } are computed using the Quadpack method as in Arellano-Valle et al. [16], Contreras-Reyes and Arellano-Valle [6] or Contreras-Reyes [18]. From the series expansions, the SE and negentropies were carried out for k = 12 as in Withers and Nadarajah [19]. However, they tend to converge for k = 4 as in the Gram–Charlier and Edgeworth expansion methods (see, e.g., Hyvärinen et al. [17] and Stehlík et al. [1], respectively). All proposed methods are implemented with R software [32].
From Figure 1, we observe that the approximations by series expansions are better in the MSN case (Panels C and D) than in the SN case (Panels A and B). Furthermore, that series expansion approximations are quite exact for small to moderate values of the skewness parameter τ ; more specifically, for 0 τ 2 in the SN case, and 0 τ 4 in the MSN case. Additionally, Panels A and C show that the SE decreases as τ increases, while Panels B and D indicate that the negentropy increases with τ . Finally, as expected in both GSN models, the SE is less than or equal to the SE of the normal model, namely H ( Z 0 ) 1.418 [6,33].
Panel A of Figure 2 shows, respectively, the behavior of the KL divergences of the SN and MSN distributions from the normal one obtained from the expansions in series given in Equations (15) and (16). As in Figure 1, the KL divergence between the SN and normal distributions increases smoothly for values of τ [ 0 , 2 ] , but rises sharply for τ > 2 . Meanwhile, the increase in KL divergence between the MSN and normal distributions seems more stable, at least for τ [ 0 , 5 ] . Crucially, for τ = | η | 2 , the SN model is close to its maximum level of asymmetry, while the MSN model does it for τ = | η | 5 (see [15] (Figure 2)).
Table 1 presents the observed power of the asymptotic test of normality obtained from Equation (18) in Section 4.1, for different sample sizes and values of the skewness parameter. All these results were obtained from 2000 simulations for a nominal level of 5%. In each simulation, the MLE of Z M S N ( η ) was obtained by maximizing the log-likelihood function:
( η ) = n 2 log ( 2 π ) 1 2 i = 1 n z i 2 + i = 1 n log ζ 0 { η u ( z i ) } ,
for shape parameter η and a random sample of size n from Z [15]. Table 1 shows that the proposed test is considerably conservative since the observed rate of incorrect rejections of the normality hypothesis is always lower than the nominal level. The proposed test is also considerably more powerful in large samples ( n 300 ) and values of the skewness parameter far from zero ( | η | 1.2 ). As expected, the power of the test increases with sample size, particularly for small values of the skewness parameter (close to normality), given that statistic 2 n K 0 depends on n despite K 0 being small (Figure 2).
Now, we compare the proposed asymptotic test with two additional tests considered by Arrué et al. [15] for null hypothesis H 0 : η = η 0 versus H 1 : η η 0 : the Likelihood Radio Test (LRT) (see Appendix A) and the asymptotic normality-based test. Since the regularity condition on MSN’s FIM at η = 0 is satisfied, the authors proposed a distributional normal theory for testing H 0 , i.e., based on asymptotic normality of MLE given by n ( θ ^ θ 0 ) d N 3 ( 0 , i M S N 1 ( θ 0 ) ) , as n , where θ ^ = ( μ ^ , σ ^ 2 , η ^ ) is the MLE of θ = ( μ , σ 2 , η ) , θ 0 = ( μ , σ 2 , 0 ) and i M S N 1 ( θ 0 ) is the inverse FIM component related to θ 0 . For asymptotic normality and LRT, they conclude that H 0 is rejected for large values of τ ^ = | η ^ | , and for large values of n, the coverage rate increases when η ^ exists ( H 0 is rejected) (see [15] (Tables 3–5)). Analogously, in Table 6 of Arrué et al. [15], the coverage rate increase when η ^ exists for large values of n.

6. Application to Condition Factor Time Series

To apply our results to a real-world problem, we considered the Condition Factor (CF) index [34], which serves as an important indicator of the fatness condition of fish [18]. The CF index,
C F = W W ^ ( 100 % ) ,
of an individual of length L is computed in terms of the observed weight W = W ( L ) and an estimation W ^ = W ^ ( L ) obtained from the morphometric relationships of the expected weight E ( W ) at length L. Then, the CF index is interpretable as food deficit (<100%) and abundance (>100%) conditions. The expected length-weight relationship is described through the non-linear relationship:
E ( W ) = α L β ,
where α is the theoretical weight at length zero and β is the weight growth rate [35]. According to (21), W ^ is computed as W ^ = α ^ L β ^ , where α ^ and β ^ are obtained by fitting the non-linear regression induced by (21) to the length-weight data obtained from a sample of the species under study.
The CF index can be mainly affected by environmental factors such as El Niño (cold events) or La Niña (warm events). These effects are conductors of threshold biological processes due to the limitation of food. For these reasons, Contreras-Reyes [18] considered a threshold autoregressive model based on the stochastic representation (12) to model CF time series. That is, by assuming an SN distribution with skewness parameter η for the CF index [20], the condition | δ | < 1 ensures the weak stationarity of the process. Additionally, when η is positive, CF values fall below 100% (food deficit). Otherwise, CF values are greater than 100% (food abundance).
We applied hypothesis testing developed in Section 4 to monthly CF time series associated with anchovy from Chile’s northern coast during the period 1990–2010, which were classified by length and sex, for length classes 12,...,18 cm and ALL (all length classes). Therefore, the sample size of each classification depends on the availability of the routine biological sampling program (see more details in [18]). CF were previously standardized, since the shape parameter η is not affected by a linear transformation of the CF [23]. Table 2 shows the η ^ ’s assuming an SN and MSN distribution based on the MLE method of Azzalini [36] and Arrué et al. [15], respectively. For MSN, we considered the log-likelihood function of Equation (20). In both models, negative and positive values of η ^ correspond to asymmetry to the right and left, respectively (see Contreras-Reyes [18] (Figure 5)). This means that CF of the above-mentioned classes are affected by extreme events. As expected, we find generally that for low values of the empirical skewness index, the shape parameter of both distributions is close to zero.
The values of η ^ obtained from the SN and MSN models are presented in Table 2. Since that SN model is not regular at η = 0 , we used only the MSN model to perform the test of normality and LRT for each sample datum. The results of this analysis appear in Table 3 and are not analogous for all the length classes in both groups. In fact, for the group of males, the null hypothesis H 0 : τ = 0 is not rejected, only in length class 15 (95% confidence level) and in class ALL (90% confidence level). In contrast, for the group of females, the null hypothesis is not rejected for length classes 12, 15, 17 (95% confidence level) and in class ALL (90% confidence level). For both tests, we obtained similar decisions on each time series.
According to Contreras-Reyes [18], the time series in which the shape parameter is close to zero or when the null hypothesis is not rejected are influenced simultaneously by both normal and extreme events as in the length class ALL, where all the fish population is included for the analysis. For length class 17 in males, for example, the CF is susceptible to some atypical events such as the moderate-strong El Niño event between 1991 and 1992 (high negative empirical skewness and high empirical kurtosis). For length class 13 in both sexes, the CF is susceptible to the strong El Niño event produced between 1997 and 1998.

7. Discussion

We have presented the methodology to compute the Shannon entropy, the negentropy and the Kullback–Leibler and Jeffrey’s divergences for a broad family of asymmetric distributions with the normal kernel called generalized skew-normal distributions. Our method considers asymptotic expansions regarding moments and cumulants for two particular cases: the skew-normal and modified skew-normal distributions. We then measured the degrees of disparity of these distributions from the normal distribution by using exact expressions for the negentropy in terms of moments and cumulants. Additionally, given the regularity conditions accomplished by the modified skew-normal distribution, normality was tested based on the modified skew-normal distribution. This test considered the asymptotic behavior of the Kullback–Leibler divergence, which is determined by the negentropy for normality disparity.
Numerical results showed that the Shannon entropy and negentropy of the modified skew-normal distribution are better approximated than the skew-normal one, at least for a wider range of the shape parameter. For small to moderate values of the asymmetry parameter, where the approximations are appropriate, we find that expansions series converge from the fourth moment/cumulant to greater, as in the Gram–Charlier and Edgeworth expansion methods [17]. For large values of the skewness parameter, where the expansions are inappropriate, the functions related to negentropy are not well approximated by Taylor expansions around zero, produced by a divergence in the moment and cumulant terms, i.e., the Taylor expansions for the expected values of the functions ζ 0 ( τ Z τ ) and ζ 0 { τ u ( Z τ ) } (SN and MSN case, respectively) if τ = | η | is too large. When this happens, the normal cdf, Φ ( τ Z τ ) and Φ { τ u ( Z τ ) } (SN and MSN case, respectively), tends to one, since according to the stochastic representation in (12), for large values of τ , the distribution of Z τ converges to the standardized half-normal distribution [37].
However, the normality test considered in the application used skewness parameters inside the appropriate range. Furthermore, we plan to investigate the negentropy of the modified skew-normal-Cauchy distribution or similar models. In addition, although the approximations are appropriate over the range of variation of the asymmetry admitted by both models, more work should be done in order to improve the asymptotic approximations for a greater range of the skewness parameter values. Besides, this is not an easy task since generally it is difficult to approximate KL divergences involving asymmetric and heavy-tailed distributions [38].
The statistical application related to condition factor time series of anchovies off northern Chile is given. The results show that the proposed methodology serves to detect non-normal events in these time series, which produces an empirical distribution with high presence of skewness [18]. The proposed test for normality is therefore useful to detect anomalies in condition factor time series, linked to food deficit (positive shape parameter) or food abundance (negative shape parameter) influenced by environmental conditions.

Acknowledgments

We are grateful to the Instituto de Fomento Pesquero (IFOP) for providing access to the data used in this work. The authors also thank Jaime Arrué for providing useful R codes of the MSN distribution. We are sincerely grateful to the four anonymous reviewers for their comments and suggestions that greatly improved an early version of this manuscript. Arellano-Valle’s research was supported by Grant FONDECYT (Chile) Nos. 1120121 and 1150325. Contreras-Reyes’s research was supported by Grant PIIC 069/2016 from UTFSM (Valparaíso, Chile) and by the CONICYT doctoral scholarship 2016 No. 21160618. Stehlík’s research was supported by FONDECYT (Chile) No. 1151441 and LIT-2016-1-SEE-023. All R codes used in this paper are available by request to the corresponding author.

Author Contributions

Javier E. Contreras-Reyes conceived the experiments and analyzed the data; Reinaldo B. Arellano-Valle, Javier E. Contreras-Reyes and Milan Stehlík designed and performed the experiments, contributed reagents/analysis tools and wrote the paper; and Javier E. Contreras-Reyes contributed materials tools. All authors have read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1. Moments of the Skew-Normal Distribution.

The moments μ k = E ( Z τ k ) are given by:
μ 2 k = ( 2 k 1 ) ! ! = ( 2 k ) ! 2 k k ! ( even moments ) , μ 2 k 1 = ( 2 k 1 ) ! b τ 2 k 1 ( 1 + τ 2 ) ( 2 k 1 ) / 2 j = 0 k 1 j ! ( 2 τ ) 2 j ( 2 j + 1 ) ! ( k j 1 ) ! ( odd moments ; [ 26 ] ) .
From Proposition 2 in Martínez et al. [39], the odd moments can also be computed as:
μ 2 k 1 = b m = 1 k a k ( m ) τ 2 m 1 ( 1 + τ 2 ) ( 2 k 1 ) / 2 ,
where the coefficient a k ( m ) is computed iteratively as follows:
a 1 ( 1 ) = 1 , a k ( 1 ) = ( 2 k 1 ) a k 1 ( 1 ) , k 2 , a k ( m ) = 2 ( k 1 ) { a k 1 ( m ) + a k 1 ( m 1 ) } , 1 < m < k , k 2 , a k ( k ) = 2 ( k 1 ) a k 1 ( k 1 ) , k 2 .

Appendix A.2. Cumulants of the Skew-Normal Distribution

The coefficients κ k = ζ k ( 0 ) , k = 1 , 2 , , are related to the cumulants of the half-normal random variable V χ 1 2 given by K ( t ) = ( 1 / 2 ) t 2 + ζ 0 ( t ) (see also [21,23]). Let K m ( 0 ) = d m d t m K t t = 0 be the m-th cumulant of V and clearly K 1 ( 0 ) = ζ 1 ( 0 ) = κ 1 , K 2 ( 0 ) = 1 + ζ 2 ( 0 ) = 1 + κ 2 , and K m ( 0 ) = ζ m ( 0 ) = κ m , m = 3 , 4 , . Furthermore, from [21,23], it emerges that:
ζ 1 ( x ) = ϕ ( x ) Φ ( x ) , ζ 2 ( x ) = ζ 1 ( x ) { x + ζ 1 ( x ) } , ζ 3 ( x ) = ζ 2 ( x ) { x + ζ 1 ( x ) } ζ 1 ( x ) { 1 + ζ 2 ( x ) } , ζ 4 ( x ) = ζ 3 ( x ) { x + ζ 1 ( x ) } 2 ζ 2 ( x ) { 1 + ζ 2 ( x ) } ζ 1 ( x ) ζ 3 ( x ) , ζ 5 ( x ) = ζ 4 ( x ) { x + ζ 1 ( x ) } 3 ζ 3 ( x ) { 1 + ζ 2 ( x ) } 3 ζ 2 ( x ) ζ 3 ( x ) ζ 1 ( x ) ζ 4 ( x ) ,
Recalling that b = 2 / π , the first five coefficients κ k = ζ k ( 0 ) are:
κ 1 = b , κ 2 = κ 1 2 , κ 3 = 2 κ 2 κ 1 κ 1 , κ 4 = 2 κ 3 κ 1 2 κ 2 2 2 κ 2 , κ 5 = 2 κ 4 κ 1 6 κ 3 κ 2 3 κ 3 ,
Thus, by letting κ 0 = 1 , a recursive rule for these coefficients is obtained as follows:
κ 1 = b , κ 2 k = ( 2 k 2 ) κ 2 k 2 2 i = 1 k 2 k 2 i 1 κ i κ 2 k i + 2 k 2 k 1 κ k 2 , κ 2 k + 1 = ( 2 k 1 ) κ 2 k 1 2 i = 1 k 2 k 1 i 1 κ i κ 2 k i + 1 , k = 1 , 2 , .

Appendix A.3. Odd Moments of the Modified Skew-Normal Distribution

Recalling that b = 2 / π , the odd moments can be computed as:
μ 2 k 1 ( τ ) = 2 k 1 ( k 1 ) ! b 2 ξ k ( τ ) 1 ,
with:
ξ k ( τ ) = 0 Φ τ u ( x ) x k 1 e x / 2 2 k Γ ( k ) d x , k = 1 , 2 , ,
where Γ ( · ) denotes the usual gamma function. Note that ξ k ( τ ) = E Φ τ u ( X k ) , where X k χ 2 k 2 ; thus, 0 < 2 ξ k ( τ ) 1 < 1 for all k = 1 , 2 , and τ > 0 . In particular, the first four moments are μ 1 ( τ ) = b { 2 ξ 1 ( τ ) 1 } , μ 2 = 1 , μ 3 ( τ ) = 2 b { 2 ξ 2 ( τ ) 1 } , and μ 4 = 3 .

Appendix A.4. Likelihood Radio Test

The Likelihood Radio Test (LRT) statistic [40] for a null hypotheses H 0 : θ Θ 0 versus H 1 : θ Θ 0 , Θ 0 Θ (the parametric space), is given by:
L R T = 2 { ( θ ^ ) ( θ ^ 0 ) } ,
where θ ^ is the unrestricted MLE of θ , θ ^ 0 is the MLE of θ under H 0 and the log-likelihood function ( θ ) for MSN distributions is presented in (20). As before, normality is typically obtained from the GSN class at η 0 = 0 . Because the MSN distribution satisfies the standard regularity conditions [15], the LRT statistic is asymptotically χ s 2 distributed under H 0 , with s = dim ( Θ ) dim ( Θ 0 ) = 1 degrees of freedom [41]. Hence, the p-value associated with the LRT is computed as 1 χ s 2 ( LRT ) , where χ s 2 ( LRT ) denotes the χ s 2 -distribution function evaluated at the observed value of the LRT statistic.
In order to test normality, we considered the particular null hypothesis H 0 : η = 0 versus H 1 : η 0 , with the rest of the parameters not specified. Therefore, by (20), the LRT statistic is given by:
L R T = 2 1 2 i = 1 n z i 2 ^ + 1 2 i = 1 n z i 0 2 ^ + i = 1 n log ζ 0 { η ^ u ( z ^ i ) } ,
where η ^ is the unrestricted MLE of η , z i ^ and z ^ i 0 are the unrestricted and restricted MLE of z i = ( y i μ ) / σ , respectively; and the p-value is computed as 1 χ 1 2 ( LRT ) .

References

  1. Stehlík, M.; Střelec, L.; Thulin, M. On robust testing for normality in chemometrics. Chemom. Intell. Lab. Syst. 2014, 130, 98–108. [Google Scholar] [CrossRef]
  2. Vidal, I.; Iglesias, P.; Branco, M.D.; Arellano-Valle, R.B. Bayesian Sensitivity Analysis and Model Comparison for Skew Elliptical Models. J. Stat. Plan. Inference 2006, 136, 3435–3457. [Google Scholar] [CrossRef]
  3. Stehlík, M. Distributions of exact tests in the exponential family. Metrika 2003, 57, 145–164. [Google Scholar] [CrossRef]
  4. Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  5. Stehlík, M. Decompositions of information divergences: Recent development, open problems and applications. In Proceedings of the 9th International Conference on Mathematical Problems in Engineering, Aerospace and Sciences: ICNPAA 2012, Vienna, Austria, 10–14 July 2012; Volume 1493, pp. 972–976. [Google Scholar]
  6. Contreras-Reyes, J.E.; Arellano-Valle, R.B. Kullback–Leibler divergence measure for Multivariate Skew-Normal Distributions. Entropy 2012, 14, 1606–1626. [Google Scholar] [CrossRef]
  7. Jeffreys, H. An invariant form for the prior probability in estimation problems. Proc. R. Soc. Lond. A 1946, 186, 453–461. [Google Scholar] [CrossRef]
  8. Gómez-Villegas, M.A.; Main, P.; Navarro, H.; Susi, R. Assessing the effect of kurtosis deviations from Gaussianity on conditional distributions. Appl. Math. Comput. 2013, 219, 10499–10505. [Google Scholar] [CrossRef]
  9. Main, P.; Arevalillo, J.M.; Navarro, H. Local Effect of Asymmetry Deviations from Gaussianity Using Information-Based Measures. In Proceedings of the 2nd International Electronic Conference on Entropy and Its Applications, Santa Barbara, CA, USA, 15–30 November 2015. [Google Scholar]
  10. Dette, H.; Ley, C.; Rubio, F.J. Natural (non-) informative priors for skew-symmetric distributions. Scand. J. Stat. 2017, in press. [Google Scholar]
  11. Gao, J.H.; Zhang, B. Estimation of seismic wavelets based on the multivariate scale mixture of Gaussians model. Entropy 2009, 12, 14–33. [Google Scholar] [CrossRef]
  12. Wang, Z.; Zhang, B.; Gao, J. The residual phase estimation of a seismic wavelet using a Rényi divergence-based criterion. J. Appl. Geophys. 2014, 106, 96–105. [Google Scholar] [CrossRef]
  13. Pires, C.A.; Ribeiro, A.F. Separation of the atmospheric variability into non-Gaussian multidimensional sources by projection pursuit techniques. Clim. Dyn. 2017, 48, 821–850. [Google Scholar] [CrossRef]
  14. Pires, C.A.; Hannachi, A. Independent Subspace Analysis of the Sea Surface Temperature Variability: Non-Gaussian Sources and Sensitivity to Sampling and Dimensionality. Complexity 2017. [Google Scholar] [CrossRef]
  15. Arrué, J.; Arellano-Valle, R.B.; Gómez, H.W. Bias reduction of maximum likelihood estimates for a modified skew-normal distribution. J. Stat. Comput. Simul. 2016, 86, 2967–2984. [Google Scholar] [CrossRef]
  16. Arellano-Valle, R.B.; Contreras-Reyes, J.E.; Genton, M.G. Shannon entropy and mutual information for multivariate skew-elliptical distributions. Scand. J. Stat. 2013, 40, 42–62. [Google Scholar] [CrossRef]
  17. Hyvärinen, A.; Karhunen, J.; Oja, E. Independent Component Analysis; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2001. [Google Scholar]
  18. Contreras-Reyes, J.E. Analyzing fish condition factor index through skew-gaussian information theory quantifiers. Fluct. Noise Lett. 2016, 15, 1650013. [Google Scholar] [CrossRef]
  19. Withers, C.S.; Nadarajah, S. Negentropy as a function of cumulants. Inf. Sci. 2014, 271, 31–44. [Google Scholar] [CrossRef]
  20. Azzalini, A. A Class of Distributions which includes the Normal Ones. Scand. J. Stat. 1985, 12, 171–178. [Google Scholar]
  21. Azzalini, A.; Capitanio, A. Statistical applications of the multivariate skew normal distributions. J. R. Stat. Soc. Ser. B 1999, 61, 579–602. [Google Scholar] [CrossRef]
  22. Gupta, A.K.; Chang, F.C.; Huang, W.J. Some skew-symmetric models. Random Oper. Stoch. Equ. 2002, 10, 133–140. [Google Scholar] [CrossRef]
  23. Azzalini, A.; Capitanio, A. The Skew-Normal and Related Families; Cambridge University Press: Cambridge, UK, 2013; Volume 3. [Google Scholar]
  24. Nadarajah, S.; Kotz, S. Skewed distributions generated by the normal kernel. Stat. Probab. Lett. 2003, 65, 269–277. [Google Scholar] [CrossRef]
  25. Arrué, J.; Gómez, H.W.; Varela, H.; Bolfarine, H. On the skew-normal-Cauchy distribution. Commun. Stat. A Theory. 2010, 40, 15–27. [Google Scholar] [CrossRef]
  26. Henze, N. A probabilistic representation of the ‘skew-normal’ distribution. Scand. J. Stat. 1986, 13, 271–275. [Google Scholar]
  27. Salicrú, M.; Menéndez, M.L.; Pardo, L.; Morales, D. On the applications of divergence type measures in testing statistical hypothesis. J. Multivar. Anal. 1994, 51, 372–391. [Google Scholar] [CrossRef]
  28. Arellano-Valle, R.B.; Azzalini, A. The centred parametrization for the multivariate skew-normal distribution. J. Multivar. Anal. 2009, 100, 816. [Google Scholar] [CrossRef]
  29. Chiogna, M. Some results on the scalar skew-normal distribution. J. Ital. Stat. Soc. 1998, 1, 1–13. [Google Scholar] [CrossRef]
  30. Kupperman, M. Further Applications of Information Theory to Multivariate Analysis and Statistical Inference. Ph.D. Thesis, George Washington University, Washington, DC, USA, January 1957; p. 270. [Google Scholar]
  31. Piessens, R.; deDoncker-Kapenga, E.; Uberhuber, C.; Kahaner, D. Quadpack: A Subroutine Package for Automatic Integration; Springer: Berlin, Germany, 1983. [Google Scholar]
  32. R Core Team. A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2015; Available online: http://www.R-project.org (accessed on 1 October 2016).
  33. Contreras-Reyes, J.E. Rényi entropy and complexity measure for skew-gaussian distributions and related families. Physica A 2015, 433, 84–91. [Google Scholar] [CrossRef]
  34. Le Cren, E.D. The length–weight relationship and seasonal cycle in gonad weight and condition in the perch (Perca fluviatilis). J. Anim. Ecol. 1951, 20, 201–219. [Google Scholar] [CrossRef]
  35. Contreras-Reyes, J.E.; Cortés, D.D. Bounds on Rényi and Shannon Entropies for Finite Mixtures of Multivariate Skew-Normal Distributions: Application to Swordfish (Xiphias gladius Linnaeus). Entropy 2016, 18, 382. [Google Scholar] [CrossRef]
  36. Azzalini, A. R package sn: The Skew-Normal and Skew-t Distributions (version 0.4-6); Università di Padova’: Padua, Italy, 2010. [Google Scholar]
  37. Arnold, B.C.; Beaver, R.J. Hidden truncation models. Sankhya A 2000, 62, 23–35. [Google Scholar]
  38. Stehlík, M.; Somorčík, J.; Střelec, L.; Antoch, J. Approximation of information divergences for statistical learning with applications. Math. Slovaca 2017, in press. [Google Scholar]
  39. Martínez, E.H.; Varela, H.; Gómez, H.W.; Bolfarine, H. A note on the likelihood and moments of the skew-normal distribution. Stat. Oper. Res. Trans. 2008, 32, 57–66. [Google Scholar]
  40. Chernoff, H. On the distribution of the likelihood ratio. Ann. Math. Stat. 1954, 25, 573–578. [Google Scholar] [CrossRef]
  41. Azzalini, A.; Arellano-Valle, R.B. Maximum penalized likelihood estimation for skew-normal and skew-t distributions. J. Stat. Plan. Inference 2013, 143, 419–433. [Google Scholar] [CrossRef]
Figure 1. Shannon entropy and negentropy for the (A,B) Skew-Normal (SN) and (C,D) Modified Skew-Normal (MSN) cases. The blue and red lines correspond to numerical integration and cumulant expansion series methods, respectively.
Figure 1. Shannon entropy and negentropy for the (A,B) Skew-Normal (SN) and (C,D) Modified Skew-Normal (MSN) cases. The blue and red lines correspond to numerical integration and cumulant expansion series methods, respectively.
Entropy 19 00528 g001aEntropy 19 00528 g001b
Figure 2. Kullback-Leibler (KL) divergence, K 0 = K ( Z τ , Z 0 ) , between SN and normal (solid line) and MSN and normal (dotted line).
Figure 2. Kullback-Leibler (KL) divergence, K 0 = K ( Z τ , Z 0 ) , between SN and normal (solid line) and MSN and normal (dotted line).
Entropy 19 00528 g002
Table 1. Observed power (both in %) of the proposed normality test using the maximum likelihood estimator (MLE) of the MSN model from 2000 simulations for the nominal level of 5% and various values of shape parameter η and sample size n.
Table 1. Observed power (both in %) of the proposed normality test using the maximum likelihood estimator (MLE) of the MSN model from 2000 simulations for the nominal level of 5% and various values of shape parameter η and sample size n.
n/ η 00.10.20.30.40.81.21.62
5097.5537.0239.4537.3937.6040.0740.8248.3751.29
10097.6546.1545.9047.9046.0550.0556.1066.2075.49
20098.0554.2554.3053.6554.5058.8070.5585.2093.25
30098.3057.3057.7058.0058.6562.0077.0592.6598.25
40098.5557.2057.6557.1557.5064.1083.9595.8099.65
50098.7058.5559.4059.9558.4566.7086.9597.4599.75
Table 2. Shape parameter estimates ( η ^ ) of SN (reported in [18]) and MSN models for each sex and length class L, together with its respective standard deviations (s.d). Sample size (n), empirical skewness ( b ^ 1 ) and kurtosis ( b ^ 2 ), as well as the log-likelihood function ( η ^ ) for each model fit are also reported.
Table 2. Shape parameter estimates ( η ^ ) of SN (reported in [18]) and MSN models for each sex and length class L, together with its respective standard deviations (s.d). Sample size (n), empirical skewness ( b ^ 1 ) and kurtosis ( b ^ 2 ), as well as the log-likelihood function ( η ^ ) for each model fit are also reported.
SexLn b ^ 1 b ^ 2 SNMSN
η ^ s.d ( η ^ ) η ^ s.d ( η ^ )
Male12213−0.2203.723−1.0650.147−301.066−1.7170.662−235.783
132380.6585.1201.7780.134−332.6382.6170.447−359.291
14260−0.1472.885−1.0860.166−367.9350.4590.542−342.881
15261−0.0302.755−0.4420.150−369.8340.0650.386450.489
162610.3073.1381.6160.149−367.5422.5650.415−354.073
17221−3.46127.958−3.0920.080−285.125−3.5970.524−312.701
181800.2643.0011.3680.201−253.9882.8210.643−269.193
All2620.0682.6870.7210.178−371.192−0.2870.517−131.335
Female121980.0412.7380.5520.196−280.434−0.1420.386−209.551
132280.9176.1032.2670.128−315.3053.1010.524−326.665
142600.1902.9071.2420.157−367.5551.4590.867−331.373
152600.0762.6720.7280.174−368.344−0.1890.383−425.212
162610.3493.0911.7020.155−367.1602.6310.475−346.705
17246−0.0563.115−0.6890.149−348.4870.0410.754−373.068
18208−0.5394.349−1.5810.136−291.484−2.2230.472−313.160
All2620.0722.7640.7480.172−371.174−0.2670.401−110.108
Table 3. MSN Shannon entropy (H) and negentropy (N) for each sex and length class L using expansion series of cumulants. For each time series, the KL divergence K 0 = K ( Z τ ^ , Z 0 ) , statistic 2 n K 0 of Equation (18), the Likelihood Ratio Test (LRT) statistic and its respective p-values are reported. All values reported consider estimates η ^ (for τ ^ = | η ^ | ) and sample size n from Table 2.
Table 3. MSN Shannon entropy (H) and negentropy (N) for each sex and length class L using expansion series of cumulants. For each time series, the KL divergence K 0 = K ( Z τ ^ , Z 0 ) , statistic 2 n K 0 of Equation (18), the Likelihood Ratio Test (LRT) statistic and its respective p-values are reported. All values reported consider estimates η ^ (for τ ^ = | η ^ | ) and sample size n from Table 2.
SexLHN K 0 Asymptotic TestLRT
2 nK 0 p-ValueStatisticp-Value
Male121.187−0.0740.23298.950<0.00198.304<0.001
131.030−0.0630.389185.087<0.001180.469<0.001
141.396−0.0080.02311.6990.00111.6930.001
151.4180.0000.0010.2420.6230.2460.620
161.038−0.0660.381198.700<0.001193.567<0.001
170.8730.0500.546241.472<0.001211.924<0.001
180.999−0.0470.420151.341<0.001146.427<0.001
All1.410−0.0030.0094.6820.0314.6730.030
Female121.417−0.0010.0020.8730.3500.9030.342
130.956−0.0190.463211.320<0.001198.516<0.001
141.237−0.0620.18394.892<0.00193.300<0.001
151.415−0.0010.0042.0270.1552.0710.150
161.028−0.0620.391204.117<0.001199.015<0.001
171.4180.0000.00020.0910.7630.0890.765
181.095−0.0800.324134.967<0.001133.629<0.001
All1.411−0.0030.0084.0580.0444.1120.042

Share and Cite

MDPI and ACS Style

Arellano-Valle, R.B.; Contreras-Reyes, J.E.; Stehlík, M. Generalized Skew-Normal Negentropy and Its Application to Fish Condition Factor Time Series. Entropy 2017, 19, 528. https://doi.org/10.3390/e19100528

AMA Style

Arellano-Valle RB, Contreras-Reyes JE, Stehlík M. Generalized Skew-Normal Negentropy and Its Application to Fish Condition Factor Time Series. Entropy. 2017; 19(10):528. https://doi.org/10.3390/e19100528

Chicago/Turabian Style

Arellano-Valle, Reinaldo B., Javier E. Contreras-Reyes, and Milan Stehlík. 2017. "Generalized Skew-Normal Negentropy and Its Application to Fish Condition Factor Time Series" Entropy 19, no. 10: 528. https://doi.org/10.3390/e19100528

APA Style

Arellano-Valle, R. B., Contreras-Reyes, J. E., & Stehlík, M. (2017). Generalized Skew-Normal Negentropy and Its Application to Fish Condition Factor Time Series. Entropy, 19(10), 528. https://doi.org/10.3390/e19100528

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop