Next Article in Journal
Measures of Qualitative Variation in the Case of Maximum Entropy
Previous Article in Journal
Divergence and Sufficiency for Convex Optimization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Objective Bayesian Entropy Inference for Two-Parameter Logistic Distribution Using Upper Record Values

1
Department of Statistics, Daejeon University, Daejeon 34520, Korea
2
Department of Statistics, Kyungpook National University, Daegu 41566, Korea
*
Author to whom correspondence should be addressed.
Entropy 2017, 19(5), 208; https://doi.org/10.3390/e19050208
Submission received: 24 March 2017 / Revised: 28 April 2017 / Accepted: 29 April 2017 / Published: 3 May 2017

Abstract

:
In this paper, we provide an entropy inference method that is based on an objective Bayesian approach for upper record values having a two-parameter logistic distribution. We derive the entropy that is based on the i-th upper record value and the joint entropy that is based on the upper record values. Moreover, we examine their properties. For objective Bayesian analysis, we obtain objective priors, namely, the Jeffreys and reference priors, for the unknown parameters of the logistic distribution. The priors are based on upper record values. Then, we develop an entropy inference method that is based on these objective priors. In real data analysis, we assess the quality of the proposed models under the objective priors and compare them with the model under the informative prior.

1. Introduction

Shannon [1] proposed information theory for quantifying information loss and introduced statistical entropy. Baratpour et al. [2] obtained the entropy of a continuous probability distribution using upper record values. Moreover, they obtained several bounds for this entropy using the hazard rate function. Abo-Eleneen [3] suggested an efficient computation method for the entropy in progressively Type-II censored samples. Kang et al. [4], using maximum likelihood estimators (MLE) and approximate MLE (AMLE), derived estimators of the entropy of a double-exponential distribution that are based on multiply Type-II censored samples. Seo and Kang [5], using estimators of the shape parameter in the generalized half-logistic distribution, developed methods for estimating entropy that are based on Type-II censored samples.
In this paper, we provide an entropy inference method that is based on an objective Bayesian approach for upper record values having the two-parameter logistic distribution. The cumulative distribution function (cdf) and probability density function (pdf) of a random variable X with this distribution are given by
F ( x ) = 1 1 + e ( x μ ) / σ
and
f ( x ) = e ( x μ ) / σ σ 1 + e ( x μ ) / σ 2 , x R , μ R , σ > 0 ,
respectively, where μ is the location parameter and σ is the scale parameter.
The paper is organized as follows: In Section 2, we obtain the Jeffreys and reference priors and derive an entropy inference method that is based on the two non-informative priors. In Section 3, we analyze a real data set in order to demonstrate the validity of the proposed method. Section 4 concludes this paper.

2. Objective Bayesian Analysis

2.1. Entropy

The entropy of f ( x ) is defined by
H ( f ) = f ( x ) log f ( x ) d x .
Then, the entropy based on the i-th upper record value X U ( i ) is
H U ( i ) = f X U ( i ) ( x ) log f X U ( i ) ( x ) d x ,
where f X U ( i ) ( x ) is the marginal density function of X U ( i ) , defined as
f X U ( i ) ( x ) = 1 Γ ( i ) log ( 1 F ( x ) ) i 1 f ( x ) .
Assuming that X U ( i ) is the i-th upper record value from the logistic distribution with pdf f ( x ) as in (1), the marginal density function (2) is given by
f X U ( i ) ( x ) = 1 Γ ( i ) x μ σ + log 1 + e ( x μ ) / σ i 1 e ( x μ ) / σ σ 1 + e ( x μ ) / σ 2 .
Then, the corresponding entropy is given by
H U ( i ) = log Γ ( i ) + log σ + i ( i 1 ) ψ ( i ) + j = 1 1 j ( j + 1 ) i .
This only depends on the scale parameter σ and it is clear that it is an increasing function of σ . Therefore, as σ increases, less information is provided by the distribution.
Remark 1.
We can obtain the following relationship between the entropies corresponding to two consecutive record times:
lim i H U ( i ) H U ( i 1 ) = lim i log ( i 1 ) ψ ( i 1 ) + j = 1 1 j ( j + 1 ) i 1 j ( j + 1 ) i 1 = lim i j = 0 1 i 1 + j log 1 + 1 i 1 + j = 0 .
Theorem 1.
The joint entropy that is based on X U ( 1 ) , , X U ( k ) from the logistic distribution with pdf as in (1) is
H U ( 1 ) , , U ( k ) = k 1 + log σ + i = 1 k j = 1 1 j ( j + 1 ) i .
This is an increasing function of σ, as is the case with H U ( i ) .
Proof. 
The joint entropy based on the upper record values X U ( 1 ) , , X U ( k ) is defined by Park [6] as
H U ( 1 ) , , U ( k ) = x U ( 2 ) f X U ( 1 ) , , X U ( k ) ( x U ( 1 ) , , x U ( k ) ) × log f X U ( 1 ) , , X U ( k ) ( x U ( 1 ) , , x U ( k ) ) d x U ( 1 ) , , d x U ( k ) ,
where f X U ( 1 ) , , X U ( k ) ( x U ( 1 ) , , x U ( k ) ) is the joint density function of X U ( 1 ) , , X U ( k ) . In addition, it is simplified to a single integral by Rad et al. [7] as follows
H U ( 1 ) , , U ( k ) = k ( 1 k ) 2 i = 1 k 1 Γ ( i ) log ( 1 F ( x ) ) i 1 f ( x ) log f ( x ) d x .
Let X U ( 1 ) , , X U ( k ) be the upper record values from the logistic distribution with pdf as in (1) and
Y = X μ σ + log 1 + e ( X μ ) / σ .
Then, the integral term in (5) is given by
1 Γ ( i ) log ( 1 F ( x ) ) i 1 f ( x ) log f ( x ) d x = log σ 2 0 y i e y d y + 0 y i 1 e y log e y 1 d y .
Finally, using the series expansion
log z z 1 = j = 1 1 j z j , z 1   or   z > 1 ,
we can complete the proof.  ☐
Remark 2.
The entropies (3) and (4) can take negative values because of the term log σ . This is because the marginal density function f X U ( i ) ( x ) and the joint density function f X U ( 1 ) , , X U ( k ) ( x U ( 1 ) , , x U ( k ) ) can have values greater than one for very small σ.
We present the values of the entropies H U ( i ) and H U ( 1 ) , , U ( k ) for various values of σ , i and k in Table 1 and Table 2 and Figure 1.
Table 1 shows that H U ( i ) is an increasing function of σ for fixed i. Symmetrically, it is an increasing function of i for fixed σ and i 3 . Likewise, Table 2 shows that H U ( 1 ) , , U ( k ) increases as σ and k increase, except for σ = 0.1 .
We note that σ in (3) and (4) is an unknown parameter. Thus, it should be estimated when the upper record values are observed. The following theorem provides an estimator of the joint entropy H U ( 1 ) , , U ( k ) in the Bayesian framework.
Theorem 2.
The Bayes estimator of H U ( 1 ) , , U ( k ) is
H ^ U ( 1 ) , , U ( k ) B = k 1 + E π | x log σ + i = 1 k j = 1 1 j ( j + 1 ) i ,
where the posterior expectation E π | x · exists and is finite.
Proof. 
In the Bayesian framework, the entropy estimator that is based on X U ( 1 ) , , X U ( k ) is defined as
H ^ U ( 1 ) , , U ( k ) B = μ σ H U ( 1 ) , , U ( k ) π ( μ , σ | x ) d μ d σ .
Then, the estimator is given by
H ^ U ( 1 ) , , U ( k ) B = k + k μ σ log σ π ( μ , σ | x ) d μ d σ + i = 1 k j = 1 1 j ( j + 1 ) i .
This completes the proof.  ☐
In the following subsection we will provide a method for obtaining the term E π | x log σ in (6).

2.2. Posterior Analysis Based on Objective Priors

Asgharzadeh et al. [8] proposed a subjective prior distribution for π ( μ , σ ) as follows
π I B ( μ , σ ) exp ( b + μ μ 0 ) / σ σ a + 2 1 + exp ( μ μ 0 ) / σ 2 , μ 0 R , a , b > 0 .
If one has sufficient prior information, the hyperparameters a , b and μ 0 can be easily determined; otherwise one should depend on objective or non-informative priors. In fact, it is not easy to elicit suitable prior information. We will not consider a method for eliciting the values of the hyperparameters, but rather an inference method that is based on objective priors. We will now obtain objective priors (the Jeffreys and reference priors) that are based on the Fisher information matrix for ( μ , σ ) . See [8].
Let X U ( 1 ) , , X U ( k ) be the upper record values of X 1 , , X n from the logistic distribution with pdf as in (1). Then, the corresponding likelihood function is given by
L ( μ , σ ) = f x U ( k ) i = 1 k 1 f x U ( i ) 1 F x U ( i ) = 1 σ k exp x U ( k ) μ / σ 1 + exp x U ( k ) μ / σ i = 1 k 1 1 + exp x U ( i ) μ / σ .
In addition, the Fisher information matrix for ( μ , σ ) is given by
I ( μ , σ ) = E 2 μ 2 log L ( μ , σ ) E 2 μ σ log L ( μ , σ ) E 2 σ μ log L ( μ , σ ) E 2 σ 2 log L ( μ , σ )
By the result in [8], all elements of the Fisher information matrix are proportional to 1 / σ 2 . Therefore, the Jeffreys prior is
π J ( μ , σ ) 1 σ 2 ,
since it is proportional to the square root of the determinant of the Fisher information matrix. However, the Jeffreys prior has some drawbacks in the multi-parameter case, such as the marginalization paradox and the Neyman-Scott problem. Alternatively, Bernardo [9] introduced the reference prior. Moreover, Berger et al. [10,11] provided a general algorithm for deriving the reference prior. Using this algorithm, we can obtain the reference prior π R ( μ , σ ) as follows
π R ( μ , σ ) 1 σ ,
regardless of which parameter is of interest.
Under a joint prior π ( μ , σ ) , the resulting posterior distribution is
π ( μ , σ | x ) π ( μ , σ ) 1 σ k exp x U ( k ) μ / σ 1 + exp x U ( k ) μ / σ i = 1 k 1 1 + exp x U ( i ) μ / σ .
Unfortunately, it is impossible to express in closed forms the marginal distribution for μ and σ under the derived priors (7) and (8). In order to generate Markov chain Monte Carlo (MCMC) samples from the marginal distributions, it is necessary to obtain the full conditional posterior distributions for each parameter under the joint prior π ( μ , σ ) as follows
π ( μ | σ , x ) π ( μ , σ ) exp μ / σ 1 + exp x U ( k ) μ / σ i = 1 k 1 1 + exp x U ( i ) μ / σ
and
π ( σ | μ , x ) π ( μ , σ ) 1 σ k exp x U ( k ) μ / σ 1 + exp x U ( k ) μ / σ i = 1 k 1 1 + exp x U ( i ) μ / σ ,
respectively.
Under both objective priors (7) and (8), the full conditional posterior distributions for μ are log-concave. Therefore, we can draw the MCMC samples μ i ( i = 1 , , N ) from these conditional posterior distributions using the method proposed by [12]. Moreover, we note that σ R + , whereas μ R and X U ( i ) R . Thus, it is not easy to find a suitable proposal distribution for drawing the MCMC samples σ i ( i = 1 , , N ) from the full conditional posterior distribution π ( σ | μ , x ) . Therefore, we employ the random-walk Metropolis algorithm that is based on a normal proposal distribution truncated at zero. Using the MCMC samples, the term E π | x log σ in (6) can be approximated as follows
E π | x log σ 1 N M i = M + 1 N log σ i ,
where M is the number of burn-in samples.
The following section examines the validity of the provided objective Bayesian method by analyzing a real data set.

3. Application

Asgharzadeh et al. [8] analyzed the upper record values 2.70 , 3.78 , 4.83 , 8.02 , 8.37 from the total annual rainfall (in inches) during March that was recorded at Los Angeles Civic Center from 1973 to 2006. To obtain Bayes estimates under the subjective prior π I ( μ , σ ) , we use the same values that [8] used (i.e., a = b = 0.00001 and μ 0 = 0 ). The MCMC samples are generated using the algorithm that is described in Section 2.1. To obtain the optimal acceptance rate under priors (7) and (8), the variances in a truncated normal proposal are set to 0.7 and 0.8, respectively [13]. Based on 5500 MCMC samples with 500 brun-in samples, the Bayes estimates under the square error loss function and the corresponding 95% HPD CrIs are computed in order to compare the MLE. The results are presented in Table 3. To verify the validity of the MCMC samples, we present their autocorrelation functions (ACF) and trace plots in Figure 2 and Figure 3.
From Figure 2 and Figure 3, we can see that the MCMC samples are mixing and converge to the stationary distribution well.
Table 3 shows that the length of the HPD CrIs under the objective priors π J ( μ , σ ) and π R ( μ , σ ) is smaller than it is under the subjective prior π I ( μ , σ ) with a = b = 0.00001 and μ 0 = 0 .
Furthermore, we assess the quality of the Bayesian models under priors (7) and (8) based on the replications X U ( i ) r e p ( i = 1 , , 5 ) of the observed upper record values from the posterior predictive distributions that are given by
f X r e p J ( x r e p | x ) = μ σ f X r e p ( x r e p | μ , σ ) π J ( μ , σ | x ) d μ d σ
and
f X r e p R ( x r e p | x ) = μ σ f X r e p ( x r e p | μ , σ ) π R ( μ , σ | x ) d μ d σ ,
where f X r e p ( x r e p ) is the marginal density function of X r e p . The replications are obtained as follows
X U ( i ) r e p = 1 N M j = M + 1 N X U ( i ) r e p ( j ) , i = 1 , , k ,
where X U ( i ) r e p ( j ) is a sample from the marginal density function f X r e p · ( x r e p ) . The replications and their mean and standard deviation (std) are given in Table 4. The mean and standard deviation (std) of the observed upper record values are 5.54 and 2.541, respectively.
The model under the Jeffreys prior (7) exhibits better performance with respect to the replications X U ( i ) r e p ( i = 1 , 2 , 3 , 5 ) and the mean, whereas, under the reference prior (8), it exhibits better performance with respect to the replication X U ( 4 ) r e p . However, there is no significant difference between the replications under the priors.
Finally, we present the estimation results for the joint entropy H ^ U ( 1 ) , , U ( k ) B under the subjective prior π I B ( μ , σ ) with a = b = 0.00001 and μ 0 = 0 and the objective priors π J ( μ , σ ) and π R ( μ , σ ) in Table 5. In addition, we present the kernel density of the joint entropy based on the MCMC samples in Figure 4.
Table 5 shows that the joint entropy under the informative prior distribution π I B ( μ , σ ) is larger than it is under the objective priors (7) and (8). In addition, Figure 4 shows that the tail of the kernel density under the subjective prior π I ( μ , σ ) with a = b = 0.00001 and μ 0 = 0 is heavier than it is under the objective priors π J ( μ , σ ) and π R ( μ , σ ) . This is due to the fact that σ ^ I B is estimated to be larger than σ ^ J B and σ ^ R B (see Table 3).

4. Conclusions

In this paper, we proposed an entropy inference method that is based on an objective Bayesian approach for upper record values having the two-parameter logistic distribution. We first obtained non-informative priors, namely, the Jeffreys and reference priors, for the unknown parameters of the two-parameter logistic distribution. Subsequently, we derived the joint entropy based on the upper record values and examined its properties. We evaluated the objective Bayesian models under the two objective priors through the posterior predictive checking that was based on the replications of the observed upper record values. The proposed objective Bayesian approach is useful when there is not enough prior information.

Acknowledgments

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by theMinistry of Education (No. 2015R1D1A1A01057847). We are grateful to the Editor-in-chief, an Associate Editor and anonymous referees for their helpful comments.

Author Contributions

Jung In Seo and Yongku Kim conceived the idea and developed the mothod presented in this paper. Both authors performed data analysis, and wrote the paper. Both authors have read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  2. Baratpour, S.; Ahmadi, J.; Arghami, N.R. Entropy properties of record statistics. Stat. Pap. 2007, 48, 197–213. [Google Scholar] [CrossRef]
  3. Abo-Eleneen, Z.A. The entropy of progressively censored samples. Entropy 2011, 13, 437–449. [Google Scholar] [CrossRef]
  4. Kang, S.B.; Cho, Y.S.; Han, J.T.; Kim, J. An estimation of the entropy for a double exponential distribution based on multiply Type-II censored samples. Entropy 2012, 14, 161–173. [Google Scholar] [CrossRef]
  5. Seo, J.I.; Kang, S.B. Entropy estimation of generalized half-logistic distribution (GHLD) based on Type-II censored samples. Entropy 2014, 16, 443–454. [Google Scholar] [CrossRef]
  6. Park, S. Testing exponentiality based on the Kullback-Leibler information with the Type-II censored data. IEEE Trans. Reliab. 2005, 54, 22–26. [Google Scholar] [CrossRef]
  7. Rad, A.H.; Yousefzadeh, F.; Amini, M.; Arghami, N.R. Testing exponentiality based on record values. J. Iran. Stat. Soc. 2007, 6, 77–87. [Google Scholar]
  8. Asgharzadeh, A.; Valiollahi, R.; Abdi, M. Point and interval estimation for the logistic distribution based on record data. Stat. Oper. Res. Trans. 2016, 40, 1–24. [Google Scholar]
  9. Bernardo, J.M. Reference posterior distributions for Bayesian inference (with discussion). J. R. Stat. Soc. Ser. B 1979, 41, 113–147. [Google Scholar]
  10. Berger, J.O.; Bernardo, J.M. Estimating a product of means: Bayesian analysis with reference priors. J. Am. Stat. Assoc. 1989, 84, 200–207. [Google Scholar] [CrossRef]
  11. Berger, J.O.; Bernardo, J.M. On the development of reference priors (with discussion). In Bayesian Statistics IV; Bernardo, J.M., Berger, J.O., Dawid, A.P., Smith, A.F.M., Eds.; Oxford University Press: Oxford, UK, 1992; pp. 35–60. [Google Scholar]
  12. Devroye, L. A simple algorithm for generating random variates with a log-concave density. Computing 1984, 33, 247–257. [Google Scholar] [CrossRef]
  13. Roberts, G.O.; Rosenthal, J.S. Optimal scaling for various Metropolis-Hastings algorithms. Stat. Sci. 2001, 16, 351–367. [Google Scholar] [CrossRef]
Figure 1. Entropy of H U ( i ) (a) and H U ( 1 ) , , U ( k ) (b) for upper record values.
Figure 1. Entropy of H U ( i ) (a) and H U ( 1 ) , , U ( k ) (b) for upper record values.
Entropy 19 00208 g001
Figure 2. (a) Autocorrelation functions (ACF) (left) and trace plot (right) of the Markov chain Monte Carlo (MCMC) samples under the prior π I ( μ , σ ) , (b) ACF (left) and trace plot (right) of the MCMC samples under the prior π J ( μ , σ ) and (c) ACF (left) and trace plot (right) of the MCMC samples under the prior π R ( μ , σ ) .
Figure 2. (a) Autocorrelation functions (ACF) (left) and trace plot (right) of the Markov chain Monte Carlo (MCMC) samples under the prior π I ( μ , σ ) , (b) ACF (left) and trace plot (right) of the MCMC samples under the prior π J ( μ , σ ) and (c) ACF (left) and trace plot (right) of the MCMC samples under the prior π R ( μ , σ ) .
Entropy 19 00208 g002
Figure 3. (a) ACF (left) and trace plot (right) of the MCMC samples under the prior π I ( μ , σ ) , (b) ACF (left) and trace plot (right) of the MCMC samples under the prior π J ( μ , σ ) and (c) ACF (left) and trace plot (right) of the MCMC samples under the prior π R ( μ , σ ) .
Figure 3. (a) ACF (left) and trace plot (right) of the MCMC samples under the prior π I ( μ , σ ) , (b) ACF (left) and trace plot (right) of the MCMC samples under the prior π J ( μ , σ ) and (c) ACF (left) and trace plot (right) of the MCMC samples under the prior π R ( μ , σ ) .
Entropy 19 00208 g003
Figure 4. (a) Kernel density of the joint entropy based on the MCMC samples under the prior π I ( μ , σ ) , (b) Kernel density of the joint entropy based on MCMC samples under the prior π J ( μ , σ ) and (c) Kernel density of the joint entropy based on MCMC samples under the prior π R ( μ , σ ) .
Figure 4. (a) Kernel density of the joint entropy based on the MCMC samples under the prior π I ( μ , σ ) , (b) Kernel density of the joint entropy based on MCMC samples under the prior π J ( μ , σ ) and (c) Kernel density of the joint entropy based on MCMC samples under the prior π R ( μ , σ ) .
Entropy 19 00208 g004aEntropy 19 00208 g004b
Table 1. Entropy based on the i-th upper record value X U ( i ) .
Table 1. Entropy based on the i-th upper record value X U ( i ) .
i12345678910
σ
0.1−0.303−0.370−0.302−0.208−0.115−0.0290.0480.1170.1790.234
0.51.3071.2391.3071.4011.4941.5801.6571.7271.7881.844
12.0001.9322.0012.0942.1872.2732.3512.4202.4812.537
22.6932.6252.6942.7872.8802.9663.0443.1133.1753.230
43.3863.3193.3873.4803.5743.6603.7373.8063.8683.923
84.0794.0124.0804.1744.2674.3534.4304.4994.5614.616
Table 2. Joint entropy based on X U ( 1 ) , , X U ( k ) .
Table 2. Joint entropy based on X U ( 1 ) , , X U ( k ) .
k12345678910
σ
0.1−0.303−1.250−2.400−3.632−4.900−6.187−7.481−8.780−10.080−11.382
0.51.3071.9692.4292.8063.1473.4703.7854.0964.4054.712
12.0003.3554.5085.5796.6127.6298.6379.64110.64311.644
22.6934.7416.5878.35110.07811.78813.48915.18616.88118.575
43.3866.1288.66711.12413.54415.94718.34120.73123.12025.507
84.0797.51410.74613.89617.01020.10623.19326.27629.35832.438
Table 3. Estimates of μ and σ and the corresponding 95 % HPD CrIs.
Table 3. Estimates of μ and σ and the corresponding 95 % HPD CrIs.
μ ^ IB μ ^ JB μ ^ RB σ ^ IB σ ^ JB σ ^ RB
Estimate3.2253.1963.2091.2341.0101.096
HPD CrI(2.137, 4.465)(2.694, 3.904)(2.392, 4.122)(0.652, 2.287)(0.386, 1.635)(0.403, 1.935)
AR 0.3890.4380.439
Table 4. Replications and their mean and standard deviation (std).
Table 4. Replications and their mean and standard deviation (std).
X U ( 1 ) rep X U ( 2 ) rep X U ( 3 ) rep X U ( 4 ) rep X U ( 5 ) rep MeanStd
π I B 3.2245.2626.7268.0899.3536.5312.522
π J B 3.1954.8466.0467.1768.2045.8942.064
π R B 3.2215.0006.2887.5068.6296.1292.226
Table 5. Estimates and the corresponding 95 % HPD CrI of the joint entropy H ^ U ( i ) , , U ( k ) B .
Table 5. Estimates and the corresponding 95 % HPD CrI of the joint entropy H ^ U ( i ) , , U ( k ) B .
H ^ U ( i ) , , U ( k ) IB H ^ U ( i ) , , U ( k ) JB H ^ U ( i ) , , U ( k ) RB
Estimate7.4076.4316.803
HPD CrI(4.910, 11.027)(2.648, 9.417)(3.455, 10.557)

Share and Cite

MDPI and ACS Style

Seo, J.I.; Kim, Y. Objective Bayesian Entropy Inference for Two-Parameter Logistic Distribution Using Upper Record Values. Entropy 2017, 19, 208. https://doi.org/10.3390/e19050208

AMA Style

Seo JI, Kim Y. Objective Bayesian Entropy Inference for Two-Parameter Logistic Distribution Using Upper Record Values. Entropy. 2017; 19(5):208. https://doi.org/10.3390/e19050208

Chicago/Turabian Style

Seo, Jung In, and Yongku Kim. 2017. "Objective Bayesian Entropy Inference for Two-Parameter Logistic Distribution Using Upper Record Values" Entropy 19, no. 5: 208. https://doi.org/10.3390/e19050208

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop