Abstract
A recent paper introduced the interpolating family (IF) of distributions, and they also derived various mathematical properties of the family. Some of the most important properties discussed were the integer order moments of the IF distributions. The moments were expressed as an integral (which were not evaluated) or as finite sums of the beta function. In this paper, more general expressions for moments of any integer order or any real order are derived. Apart from being more general, our expressions converge for a wider range of parameter values. The expressions for entropies are also derived, the maximum likelihood estimation is considered and the finite sample performance of maximum likelihood estimates is investigated.
MSC:
62E99
1. Introduction
Ref. [1] introduced the interpolating family (IF) of size distributions, which is given by the probability density function
for , , , , and , where
Ref. [1] derived several mathematical properties of (1), including special cases, the cumulative distribution function, survival function, hazard function, quantile function, the median, random variate generation, moments, the mean, variance, unimodality and the location of the mode.
As explained in [1], the distribution given by (1) is not new. The motivation for (1) was to introduce a distribution that combines Pareto-type distributions and Weibull-type distributions into one mathematical form. The aim of this paper is to derive more of the mathematical properties of (1), and hence to add to the applicability of (1). More general expressions for the moment properties of (1) given in Section 3 can entail the development of estimation methods based on moments, L moments, trimmed L moments and probability weighted moments. The derivation of entropies for (1) can help to develop estimation methods based on entropies to fit (1) to real data. The derivation of the maximum likelihood procedure for (1) can help to use the procedure to fit (1) to real data.
Let X be a random variable with its probability density function given by (1). Ref. [1] expressed the rth moment of X as
where
Ref. [1] did not simplify (2), and they stated “It is in principle possible to write out as an infinite series of beta functions, but because this expression is rather intricate and needs to be worked out on a case-by-case basis just like , we refrain from doing so”. Ref. [1] then derived simpler expressions for (2) for the following three special cases: (i) , which was referred to as the IF1 distribution; (ii) , which was referred to as the IF2 distribution; (iii) and , which was referred to as the IF3 distribution. The derived expressions are the finite sums or doubly finite sums of the beta function.
Equation (2) is when X is an IF random variable, and it can be simplified in terms of a known special function whether r is an integer or not. Particular cases of this result are when X is an IF1 random variable or when an IF3 random variable is also derived. Apart from being more general, our expressions converge for a wider range of parameter values. In fact, some of our expressions hold for all admissible values of r, p, b, c, q and .
The expressions given in this paper involve the Wright generalized hypergeometric function, , with the p numerator and q denominator parameters ([2], Equation (1.9)) being defined by
for , where denotes the set of complex numbers, , , and for and . This function was originally introduced by [3]. If
then (4) converges absolutely for all finite values of z. If
then the radius of convergence of (4) is
(See Theorem 1.5 in [2].)
Apart from the Wright generalized hypergeometric function, the calculations in this paper use the gamma and beta functions defined by
and
respectively. The gamma function is defined if is any real number. The beta function is defined if , and are any real numbers.
The rest of this paper is organized as follows. Section 2 gives a technical lemma that is useful for subsequent calculations. Section 3 derives the rth moment of an IF random variable when is an integer or a real number, it also derives the rth moment of an IF1 random variable when is an integer or a real number, and it further derives the rth moment of an IF3 random variable when is an integer or a real number. Section 4 derives the expressions for two popular entropies. The maximum likelihood estimation for (1) is considered in Section 5. Its finite sample performance is investigated in Section 6. Finally, certain conclusions are detailed in Section 7.
2. A Technical Lemma
In this section, a technical lemma is presented. The integral in the lemma arises in the common mathematical properties of (1).
Lemma 1.
If and , then
Proof.
3. Moments of the IF Random Variable
Let X denote an IF random variable. Proposition 1 expresses the integer order moment of X as a finite sum of the Wright generalized hypergeometric function. Proposition 2 expresses the real order moment of as a single Wright generalized hypergeometric function.
Proposition 1.
Let X denote an IF random variable. If is an integer and , then
The Wright generalized hypergeometric function in (11) converges for all admissible values of p, b, c, q and , such that either or and .
Proof.
Proposition 2.
Let X denote an IF random variable. If is real and , then
The Wright generalized hypergeometric function in (13) converges for all admissible values of p, b, c, q and , such that .
Proof.
Now, let X denote an IF1 random variable. The rth moment of X if is an integer can be obtained by setting in (11). The rth moment of if is a real number can be obtained by setting in (13).
Proposition 3.
Let X denote an IF1 random variable. If is an integer and , then
The Wright generalized hypergeometric function in (16) converges for all admissible values of p, b, c, q and , such that either or and .
Proposition 4.
Let X denote an IF1 random variable. If is real and , then
The Wright generalized hypergeometric function in (17) converges for all admissible values of b, c, q and , such that .
Now, let X denote an IF3 random variable. The rth moment of X if is an integer can be obtained by setting in (11). The rth moment of if is a real number can be obtained by setting in (13).
Proposition 5.
Let X denote an IF3 random variable. If is an integer and , then
The Wright generalized hypergeometric function in (18) converges for all admissible values of p, b, c, q and .
Proposition 6.
Let X denote an IF3 random variable. If is real and , then
The Wright generalized hypergeometric function in (13) converges for all admissible values of p, b, c, q and .
4. Entropies
Two of the most popular entropies are Shannon entropy [4] and Rényi entropy [5], which are defined by
and
respectively, for and . Propositions 7 and 8 derive the explicit expressions for (20) and (21), respectively, when X is an IF random variable.
Proposition 7.
Let X denote an IF random variable. If , then
The Wright generalized hypergeometric function in (22) converges for all admissible values of p, b, c, q and , such that .
Proof.
Proposition 8.
Let X denote an IF random variable. Then,
Equation (22) provides a way through which to quantify the uncertainty in a set of data and a flexible framework with respect to capturing different aspects of information content. It can be used in addition to variance as a measure of uncertainty. (22) is a monotonic increasing function of c, the scale parameter, and it is independent of , the location parameter. The behavior of (22) with respect to other parameters depends on the Wright generalized hypergeometric function.
5. Estimation
Suppose is a random sample from (1). In this section, the maximum likelihood estimation of is considered and the associated observed information matrix is derived. The log-likelihood function is
The partial derivatives of (24) with respect to the parameters are
and
where , , , and . The maximum likelihood estimators of , say , can be obtained as the simultaneous solutions of , , , and . The maximum likelihood estimators of can also be obtained by directly maximizing (24). In Section 6, the maximum likelihood estimates were obtained by directly maximizing (24). The optim function in R software [6] was used for numerical maximization. optim was executed for a wide range of initial values. optim did not converge for all of the initial values. Whenever optim converged, the maximum likelihood estimates were unique.
Confidence intervals and tests of the hypothesis about can be based on the fact that has an asymptotic normal distribution with the mean and covariance matrix , where denotes the expected information matrix. For a large n, can be approximated by the observed information matrix . Standard calculations show that
where
and
where , , , , log−, , , = − and . In addition, denotes the Dirac delta function.
The percent confidence intervals for p, b, c, q and based on (25) are
and
respectively, where denotes the percentile of the standard normal distribution and , denotes the th element of the inverse of with when it is replaced by .
6. Simulation Study
In this section, a simulation study is conducted to check the finite sample performance of , which was detailed in Section 5. The finite sample performance is checked with respect to bias and the mean squared error. The following scheme was used:
- (a)
- Set , , , , and ;
- (b)
- Simulate 10,000 random samples each of size n from (1), the inverse method and the quantile function (detailed in Section 4.2 of [1]);
- (c)
- (d)
- Compute the biases asfor ;
- (e)
- Compute the mean squared errors asfor ;
- (f)
- Repeat steps (b) to (e) for .
The biases are plotted in Figure 1. The mean squared errors are plotted in Figure 2. The numerical values of the biases and mean squared errors are given in Table 1 and Table 2.
Figure 1.
Biases versus . The y axes are in log scale.
Figure 2.
Mean squared errors versus . The y axes are in log scale.
Table 1.
Biases of , , , and .
Table 2.
Mean squared errors of , , , and .
With the exception of p, the biases approach zero as n approaches 100. The biases appear positive for p, c, q and . The biases appear negative for b. In terms of magnitude, the biases are smallest for and largest for c and q. With the exception of p, the mean squared errors approach zero as n approaches 100. They are smallest for and the largest for c and q. The biases and mean squared errors appear small enough for b, c, q and for the n close to 100.
The observations noted are for the particular initial values , , , , and . But the same observations hold for a wide range of other values of p, b, c, q and . In particular, the magnitude of biases always decreased to zero with increasing n, and the mean squared errors always decreased to zero with increasing n (with the exception of p). Hence, the maximum likelihood estimates of the interpolating family of distributions can be considered to behave according to the large sample theory of maximum likelihood estimation.
7. Conclusions
A family of distributions, which was proposed by [1] and referred to as the interpolating family of distributions, was studied. More general expressions for the moments of these distributions, as well as expressions for entropies, were derived. The maximum likelihood estimation of the distributions was considered, and the expressions for the score functions and the observed information matrix were derived. Simulations were performed to study the finite sample performance of the estimators. The simulations showed that the maximum likelihood estimator of p did not behave well. This may be overcome by using other estimation methods, including the method of probability weighted moments, biased corrected maximum likelihood estimation, the method of L moments, the method of trimmed L moments, the minimum distance estimation and methods based on entropies.
The most notable results in the paper are as follows: Propositions 1 and 2 expressing the moments of (1) in the most general cases; Proposition 7 expressing the Rényi entropy of (1) in the most general case; Section 5 detailing the explicit expressions for the observed information matrix.
According to [7], page 371), a flexible family of distributions should have the following properties: versatility, tractability, interpretability, a data generating mechanism and a straightforward parameter estimation. With respect to versatility, (1) can exhibit unimodal shapes (see Figure 2 in [1]). However, given (1) has five parameters, one would like to see if multimodal shapes are possible. With respect to tractability, (1) takes an elementary form and so it can be computed easily. The interpretability of the parameters in (1) was discussed in Section 2.2 of [1]. The parameters control location, scale, tail weight and shape, among others, were considered. The quantile function corresponding to (1) takes an elementary form, as shown in Section 4.2 of [1]. As such, the data generation from (1) is straight forward. The maximum likelihood estimation for (1) has to be performed numerically (see Section 5). Simulation studies show that the maximum likelihood estimator of p does not behave well, even for large samples.
Author Contributions
All authors have contributed equally to the manuscript. Conceptualization, S.N. and I.E.O.; methodology, S.N. and I.E.O. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Institutional Review Board Statement
Both authors upheld the ‘Ethical Responsibilities of Authors’.
Informed Consent Statement
Both authors gave explicit consent to participate in this study. Both authors gave explicit consent to publish this manuscript.
Data Availability Statement
Data are contained within the article.
Acknowledgments
The authors would like to thank the editor and the three referees for their careful reading and comments, as their contributions considerably improved this paper.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Sinner, C.; Dominicy, Y.; Trufin, J.; Waterschoot, W.; Weber, P.; Ley, C. From Pareto to Weibull—A constructive review of distributions on R+. Int. Stat. Rev. 2023, 91, 35–54. [Google Scholar] [CrossRef]
- Kilbas, A.A.; Srivastava, H.M.; Trujillo, J.J. Theory and Applications of Fractional Differential Equations; Elsevier: Amsterdam, The Netherlands, 2006. [Google Scholar]
- Wright, E.M. The asymptotic expansion of the generalized hypergeometric function. J. Lond. Math. Soc. 1935, 10, 286–293. [Google Scholar] [CrossRef]
- Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423, 623–656. [Google Scholar] [CrossRef]
- Rényi, A. On measures of information and entropy. In Proceedings of the 4th Berkeley Symposium on Mathematics, Statistics and Probability; Neyman, J., Ed.; University of California Press: Berkeley, CA, USA, 1960; Volume 1, pp. 547–561. [Google Scholar]
- R Development Core Team. A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2024. [Google Scholar]
- Ley, C.; Babic, S.; Craens, D. Flexible models for complex data with applications. Annu. Rev. Stat. Appl. 2021, 8, 369–391. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).

