Next Article in Journal
An Operator-Based Scheme for the Numerical Integration of FDEs
Previous Article in Journal
Generalized Strongly Increasing Semigroups
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Generalized Mixtures of Exponential Distribution and Associated Inference

1
Department of Applied Mathematics, Xi’an University of Technology, Xi'an 710054, China
2
Department of Mathematical Sciences, Eastern New Mexico University, Portales, NM 88130, USA
3
Department of Mathematical Sciences, New Mexico State University, Las Cruces, NM 88003, USA
*
Author to whom correspondence should be addressed.
Mathematics 2021, 9(12), 1371; https://doi.org/10.3390/math9121371
Submission received: 1 May 2021 / Revised: 31 May 2021 / Accepted: 10 June 2021 / Published: 13 June 2021

Abstract

:
A new generalization of the exponential distribution, namely the generalized mixture of exponential distribution, is introduced. Some of its basic properties, such as hazard function, moments, order statistics, mean deviation, measures of uncertainly, and reliability probability, are studied. Three different estimation methods are investigated by the maximum likelihood estimator, least-square estimator, and weighted least-square estimator. The performances of the estimators are assessed by simulation studies. Real-world applications of the proposed distribution are explored, and data fitting results show that the new distribution performs better than its competitors.

1. Introduction

Among the parametric models, exponential distribution is perhaps the most widely applied statistical distribution in several fields and plays an important role in the statistical theory of reliability and lifetime analysis. Based on this reason, statisticians have been interested in defining new classes of univariate distributions by adding one or more shape parameters to provide greater flexibility in modeling real data in many applied fields. Gupta and Kundu [1] studied the generalized exponential distribution and used it as an alternative to gamma or Weibull distribution in many situations. Gupta and Kundu [2] used the idea of Azzalini [3], introducing a new class of weighted exponential (WE) distributions, and Kharazmi et al. [4] extended it into the generalized weighted exponential (GWE) distribution. Nadarajah and Haghighi [5] discussed a new two-parameter generalization of the exponential distribution, which had its mode at zero and allowed increasing, decreasing, and constant hazard rates.
On the other hand, generalizations of exponentiated type distributions can be obtained from the class of generalized beta distributions, in particular after the works of Eugene et al. [6] Nadarajah and Kotz [7] introduced the beta exponential distribution, generated from the logit of a beta random variable. Barreto-Souza et al. [8] discussed beta generalized exponential distribution, which includes the beta exponential and generalized exponential distributions as special cases. A generalization of the exponentiated Frenchet distribution, called the beta Frenchet distribution, was studied by Barreto-Souza et al. [9]. Ristic and Balakrishnan [10] proposed the gamma exponential distribution generated by gamma random variables. Being of a similar methodology, many X-family exponential distributions were investigated recently. These include the Weibull exponential (WED) distributions which were introduced by Oguntunde et al. [11], Marshall–Olkin generalized exponential distributions which were defined by Ristic and Kundu [12], Kumaraswamy Marshall–Olkin exponential distributions which were given by George and Thobias [13], and generalized extended exponential-Weibull (GExtEW) distributions which were proposed by Shakhatreh et al. [14].
In this paper, a new class of generalized mixture exponential (GME) distribution is introduced, which has the exponential and WE distributions as its submodels. In order to motivate interest, let us first present the definition of the generalized skew normal distribution introduced by Kumar and Anusree [15]. A random variable Z is said to have a generalized skew normal distribution if its probability density function (pdf) is of the following form,
h ( z , λ , β ) = 2 2 + β f ( z ) ( 1 + β F ( λ z ) ) ,
where f ( z ) = ϕ ( z ) , F ( λ z ) = Φ ( λ z ) , λ and β > 2 . In fact, the correct values of β should be β 1 , which has been discussed in Tian et al. [16].
The rest of the article is organized as follows. The GME distribution is introduced in Section 2. Some important properties of GME distributions, such as cumulative distribution function (cdf), hazard function, mean deviations, order statistics, measure of uncertainly, and reliability probability, are discussed in Section 3. Three different estimation methods are studied in Section 4. Simulations are conducted to investigate and compare the performances of the proposed estimation methods in Section 5. Two real data sets are analyzed for illustrating the usefulness of the proposed GME distribution in Section 6. Some conclusions are presented in Section 7.

2. Generalized Mixture Exponential Distribution

The GME distribution offers more flexible distributions with applications in lifetime modeling, which is defined as follows.
Definition 1.
A random variable X is said to have a GME distribution if its pdf is of the following form,
f ( x ; λ , α , β ) = ( α + 1 ) λ α + 1 + α β e λ x [ 1 + β ( 1 e α λ x ) ] , x > 0 ,
where α > 0 is the scale parameter, λ > 0 and β 1 are the shape parameters, and we denote it as X G M E ( λ , α , β ) .
Remark 1. 
(i) 
For β = 0 , or α 0 , or α , G M E ( λ , α , β ) is reduced into exponential distribution with parameter λ namely, E ( λ ) .
(ii) 
For β = 1 , G M E ( λ , α , β ) is reduced into E ( λ ( α + 1 ) ) .
(iii) 
For β , G M E ( λ , α , β ) is reduced into W E ( λ , α ) .
For different values of λ , α , β , the pdfs of G M E ( λ , α , β ) are presented in the Figure 1, which indicate that the GME distribution can generate distributions with various shapes.
Proposition 1.
The cdf, the survival function and the hazard function of X G M E ( λ , α , β ) are given by
F ( x ; λ , α , β ) = 1 + β e ( α + 1 ) λ x α + 1 + α β ( α + 1 ) ( β + 1 ) e λ x α + 1 + α β , h ( x ; λ , α , β ) = ( α + 1 ) λ 1 + β ( 1 e λ α x ) ( α + 1 + α β ) + β ( 1 e λ α x ) , S ( x ; λ , α , β ) = ( α + 1 ) ( β + 1 ) e λ x α + 1 + α β β e ( α + 1 ) λ x α + 1 + α β .
Proof of Proposition 1
According to the Equation (1), we have
F ( x ; λ , α , β ) = λ ( α + 1 ) α + 1 + α β ( 1 + β ) 0 x e λ t d t β 0 x e λ α t d t = 1 e λ x + β α + 1 + α β e λ ( 1 + α ) x e λ x .
Therefore,
h ( x ; λ , α , β ) = f ( x ; λ , α , β ) 1 F ( x ; λ , α , β ) = ( α + 1 ) λ 1 + β ( 1 e λ α x ) ( α + 1 + α β ) + β ( 1 e λ α x ) ,
S ( x ; λ , α , β ) = 1 F ( x ; λ , α , β ) = ( α + 1 ) ( β + 1 ) e λ x α + 1 + α β β e ( α + 1 ) λ x α + 1 + α β .
This ends the proof of Proposition 1. □
Figure 2 shows that the GME distribution produces flexible hazard rate shapes, such as decreasing, increasing, and stable.

3. General Properties of the GME Distribution

In what follows, we discuss various properties associated with the proposed distribution.
Proposition 2.
The shapes of density function of X G M E ( λ , α , β ) can be characterized as follows,
(i) 
f ( x ) is monotone decreasing, if 1 β 0 or 0 < α β 1 ,
(ii) 
f ( x ) is unimodal, if α β > 1 .
Proof of Proposition 2
The derivatives of f ( x ) are obtained by Equation (1),
f ( x ) = λ 2 ( α + 1 ) e λ x α + 1 + α β β ( α + 1 ) β + 1 e α λ x 1 .
(i)
if 1 β 0 , we have λ 2 ( α + 1 ) e λ x α + 1 + α β > 0 , and ( α + 1 ) β 0 , thus,
λ 2 ( α + 1 ) e λ x α + 1 + α β β ( α + 1 ) β + 1 e α λ x 1 < 0 ;
if 0 < α β 1 , we get e α λ x < 1 and β ( α + 1 ) β + 1 < 1 , thus, β ( α + 1 ) β + 1 e α λ x 1 < 0 .
(ii)
Setting f ( x ) = 0 , we have x 0 = 1 α λ l o g β + 1 β + β α , and f ( x 0 ) > 0 . Thus, if α β > 1 , we have f ( x ) is monotone increasing on 0 < x < x 0 and monotone decreasing on x > x 0 . Thus, f ( x ) is unimodal.
This ends the proof of Proposition 2. □
Remark 2.
The two properties in Proposition 2 are actually exclusive.
Proposition 3.
Let X G M E ( λ , α , β ) , the moment generating function of X is
M X ( t ) = λ ( α + 1 ) α + 1 + α β 1 + β λ t β λ α + λ t , t < λ .
Proof of Proposition 3
According to Equation (1) and the definition of moment generating function,
M X ( t ) = λ ( α + 1 ) α + 1 + α β ( 1 + β ) 0 e ( t λ ) x d x β 0 e ( t λ λ α ) x d x = λ ( α + 1 ) α + 1 + α β 1 + β λ t β λ α + λ t , t < λ .
This ends the proof of Proposition 3. □
Corollary 1.
Let X G M E ( λ , α , β ) , the first four moments of X are
E [ X ] = ( 1 + α ) 2 ( 1 + β ) β λ ( α + 1 + α β ) ( 1 + α ) , E [ X 2 ] = 2 ( 1 + α ) 3 ( 1 + β ) 2 β λ 2 ( α + 1 + α β ) ( 1 + α ) 2 , E [ X 3 ] = 6 ( 1 + α ) 4 ( 1 + β ) 6 β λ 3 ( α + 1 + α β ) ( 1 + α ) 3 , E [ X 4 ] = 24 ( 1 + α ) 5 ( 1 + β ) 24 β λ 4 ( α + 1 + α β ) ( 1 + α ) 4 .
Proposition 4.
Let X G M E ( λ , α , β ) and μ = E [ X ] , then the mean deviation about the mean of X is given by
D ( μ ) = 2 ( α + 1 + α β + β ) λ ( α + 1 + α β ) e λ μ 2 β λ ( α + 1 + α β ) ( 1 + α ) e λ μ ( 1 + α ) .
Proof of Proposition 4
According to Equation (1) and D ( μ ) = E [ X μ ] , with μ = E [ X ] , we have
D ( μ ) = 0 μ ( μ x ) f ( x ) d x + μ ( x μ ) f ( x ) d x = μ 0 μ f ( x ) d x μ f ( x ) d x + 0 μ x f ( x ) d x + μ x f ( x ) d x = μ 2 F ( μ ) 1 + μ 2 0 μ x f ( x ) d x = 2 β λ ( α + 1 + α β ) ( 1 + α ) 1 e λ μ ( 1 + α ) 2 ( α + 1 + α β + β ) λ ( α + 1 + α β ) ( 1 e λ μ ) + 2 μ = 2 ( α + 1 + α β + β ) λ ( α + 1 + α β ) e λ μ 2 β λ ( α + 1 + α β ) ( 1 + α ) e λ μ ( 1 + α ) .
This ends the proof of Proposition 4. □
The entropy of a random variable is a measure of uncertainty, which is an important topic in the fields of communication theory, statistical physics, and probability theory. In the following, we study the entropy measures for X G M E ( λ , α , β ) .
Proposition 5.
Let X G M E ( λ , α , β ) , then the Shannon entropy, S(x), and Renyi entropy, R(x), of X are given by
S ( x ) = log λ ( α + 1 ) α + 1 + α β 1 + 1 α log ( 1 + β ) + log ( β ) α log ( u ) α log ( 1 u ) , R γ ( x ) = 1 1 γ γ log λ ( α + 1 ) α + 1 + α β + γ + γ α log ( 1 + β ) γ log ( β ) α log ( λ α ) + log B β 1 + β ; γ α , γ + 1 , γ > 0 a n d γ 1 ,
respectively, where B ( z ; a , b ) = 0 z u a 1 ( 1 u ) b 1 d u is the incomplete beta function.
Proof of Proposition 5
For any γ > 0 and γ 1 , the Reni entropy of X is defined as
R γ ( x ) = 1 1 γ log 0 f γ ( x ; λ , β , α ) d x = 1 1 γ log λ ( α + 1 ) α + 1 + α β γ 0 e λ γ x 1 + β ( 1 e λ α x ) γ d x = 1 1 γ log λ ( α + 1 ) α + 1 + α β γ ( 1 + β ) γ + γ α β γ α × 0 β 1 + β e λ α x γ α 1 β 1 + β e λ α x γ d x = 1 1 γ γ log λ ( α + 1 ) α + 1 + α β + γ + γ α log ( 1 + β ) γ α log ( β ) log ( λ α ) + log B β 1 + β ; γ α , γ + 1 .
The Shannon entropy S ( x ) is the limiting value of R γ ( x ) as γ 1 and, thus, the results are obtained.
This ends the proof of Proposition 5. □
In the next proposition, we study the probability that one of the two independent GME random variables exceeds the other, which is named as the reliability probability.
Proposition 6.
Suppose two independent random variables X and Y follow G M E ( λ , α , β ) , then the reliability probability is given by
P ( X > Y ) = ( 1 + α ) 2 ( 1 + β ) 2 2 ( 1 + α + α β ) 2 + λ β 2 2 ( 1 + α + α β ) 2 β ( 1 + β ) ( 1 + α ) 2 ( 2 + α ) ( 1 + α + α β ) 2 β ( 1 + β ) ( 1 + α ) ( 2 + α ) ( 1 + α + α β ) 2 .
Proof of Proposition 6
Let Z = Y X and X = X , the joint density function of X and Z is obtained as
f ( x , z ; λ , β , α ) = λ 2 ( 1 + α ) 2 ( 1 + β ) 2 ( 1 + α + α β ) 2 e λ ( 2 x + z ) λ 2 β ( 1 + β ) ( 1 + α ) 2 ( 1 + α + α β ) 2 e λ ( 1 + α ) z + ( 2 + α ) x λ 2 β ( 1 + β ) ( 1 + α ) 2 ( 1 + α + α β ) 2 e λ z + ( 2 + α ) x + λ 2 β 2 ( 1 + α ) 2 ( 1 + α + α β ) 2 e λ ( 1 + α ) z + ( 2 + 2 α ) x .
Therefore, the marginal density function of Z is
f ( z ; λ , β , α ) = ( α + 1 ) 2 ( 1 + β ) 2 ( 1 + α + α β ) 2 e λ z + λ β 2 2 ( 1 + α + α β ) 2 e λ ( 1 + α ) z β ( 1 + β ) ( 1 + α ) 2 ( 2 + α ) ( 1 + α + α β ) 2 e λ z β ( 1 + β ) ( 1 + α ) ( 2 + α ) ( 1 + α + α β ) 2 e λ ( 1 + α ) z .
Thus, the result is obtained by P ( X > Y ) = P ( Z < 0 ) .
This ends the proof of Proposition 6. □
Order statistics are fundamental tools in non-parametric statistics and inference. In what follows, we derive an expression for the density function of the r t h order statistic in a random sample size n r from the GME distribution.
Proposition 7.
Suppose X 1 , X 2 , , X n is a random sample from G M E ( λ , α , β ) . Let X 1 : n X 2 : n X n : n denote the corresponding order statistics. Then the pdf and cdf of r t h order statistic, X r : n , 1 r n , are respectively,
f r : n ( x ) = n ! ( r 1 ) ! ( n r ) ! ( 1 + α ) λ α + 1 + α β ( e λ ( 1 + α ) x e λ x ) e λ x + 1 r 1 × ( 1 + α ) λ α + 1 + α β ( e λ x e λ ( 1 + α ) x ) + e λ x n r ( α + 1 ) λ α + 1 + α β e λ x [ 1 + β ( 1 e α λ x ) ] , F r : n ( x ) = l = r n u = 0 n r ( 1 ) u n l n r u ( 1 + α ) λ α + 1 + α β ( e λ ( 1 + α ) x e λ x ) e λ x + 1 l + u .
Proof of Proposition 7
It is well known that the pdf and cdf of X r : n , 1 r n are given by
f r : n ( x ) = n ! ( r 1 ) ! ( n r ) ! [ F ( x ) ] r 1 [ 1 F ( x ) ] n r f ( x ) , F r : n ( x ) = l = r n n l [ F ( x ) ] l [ 1 F ( x ) ] n l ,
respectively. Thus, the results are derived straightly from Equation (1) and Proposition 1.
This ends the proof of Proposition 7. □
Proposition 8.
Let X G M E ( λ , α , β ) , the quantile function of GME distribution, x q , wherein 0 < q < 1 , can be obtained by solving the following equation,
β e ( α + 1 ) λ x q ( α + 1 ) ( β + 1 ) e λ x q = ( q 1 ) ( 1 + α + α β ) .
We can see from Equation (2) that there is no closed form of the solution in x q and, thus, we have to use numerical techniques to obtain the quantile.
The mean residual life (MRL) function plays a very important role in reliability engineering, survival analysis and many other fields. It represents the period from time t till the time of failure, and the MRL also represents the expected additional life length for a unit.
Proposition 9.
Let X G M E ( λ , α , β ) , the MRL function of GME distribution, defined as μ X ( t ) , is given by
μ X ( t ) = ( 1 + α ) 2 ( 1 + β ) e λ t β e ( 1 + α ) λ t ( 1 + α ) 2 ( 1 + β ) λ e λ t ( 1 + α ) λ β e ( 1 + α ) λ t , t > 0 .
Proof of Proposition 9
For t > 0 , we have
μ X ( t ) = E ( X t | X > t ) = t S ( x ; λ , α , β ) d x S ( t ; λ , α , β ) = t ( α + 1 ) ( β + 1 ) e λ x α + 1 + α β β e ( α + 1 ) λ x α + 1 + α β d x S ( t ; λ , α , β ) ,
where S ( · ) is survival function of GME distribution. We know that
t ( α + 1 ) ( β + 1 ) e λ x α + 1 + α β β e ( α + 1 ) λ x α + 1 + α β d x = t ( α + 1 ) ( β + 1 ) e λ x α + 1 + α β d x t β e ( α + 1 ) λ x α + 1 + α β d x = ( 1 + α ) 2 ( 1 + β ) e λ t β e ( 1 + α ) λ t ( 1 + α + α β ) ( 1 + α ) λ .
Thus, the result is obtained.
This ends the proof of Proposition 9. □

4. Methods of Estimation

In this section, we consider the methods of maximum likelihood, least squares, and weighted least squares to estimate the unknown parameters, θ = ( λ , α , β ) , of the GME distribution. Suppose x 1 , x 2 , , x n is a random sample from G M E ( λ , α , β ) .

4.1. Maximum Likelihood Estimator

The method of maximum likelihood is the most frequently used method for parameter estimation. According to the Equation (1), the likelihood function is calculated as
L ( λ , α , β | x 1 , , x n ) = ( α + 1 ) λ α + 1 + α β n e λ i = 1 n x i i = 1 n [ 1 + β ( 1 e α λ x i ) ] .
The log-likelihood function is given by
( λ , α , β | x 1 , , x n ) = n [ log ( α + 1 ) + log ( λ ) log ( α + 1 + α β ) ] λ i = 1 n x i + i = 1 n log [ 1 + β ( 1 e α λ x i ) ] .
We denote the first partial derivatives of (3) by λ , α and β . Setting λ = 0 , α = 0 , and β = 0 , we have
λ = n λ i = 1 n x i + i = 1 n β α x i e α λ x i 1 + β ( 1 e α λ x i ) = 0 , α = n α + 1 n ( 1 + β ) α + 1 + α β + i = 1 n β λ x i e α λ x i 1 + β ( 1 e α λ x i ) = 0 , β = n α α + 1 + α β + i = 1 n 1 e α λ x i 1 + β ( 1 e α λ x i ) = 0 .
The maximum likelihood estimator (MLE) θ ^ of the unknown parameters θ can be obtained by optimizing the log-likelihood function with respect to the involved parameters. Due to the non-linearity of these equations, the MLEs of parameters can be obtained numerically. These estimators can be easily obtained by using the functions from the statistical software R.
Fisher information is helpful to get the reference priors for the model parameters. In the following, we observe that the Fisher information is given by
I ( θ ) = E λ λ λ α λ β α λ α α α β β λ β α β β ,
where
λ λ = n λ 2 i = 1 n β ( 1 + β ) α 2 x i 2 e 2 α λ x i [ 1 + β ( 1 e α λ x i ) ] 2 , α α = n ( α + 1 ) 2 + n ( 1 + β ) 2 ( 1 + α + α β ) 2 i = 1 n β ( 1 + β ) λ 2 x i 2 e 2 α λ x i [ 1 + β ( 1 e α λ x i ) ] 2 , β λ = i = 1 n α x i e α λ x i [ 1 + β ( 1 e α λ x i ) ] 2 = λ β , β α = i = 1 n λ x i e α λ x i [ 1 + β ( 1 e α λ x i ) ] 2 = α β , λ α = i = 1 n β ( 1 + β ) ( 1 α λ x i ) x i e α λ x i β 2 x i e 2 α λ x i [ 1 + β ( 1 e α λ x i ) ] 2 = α λ , β β = n α 2 ( 1 + α + α β ) 2 i = 1 n ( 1 e α λ x i ) 2 [ 1 + β ( 1 e α λ x i ) ] 2 .

4.2. Least-Square Estimator

Suppose F ( x ( j ) ) denotes the distribution function of the ordered random variables x ( 1 ) < < x ( n ) . Denote the following function
h ( λ , α , β ) = i = 1 n F ( x ( i ) ; λ , α , β ) i n + 1 2 ,
where F ( x ; λ , α , β ) = β α + 1 + α β [ e λ ( 1 + α ) x e λ x ] e λ x + 1 , and the least-square estimator (LS) of θ can be obtained by minimizing h ( λ , α , β ) . Therefore, θ ^ can be obtained by solving the following equations,
h ( λ , α , β ) λ = i = 1 n 2 Q i β α + 1 + α β ( 1 + α ) C i + B i + B i = 0 , h ( λ , α , β ) α = i = 1 n 2 Q i ( 1 + β ) ( α + 1 + α β ) 2 [ e λ ( 1 + α ) X ( i ) e λ X ( i ) ] β λ α + 1 + α β C i = 0 , h ( λ , α , β ) β = i = 1 n 2 Q i 1 α ( α + 1 + α β ) 2 [ e λ ( 1 + α ) X ( i ) e λ X ( i ) ] = 0 ,
where Q i = β α + 1 + α β [ e λ ( 1 + α ) x ( i ) e λ x ( i ) ] e λ x ( i ) + 1 i n + 1 , B i = x ( i ) e λ x ( i ) and
C i = x ( i ) e λ ( 1 + α ) x ( i ) .

4.3. Weighted Least-Square Estimator

The weighted least-square estimator (WLS) is an extension of LS and proposed by Swain et al. [17], which studied the WLS is obtained by minimizing the function,
W ( λ , α , β ) = i = 1 n ( n + 1 ) 2 ( n + 2 ) i ( n i + 1 ) F ( x ( i ) ; λ , α , β ) i n + 1 2 ,
where F ( · ) function has been given in Equation (4). Therefore, the WLS of θ can be obtained by
W ( λ , α , β ) λ = i = 1 n 2 ( n 1 ) 2 ( n + 2 ) i ( n i + 1 ) Q i β α + 1 + α β ( 1 + α ) C i + B i + B i = 0 , W ( λ , α , β ) α = i = 1 n { 2 ( n 1 ) 2 ( n + 2 ) i ( n i + 1 ) Q i { ( 1 + β ) ( α + 1 + α β ) 2 [ e λ ( 1 + α ) x ( i ) e λ x ( i ) ] β λ α + 1 + α β C i } } = 0 , W ( λ , α , β ) β = i = 1 n 2 ( n 1 ) 2 ( n + 2 ) i ( n i + 1 ) Q i 1 α ( α + 1 + α β ) 2 [ e λ ( 1 + α ) x ( i ) e λ x ( i ) ] = 0 ,
where Q i , B i and C i , i = 1 , , n , are defined as above.

5. Simulation Studies

In this section, we assess the performance of the estimation methods proposed in the previous section by conducting several simulations for different sample sizes and values of the parameter.
As indicated in Proposition 1, the F ( x ; λ , α , β ) there is used to generate pseudo-random numbers from the GME distribution. This technique is called the inverse transform method, which consists of the following steps:
(i)
Generate a random number u from the standard uniform distribution in the interval [0,1].
(ii)
Apply the numerical techniques to solve the equation F ( x ) = u with given λ , α , β .
We take the sample size n = 50 , 100 , 200 , 300 , 400 , 500 , 1000 for each simulation, and each sample was replicated N = 1000 times. The values of parameter θ = ( 1 , 5 , 0.5 ) , ( 2 , 1 , 1.5 ) , and ( 3 , 10 , 5 ) are considered, respectively. All the results were computed using the R programming. The evaluation of the estimators are performed based on the average bias and the standard error (SE) for each single parameter, where B i a s ( θ ^ j ) = 1 N i = 1 N ( θ ^ j ( i ) θ j ) , S E ( θ ^ j ) = 1 N i = 1 N ( θ ^ j ( i ) θ j ) 2 , and θ j is the j t h component of θ . Moreover, the overall bias and mean squared error (MSE) of θ ^ are also considered, where the B i a s ( θ ^ ) = j = 1 3 B i a s ( θ ^ j ) , M S E ( θ ^ ) = 1 N i = 1 N | | θ ^ ( i ) θ | | 2 , and | | · | | is the Euclidean norm. The simulation results for different scenarios are given in the Table 1, Table 2 and Table 3.
From Table 1, Table 2 and Table 3, we find that the SE of all three estimator decrease as the sample size n increases and all estimators will tend to more accuracy when n is large. In addition, the estimated values obtained by the three estimator are close to the true values. Furthermore, the plots of bias of the simulated estimators of λ , α , and β , corresponding with different sample size n, are shown in Figure 3, Figure 4 and Figure 5, respectively.
From Figure 3, Figure 4 and Figure 5, we observe that the magnitude of bias of all estimators tends to zero as n grows, which means these estimators are asymptotically unbiased and consistent for the parameters. Thus, these estimator techniques perform well for estimating the parameters in the GME distribution. For further studying, we draw the plots of overall bias and MSE for θ ^ in Figure 6 and Figure 7.
Figure 6 and Figure 7 show the bias and MSE of θ ^ , and we can find that as n increases, the bias of θ ^ towards to zero, and the WLS always has the smallest value of MSE. The LS estimators have the largest MSE among the three considered estimators. Thus, we can conclude that WLS can be chosen as a more reliable estimator for the GME distribution.

6. Real Data Analysis

In this section, we use the weighted least-square estimator to analyze two real data sets for investigating the advantage of proposed GME distribution and compare it with some other distributions, including the exponential distribution distribution, WED distribution, GExtEW distribution, and GWE distribution, where the pdfs are given as follows.
(1)
Exponential distribution: E ( λ )
f E ( x ) = λ e λ x , x 0 , λ > 0 .
(2)
Weibull exponential distribution: W E D ( λ , α , β )
f W E D ( x ) = α β ( λ e λ x ) ( 1 e λ x ) β 1 ( e λ x ) β + 1 exp α 1 e λ x e λ x β , x > 0 , α , β , λ > 0 .
(3)
Generalized extended exponential-Weibull distribution: G E x t E W ( λ , α , β , r , c )
f G E x t E W ( x ) = c α ( r β x r 1 + λ ) ( β x r + λ x ) c 1 e ( β x r + λ x ) c , x > 0 , c , β , λ > 0 . r ( 0 , ) { 1 } .
(4)
Generalized weighted exponential distribution: G W E ( λ , α , k )
f G W E ( x ) = α B ( 1 / α , k + 1 ) λ e λ x ( 1 e λ α x ) k , x > 0 , α , λ > 0 , k Z + .

6.1. Data Set 1: Waiting Times

This data set represents the waiting times (in minutes) before the service of 100 bank customers, which has been previously used by Ghitany et al. [18]. It can be seen in Appendix A.1 of Appendix A. Table 4 shows the parameter estimator results of the GME, E, WED, GExtEW, and GWE distributions for these data. The corresponding minus log-likelihood, Akaike information criterion (AIC), and Bayesian information criterion (BIC) are also presented. From Table 4, we find that the GME has the smallest values of all criteria for comparing all other distributions.
Figure 8 shows the fitted models for data set 1. The first subgraph of Figure 8 shows the fitted densities to the data set histogram and some estimated distributions, and the second subgraph displays the empirical distribution function for the data set and the estimated distributions. Both figures reveal that the GME distribution provides a qualified fit for the data set.

6.2. Date Set 2: Survival Times

The second data set represents the survival times of 121 patients with breast cancer obtained from a large hospital in a period from 1929 to 1938. This data set has recently been studied by Lee [19] and Tahir et al. [20]. The data set can be seen in Appendix A.2 of Appendix A. We compare the GME distribution with the E, WED, GExtEW, and GWE distributions. The estimated value of the parameters, AIC and BIC statistics of these distributions are listed in Table 5. It can be seen that GME distribution provides the best fit among these competing models. Figure 9 displays the fitted pdfs and cdfs of the GME, E, WED, GExtEW, and GWE distributions for data set 2, and suggests that the fit of the GME distribution is reasonable.

7. Conclusions

In this paper, we introduce a new lifetime distribution, GME distribution, and propose several statistical properties of it. As it is not feasible to compare these methods theoretically, we have studied several simulations to identify the most efficient estimation method for GME distribution. The simulation results show that weighted least-square estimator (WLS) is the best performing estimator in terms of MSE. That is, the weighted least-square estimation method is more feasible for estimating parameters in the GME distribution. Finally, two real data sets were analyzed to indicate the importance and flexibility of GME distribution in comparison to some existing lifetime distributions. In the future, the development of properties and proper estimation procedure of the bivariate model and multivariate generalization will be of interest, and more work is needed along that direction.

Author Contributions

W.T.: Conceptualization, Methodology, Validation, Investigation, Resources, Supervision, Project Administration, Visualization, Writing review and editing; Y.Y. and T.T.: Software, Formal analysis, Data curation, Writing—original draft preparation, Visualization. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Datasets are provided in the paper.

Acknowledgments

The authors would like to thank the editor and three anonymous referees for their careful reading of this article and for their constructive suggestions, which considerably improved this article.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
WEweighted exponential distribution
GWEgeneralized weighted exponential distribution
WEDWeibull exponential distribution
GExtEWgeneralized extented exponential-Weibull distributions
GMEgeneralized mixture exponential distribution
MLEmaximum likelihood estimator
LSleast-square estimator
WLSweighted least-square estimator

Appendix A. Data Set

Appendix A.1. Data Set 1

0.8 0.8 3.2 3.3 4.6 4.7 6.2 6.2 7.7 8 9.7 9.8 12.5 12.9 17.3 18.1 27 31.6 1.3 3.5 4.7 6.2 8.2 10.7 13 18.2 33.1 1.5 1.8 1.9 3.6 4 4.1 4.8 4.9 4.9 6.3 6.7 6.9 8.6 8.6 8.6 10.9 11 11 13 13.3 13.6 18.4 18.9 19 38.5 1.9 2.1 2.6 4.2 4.2 4.3 5 5.3 5.5 7.1 7.1 7.1 8.8 8.8 8.9 11.1 11.2 11.2 13.7 13.9 14.1 19.9 20.6 21.3 2.7 2.9 3.1 4.3 4.4 4.4 5.7 5.7 6.1 7.1 7.4 7.6 8.9 9.5 9.6 11.5 11.9 12.4 15.4 15.4 17.3 21.4 21.9 23.

Appendix A.2. Data Set 2

0.3 0.3 4.0 5.0 5.6 6.2 6.3 6.6 6.8 7.4 7.5 8.4 8.4 10.3 11.0 11.8 12.2 12.3 13.5 14.4 14.4 14.8 15.5 15.7 16.2 16.3 16.5 16.8 17.2 17.3 17.5 17.9 19.8 20.4 20.9 21.0 21.0 21.1 23.0 23.4 23.6 24.0 24.0 27.9 28.2 29.1 30.0 31.0 31.0 32.0 35.0 35.0 37.0 37.0 37.0 38.0 38.0 38.0 39.0 39.0 40.0 40.0 40.0 41.0 41.0 41.0 42.0 43.0 43.0 43.0 44.0 45.0 45.0 46.0 46.0 47.0 48.0 49.0 51.0 51.0 51.0 52.0 54.0 55.0 56.0 57.0 58.0 59.0 60.0 60.0 60.0 61.0 62.0 65.0 65.0 67.0 67.0 68.0 69.0 78.0 80.0 83.0 88.0 89.0 90.0 93.0 96.0 103.0 105.0 109.0 109.0 111.0 115.0 117.0 125.0 126.0 127.0 129.0 129.0 139.0 154.0.

References

  1. Gupta, R.D.; Kundu, D. Generalized exponential distribution: Different method of estimations. J. Stat. Comput. Simul. 2001, 69, 315–337. [Google Scholar] [CrossRef]
  2. Gupta, R.D.; Kundu, D. A new class of weighted exponential distributions. Statistics 2009, 43, 621–634. [Google Scholar] [CrossRef]
  3. Azzalini, A. A class of distributions which includes the normal ones. Scand. J. Stat. 1985, 12, 171–178. [Google Scholar]
  4. Kharazmi, O.; Mahdavi, A.; Fathizadeh, M. Generalized weighted exponential distribution. Commun. Stat. Simul. Comput. 2015, 44, 1557–1569. [Google Scholar] [CrossRef]
  5. Nadarajah, S.; Haghighi, F. An extension of the exponential distribution. Statistics 2011, 45, 543–558. [Google Scholar] [CrossRef]
  6. Eugene, N.; Lee, C.; Famoye, F. Beta-normal distribution and its applications. Commun. Stat. Theory Methods 2002, 31, 497–512. [Google Scholar] [CrossRef]
  7. Nadarajah, S.; Kotz, S. The beta exponential distribution. Reliab. Eng. Syst. Saf. 2006, 91, 689–697. [Google Scholar] [CrossRef]
  8. Barreto-Souza, W.; Santos, A.H.; Cordeiro, G.M. The beta generalized exponential distribution. J. Stat. Comput. Simul. 2010, 80, 159–172. [Google Scholar] [CrossRef] [Green Version]
  9. Barreto-Souza, W.; Cordeiro, G.M.; Simas, A.B. Some results for beta Frenchet distribution. Commun. Stat. Theory Methods 2011, 40, 798–811. [Google Scholar] [CrossRef]
  10. Ristic, M.M.; Balakrishnan, N. The gamma-exponentiated exponential distribution. J. Stat. Comput. Simul. 2012, 82, 1191–1206. [Google Scholar] [CrossRef]
  11. Oguntunde, P.E.; Balogun, O.S.; Okagbue, H.I.; Bishop, S.A. The Weibull-exponential distribution: Its properties and applications. J. Appl. Sci. 2015, 15, 1305–1311. [Google Scholar] [CrossRef]
  12. Ristic, M.M.; Kundu, D. Marshall-Olkin generalized exponential distribution. Metron 2015, 73, 317–333. [Google Scholar] [CrossRef]
  13. George, R.; Thobias, S. Kumaraswamy Marshall-Olkin Exponential Distribution. Commun. Stat. Theory Methods 2019, 48, 1920–1937. [Google Scholar] [CrossRef]
  14. Shakhatreh, M.K.; Lemonte, A.J.; Cordeiro, G.M. On the generalized extended exponential-Weibull distribution: Properties and different methods of estimation. Int. J. Comput. Math. 2020, 97, 1029–1057. [Google Scholar] [CrossRef]
  15. Kumar, C.S.; Anusree, M.R. On a generalized mixture of standard normal and skew normal distributions. Stat. Probab. Lett. 2011, 81, 1813–1821. [Google Scholar] [CrossRef]
  16. Tian, W.; Wang, C.; Wu, M.; Wang, T. The multivariate extended skew normal distribution and its quadratic forms. In Causal Inference in Econometrics; Springer: Cham, Switzerland, 2016; pp. 153–169. [Google Scholar]
  17. Swain, J.J.; Venkatraman, S.; Wilson, J.R. Least-squares estimation of distribution functions in Johnson’s translation system. J. Stat. Comput. Simul. 1988, 29, 271–297. [Google Scholar] [CrossRef]
  18. Ghitany, M.E.; Atieh, B.; Nadarajah, S. Lindley distribution and its application. Math. Comput. Simul. 2008, 78, 493–506. [Google Scholar] [CrossRef]
  19. Lee, E.T. Statistical Methods for Survival Data Analysis; John Wiley: New York, NY, USA, 1992. [Google Scholar]
  20. Tahir, M.H.; Mansoor, M.; Zubair, M.; Hamedani, G. McDonald log-logistic distribution with an application to breast cancer data. J. Stat. Theory Appl. 2014, 13, 65–82. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The pdf curves for different parameters of G M E ( λ , α , β ) .
Figure 1. The pdf curves for different parameters of G M E ( λ , α , β ) .
Mathematics 09 01371 g001
Figure 2. The hazard function curves for different values of parameter in G M E ( λ , α , β ) .
Figure 2. The hazard function curves for different values of parameter in G M E ( λ , α , β ) .
Mathematics 09 01371 g002
Figure 3. Bias of estimator λ ^ versus sample size n for different scenarios.
Figure 3. Bias of estimator λ ^ versus sample size n for different scenarios.
Mathematics 09 01371 g003aMathematics 09 01371 g003b
Figure 4. Bias of estimator α ^ versus sample size n for different scenarios.
Figure 4. Bias of estimator α ^ versus sample size n for different scenarios.
Mathematics 09 01371 g004
Figure 5. Bias of estimator β ^ versus sample size n for different scenarios.
Figure 5. Bias of estimator β ^ versus sample size n for different scenarios.
Mathematics 09 01371 g005
Figure 6. Bias of estimator θ ^ versus sample size n for different scenarios.
Figure 6. Bias of estimator θ ^ versus sample size n for different scenarios.
Mathematics 09 01371 g006
Figure 7. MSE of estimator θ ^ versus sample size n for different scenarios.
Figure 7. MSE of estimator θ ^ versus sample size n for different scenarios.
Mathematics 09 01371 g007
Figure 8. Fitted pdfs and the relative histogram, empirical and fitted cdfs.
Figure 8. Fitted pdfs and the relative histogram, empirical and fitted cdfs.
Mathematics 09 01371 g008
Figure 9. Fitted pdfs and the relative histogram, empirical and fitted cdfs.
Figure 9. Fitted pdfs and the relative histogram, empirical and fitted cdfs.
Mathematics 09 01371 g009
Table 1. λ = 1 , α = 5 , β = 0.5 .
Table 1. λ = 1 , α = 5 , β = 0.5 .
Sample SizeMethod λ ^ ( SE ) α ^ ( SE ) β ^ ( SE ) MSE
n = 50MLE0.9337 (0.3743)5.0838 (2.4378)−0.1080 (1.7229)1.5273
LS0.9747 (0.2605)4.7614 (2.3440)−0.4964 (0.2255)0.9491
WLS0.9162 (0.2344)5.0376 (1.1337)−0.4444 (0.7600)0.7204
n = 100MLE0.8927 (0.3087)5.7791 (3.1581)−0.5004 (0.8806)1.4852
LS1.0136 (0.2050)4.6547 (2.2373)−0.4550 (0.2004)0.8934
WLS0.9433 (0.1481)4.6418 (0.9328)−0.4858 (0.4072)0.5307
n = 200MLE0.9407 (0.2054)5.3528 (2.7663)−0.5606 (0.2314)1.0774
LS0.9954 (0.1574)4.7508 (2.2406)−0.4717 (0.1863)0.8692
WLS0.9646 (0.1087)4.9761 (0.9436)−0.49608 (0.1881)0.4350
n = 300MLE0.9780 (0.1157)4.9851 (2.1748)−0.5229 (0.1454)0.8116
LS1.0224 (0.1291)5.4474 (2.1971)−0.4555 (0.1605)0.8504
WLS0.9654 (0.0982)4.5041 (0.8171)−0.5127 (0.1298)0.4120
n = 400MLE0.9731 (0.1137)4.0055 (1.4628)−0.5137 (0.1432)0.6742
LS1.0158 (0.1180)5.1742 (1.9422)−0.4639 (0.1519)0.7472
WLS0.9980 (0.0733)5.0675 (0.9835)−0.4948 (0.1078)0.4018
n = 500MLE0.9841 (0.0970)4.8382 (1.9489)−0.5135 (0.1202)0.7242
LS1.0023 (0.1071)4.9641 (1.7818)−0.4779 (0.1345)0.6818
WLS0.9821 (0.0610)4.9304 (0.9001)−0.5062 (0.0849)0.3783
n = 1000MLE0.9938 (0.0705)4.9315 (1.6172)−0.5020 (0.0932)0.5937
LS1.0060 (0.0763)4.9750 (1.7033)−0.4797 (0.1029)0.6367
WLS0.9894 (0.0501)4.8743 (0.8206)−0.5019 (0.0635)0.3545
Table 2. λ = 2 , α = 1 , β = 1.5 .
Table 2. λ = 2 , α = 1 , β = 1.5 .
Sample SizeMethod λ ^ ( SE ) α ^ ( SE ) β ^ ( SE ) MSE
n = 50MLE1.4458 (0.9097)1.2917 (0.5589)1.6460 (3.2067)1.6314
LS1.8709 (0.3747)1.4399 (0.5717)2.0283 (2.1736)1.1183
WLS1.9537 (0.3217)0.9440 (0.6706)2.2223 (1.8281)0.9885
n = 100MLE1.7026 (0.7886)1.3034 (0.5729)1.9781 (2.6490)1.3911
LS1.9554 (0.2942)1.2025 (0.6341)2.3221 (2.0580)0.9073
WLS1.9519 (0.2279)0.9008 (0.6066)1.9902 (1.6569)0.8606
n = 200MLE1.7535 (0.6525)1.3555 (0.5795)1.6087 (1.8850)1.0782
LS1.9669 (0.2506)1.1475 (0.6229)1.9986 (1.5482)0.7287
WLS1.9578 (0.1269)1.0596 (0.4735)1.5661 (0.6906)0.4392
n = 300MLE1.8438 (0.5467)1.2813 (0.5962)1.6082 (1.5177)0.9151
LS1.8927 (0.1585)1.6745 (0.4603)1.3887 (0.8766)0.5973
WLS1.9835 (0.1121)1.0563 (0.4295)1.5963 (0.6744)0.4141
n = 400MLE1.8684 (0.4775)1.3238 (0.5647)1.5057 (1.2785)0.8075
LS1.9413 (0.1308)1.3760 (0.5588)1.5542 (0.8849)0.5718
WLS1.9638 (0.0916)1.1273 (0.3971)1.4569 (0.3923)0.3095
n = 500MLE1.9002 (0.4561)1.2733 (0.5697)1.6020 (1.1967)0.7662
LS2.0143 (0.1602)1.0437 (0.5819)1.9070 (0.9159)0.5142
WLS1.9573 (0.0739)1.0416 (0.3624)1.5810 (0.5662)0.3489
n = 1000MLE1.9189 (0.3525)1.2974 (0.5674)1.5166 (0.8654)0.6228
LS1.9928 (0.1392)1.1979 (0.6193)1.7116 (0.7494)0.5268
WLS1.9700 (0.0613)1.1250 (0.291701.4208 (0.2599)0.2280
Table 3. λ = 3 , α = 10 , β = 5 .
Table 3. λ = 3 , α = 10 , β = 5 .
Sample SizeMethod λ ^ ( SE ) α ^ ( SE ) β ^ ( SE ) MSE
n = 50MLE3.0424 (0.4181)9.7778 (2.4919)5.6122 (3.8162)2.2299
LS3.0549 (0.5084)9.9521 (2.3594)4.4837 (4.1413)2.3605
WLS3.0281 (0.4395)10.1081 (2.0774)5.0283 (3.9031)2.1545
n = 100MLE3.0380 (0.3025)9.8275 (2.4712)5.5335 (3.4914)2.0887
LS3.0416 (0.3537)9.8668 (2.2719)5.0004 (4.0233)2.2335
WLS2.9993 (0.3392)10.1385 (2.0807)4.6623 (3.5851)2.0216
n = 200MLE3.0172 (0.2143)9.5225 (2.1802)5.4294 (3.2236)1.9055
LS3.0208 (0.2368)9.7251 (2.2322)4.9982 (3.6205)2.0550
WLS3.0072 (0.2337)10.3121 (2.2217)4.6433 (2.9485)1.8293
n = 300MLE3.0239 (0.1718)9.5911 (2.2102)5.5706 (3.1488)1.8680
LS3.0251 (0.2052)9.5669 (1.9457)5.3041 (3.7496)2.0098
WLS3.0066 (0.1789)10.052 (2.0334)4.6414 (2.7485)1.6785
n = 400MLE3.0222 (0.1455)9.3730 (1.9607)5.4066 (2.9740)1.7391
LS3.0285 (0.1623)9.5468 (2.1307)5.1461 (3.3433)1.9214
WLS3.0104 (0.1522)9.9920 (1.9927)4.6140 (2.4915)1.5730
n = 500MLE3.0166 (0.1405)9.3607 (1.9663)5.3163 (2.8919)1.6961
LS3.0163 (0.1561)9.4042 (1.9839)4.9983 (3.4092)1.9053
WLS3.0071 (0.1374)10.0114 (1.9581)4.7762 (2.4242)1.5292
n = 1000MLE3.0139 (0.0984)9.3293 (1.1715)5.1518 (2.4282)1.4799
LS3.0235 (0.1120)9.2747 (1.8501)5.1674 (3.4079)1.8668
WLS3.0136 (0.1032)9.7242 (1.6997)4.9825 (2.2380)1.3759
Table 4. Data set 1: Comparison between the GME, E, WED, GExtEW, and GWE by using different criteria.
Table 4. Data set 1: Comparison between the GME, E, WED, GExtEW, and GWE by using different criteria.
GME ( λ , α , β ) E ( λ ) WED ( λ , α , β ) GExtEW ( λ , α , β , r , c ) GWE ( λ , α , k )
λ 0.15540.08830.02340.08070.1284
α 0.8202-5.37033.95096.2681
β 124.9639-1.34860.0812-
r---1.7231-
c---0.4711-
k----4
-loglike317.3592329.9063321.8420317.1620318.4602
AIC640.7184661.8126649.6839644.3241642.9205
BIC648.5339664.4177657.4995657.3499650.7360
Table 5. Data set 2: Comparison between the GME, E, WED, GExtEW, and GWE by using different criteria.
Table 5. Data set 2: Comparison between the GME, E, WED, GExtEW, and GWE by using different criteria.
GME ( λ , α , β ) E ( λ ) WED ( λ , α , β ) GExtEW ( λ , α , β , r , c ) GWE ( λ , α , k )
λ 0.03110.01970.01140.02800.0237
α 0.6384-1.22631.833811.8562
β 7.5142-1.00350.0069-
r---0.8196-
c---0.9454-
k----3
-loglike578.9263585.5995580.1301580.6417590.0420
AIC1163.85301173.19901166.26001171.28301186.0850
BIC1172.24001175.99501174.64801185.26201194.4720
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, Y.; Tian, W.; Tong, T. Generalized Mixtures of Exponential Distribution and Associated Inference. Mathematics 2021, 9, 1371. https://doi.org/10.3390/math9121371

AMA Style

Yang Y, Tian W, Tong T. Generalized Mixtures of Exponential Distribution and Associated Inference. Mathematics. 2021; 9(12):1371. https://doi.org/10.3390/math9121371

Chicago/Turabian Style

Yang, Yaoting, Weizhong Tian, and Tingting Tong. 2021. "Generalized Mixtures of Exponential Distribution and Associated Inference" Mathematics 9, no. 12: 1371. https://doi.org/10.3390/math9121371

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop