Abstract
Entropy is an uncertainty measure of random variables which mathematically represents the prospective quantity of the information. In this paper, we mainly focus on the estimation for the parameters and entropy of an Inverse Weibull distribution under progressive first-failure censoring using classical (Maximum Likelihood) and Bayesian methods. For Bayesian approaches, the Bayesian estimates are obtained based on both asymmetric (General Entropy, Linex) and symmetric (Squared Error) loss functions. Due to the complex form of Bayes estimates, we cannot get an explicit solution. Therefore, the Lindley method as well as Importance Sampling procedure is applied. Furthermore, using Importance Sampling method, the Highest Posterior Density credible intervals of entropy are constructed. As a comparison, the asymptotic intervals of entropy are also gained. Finally, a simulation study is implemented and a real data set analysis is performed to apply the previous methods.
1. Introduction
Usually, in lifetime experiments, due to the restrictions of limited time and cost, accurate product lifetime data cannot be observed so we have censored data. The most common censoring schemes are so-called Type-I and Type-II censoring. In the first one, place N units in a life experiment and terminate the experiment after a predetermined time; for the other, terminate the experiment after the predetermined units number m has failed. Progressive censoring is a generalization of Type-II censoring which permits the units to be randomly removed at various time points instead of the end of the time.
Compared to conventional Type-I and Type-II censoring, progressive censoring, i.e., withdrawal of non-failed items, decreases the accuracy of estimation. However, in certain practical circumstances, experimenters are forced to withdraw items from tests. Thus, the application of the progressive censoring methodology allows profiting from information related to withdrawn items.
When the above methods still fail to meet the time and cost constraints, to further improve efficiency, other censoring schemes are successively filed by researchers. One of the successful attempts is the first failure censoring. In this censoring scheme, units are assigned to n groups in random with k identical units in each group. The lifetime experiment is conducted by testing all groups simultaneously until the first failure is observed in each group.
Since progressive censoring and first-failure censoring can both greatly enhance the efficiency of the lifetime experiment, Ref. [1] united these two items and developed a novel censoring scheme called the progressive first-failure censoring. In this censoring, samples are divided into n disjoint groups in random with k identical units at the beginning of the life experiment, and the experiment is terminated when the mth unit fails. When the ith unit fails, the group containing the ith is removed together with randomly selected groups, and when the mth fails, all the surviving groups are removed. Here, and m are set in advance. Note that
(1) When , the progressive first failure censoring can be reduced to the well-known progressive Type-II censoring.
(2) When , this censoring becomes the mentioned first-failure censoring.
(3) When , and , this censoring corresponds to Type-II censoring.
Since it is more efficient than other censoring schemes, many researchers have discussed the study of the progressive first-failure censoring. Ref. [2] considered both the point and interval estimation of two parameters from a Burr-XII distribution when both of the parameters are unknown; Ref. [3] dealt with the reliability function of GIED (Generalized inverted exponential distribution) under progressive first-failure censoring; Ref. [4] established different reliability sampling plans using two criteria from a Lognormal distribution based on the progressive first-failure censoring; Ref. [5] chose a competing risks data model under progressive first-failure censoring from a Gompertz distribution and estimated the model using Bayesian and non-Bayesian methods; Ref. [6] considered the lifetime performance index () under the progressive first-failure censoring schemes of a Pareto model, solved the problem of the hypothesis testing of , and gave a lower specification limit.
The Weibull distribution is used in a widespread manner in analyzing lifetime data. Nevertheless, the Weibull distribution possesses a constant, decreasing or increasing failure rate function, its failure rate function cannot be non-monotone, such as unimodal. In practice, if the research shows that the empirical failure rate function is non-monotone, then the Inverse Weibull model is a more suitable choice than the Weibull model. The Inverse Weibull model has a wide variety of applications in pharmacy, economics and chemistry.
The cumulative distribution function and the probability density function of the Inverse Weibull distribution (IWD) are separately written as
and
where , is the scale parameter and is the shape parameter.
The failure rate function is
One of the most important properties of the IWD is that its failure rate function can be unimodal. Figure 1 also evidently supports this conclusion, and we can observe that the distribution whose failure rate function is unimodal is more flexible in application.
Figure 1.
The failure rate function.
Many researchers have studied the Inverse Weibull distribution. Ref. [7] invesigated the Bayesian inference and successfully predicted the IWD for the type-II censoring scheme; Ref. [8] not only considered the Baysian estimation but also the generalized Bayesian estimation for the IWD parameters; Ref. [9] used three classical methods to estimate the parameters from IWD; Ref. [10] estimated the unknown parameters from IWD under the progressive type-I interval censoring and chose the optimal censoring schemes; Ref. [11] adopted two methods to get bias corrections of unknown parameters using maximum likelihood method of the IWD.
Entropy is a quantitive measure of the uncertainty of each probability distribution. For the random variable X, of the probability density distribution , the Shannon entropy, recorded as , is written as:
Many studies about entropy can be found in the literature. Ref. [12] proposed an indirect method using a decomposition to simplify the entropy’s calculation under the progressive Type-II censoring; Ref. [13] estimated the entropy for several exponential distributions and extended the results to other circumstances; Ref. [14] estimated the Shannon entropy of a Rayleigh model under doubly generalized Type-II hydrid censoring, and compared the performance by two criteria.
The Shannon entropy of the IWD is given by:
where is a Euler constant.
In this paper, we discuss the maximum likelihood and Bayesian estimation of the paramaters and entropy of IWD under progressive first-failure censoring. As far as we know, this topic is very new and few researchers study it. However, it needs in-depth research and innovation. The rest of this paper is elaborated as follows:
In Section 2, we derive the maximum likelihood estimation of entropy and parameters. In Section 3, we present the asymptotic intervals for the entropy and parameters. In Section 4, we work out the Bayesian estimation of entropy and parameters using Lindley and Importance Sampling methods. In Section 5, a simulation study is organized to compare different estimators. In Section 6, we analyze a real data set to explain the previous conclusions. Finally, in Section 7, a conclusion is presented.
2. Maximum Likelihood Estimation
We consider the maximum likelihood estimates (MLEs) for the entropy and parameters of an Inverse Weibull distribution under progressive first-failure censoring. Set be a sample from IWD under the progressive first-failure censoring . For simplicity, we choose for representing , . The joint probability density function is
where and is a normalizing constant.
Combining (1), (2), and (5), the likelihood function (LF) is
Then, the log-likelihood function is written as
For partial derivatives with respect to and , the corresponding score equations are
The MLEs and , separately, are the roots of Equations (8) and (9). The equations don’t have an explicit solution, so we need some numerical techniques to approximate the values of these parameters. Furthermore, according to the invariance property of MLE, we derive the ML estimator of entropy as:
3. Confidence Intervals
3.1. Asymptotic Intervals for MLEs
The confidence intervals (CIs) for the two parameters and can be constructed by the asymptotic normality of MLEs with and which are obtained by the inverse of the observed Fisher matrix.
From Equation (7), find second-order partial derivatives for and as follows:
The Fisher information matrix of two parameters and is . Here, we approximate that is a bivariate normal vector with mean and covariance matrix . As a matter of fact, we use to make an estimation of . In other words,
where
Thus, based on the normal approximation, the CIs for two parameters and are
Here, is the percentile of the standard normal distribution. Thus, as to obtain the approximate estimation of the variance of entropy, we use the delta method. Let
where
Then, the approximate estimate of is obtained by
Therefore, we approximate that
The asymptotic CI for entropy is derived as
3.2. Asymptotic Intervals for Log-Transformed MLE
Ref. [15] proposed that the asymptotic CI using log-transformed MLE has a more precise coverage probabilty. It is clear that , , and entropy are all positive. Then, we obtain that asymptotic approximate CIs for log-transformed MLEs are
Thus, based on the normal approximation of log-transformed MLE, the CIs for two parameters and are
Furthermore, a CI for entropy is
4. Bayes Estimation
4.1. Prior and Posterior Distribution
Both and are unknown parameters, so we don’t have any conjugate prior for both and . Usually, we choose independent priors of and which are both Gamma distributions. However, for the Inverse Weibull distribution, it is not appropriate to choose gamma for both priors. The specific reason is explained in detail in the Importance Sampling procedure subsection. Thus, in this case, we consider the following prior distributions:
possesses a Gamma prior with the probability density function
has a non-informative prior with the following probability density function
where a and b are pre-fixed to be known and positive.
Now, the joint prior distribution of the two parameters and can be obtained by
Then, the joint posterior PDF of two parameters and is derived by
4.2. Symmetric and Asymmetric Loss Functions
Choosing loss function is an important part of Bayesian inference. In this subsection, we consider the Bayes estimation for two parameters , , and entropy of an IWD under both the asymmetric and symmetric loss functions. A widely used symmetric loss function is the squared error loss function (SELF). As for asymmetric loss functions, we choose the general entropy loss function (GELF) and linex loss function (LLF). The SELF, LLF, and GELF are defined as
where means an estimate of ℵ. In LLF and SELF, the symbols of p and q indicate the direction of the asymmetry, and their sizes mean the different level. Neither of them are zero.
The Bayes estimates of ℵ under above loss functions are
where means the posterior expectation under the parameter ℵ. Now, we can derive the Bayes estimates of , , and entopy under SELF, LLF, and GELF.
To begin with, Bayes estimate of under SELF is
Let takes the value of , , and entropy, then we can easily obtain the corresponding estimation under SELF.
Moreover, Bayes estimate of under LLF is
Let take the value of , , and entropy; then, we can obviously obtain the corresponding estimation under LLF.
Finally, Bayes estimate of under GELF is
Let take the value of , , and entropy; then, we can evidently obtain the corresponding esitimation under GELF.
Obviously, the Bayesian estimation cannot be accurately expressed in a closed form. Hence, we recommend using Lindley method as well as Importance Sampling procedure to derive the Bayesian estimation.
4.3. Lindley Approximation
The Bayes estimates are in the shape of a specific ratio of two integals which can’t be reduced to a closed form. Therefore, we utilize Lindley approximation method to derive the Bayes estimates. For the , the Bayesian estimate is
where
All terms are estimated by MLEs of and . For our problem, we take for the estimation of . Then, we obtain
- (1)
- Squared error loss function
Taking or , the Bayesian estimates of two parameters and under SELF, separately, are obtained by
Next, we derive Bayes estimate of entropy under SELF. We consider that
The Bayes estimator of entropy can be obtained as earlier, and it is given by
- (2)
- Linex loss function
For parameter , taking , we can obtain that
Utilizing the above expression in Equation (28), the Bayesian estimate of is derived by
The Bayesian estimate of is obtained likewise:
Next, we derive the Bayesian estimate of entropy. It is clear that
The requested estimation of entropy can be derived in a similar method.
- (3)
- General entropy loss function
For the parameter ,
Applying the above expression in Equation (28), the Bayesian estimate of is derived by
The approximate Bayes estimator of is computed likewise.
Finally, we derive Bayesian estimate of entropy under GELF:
The Bayes estimator of entropy under GELF can be calculated as earlier, and it is given by
4.4. Importance Sampling Procedure
Using the Lindley approximation method, we can get the Bayesian estimates of the unknown parameters and entropy. Although the Lindley method can make point estimation, it cannot determine the Highest Posterior Density (HPD) credible intervals. Thus, we recommend using the Importance Sampling to get Bayesian estimates and to derive HPD credible intervals as well.
To begin with, let’s solve the doubts before. If we choose two Gammas for prior distributions, record it as and . Then, the joint prior distribution is
Correspondingly, the joint posterior distribution is
where .
We observe that seems like Gamma distribution, but the second parameter can not be proven to be strictly positive, so it cannot be considered to be a Gamma distribution. Obviously, it is not possible to generate its random samples according to the Gamma distribution, and it is also difficult to generate its random samples using other methods. Therefore, it is not appropriate to choose both Gammas as priors.
Then, we return to the prior distribution we selected before. To implement the Importance Sampling, the joint posterior distribution can be adapted as
where is a Gamma function and
Here, K is a normalizing constant and
Note that, in order to get the Bayesian estimates of parameters using the Importance Sampling, we demand to produce corresponding samples from and . It is uncomplicated and clear to produce samples from because it is a simple Gamma distribution. As for producing samples from , we have a Lemma.
Lemma 1.
is log-concave.
Proof.
Since m is a postive number, the second-order partial derivative of is constantly negative. Thereby, is log-concave. □
Then, using the approach originally proposed by [16], we can easily produce samples from .
Using the following steps, we can produce several samples from the request scenario:
- Produce from .
- Produce from .
- Repeat Step 1 and 2 to derive .
Then, the required Bayesian estimate of ℧(,) can be represented by
Furthermore, samples produced above can also be chosen to establish the HPD intervals for the parameters and entropy. Suppose that , and makes . For a given p, we purpose an approach to make a estimation of and then to establish the HPD intervals for .
Firstly, we suppose
For simplicity, we replace with . Then, sort into , where and is related to for . Then, the Bayesian estimate of is , where is an integer which satifies
Therefore, a HPD interval of can be derived by for . Finally, a HPD interval of transforms , where makes
for all .
The next section will use Monte Carlo simulation to numerically and systematically compare previously proposed estimators.
5. Simulation Results
We will use the Monte Carlo simulation method to analyze the behavior of different estimators obtained by the above sections based on the expected value (EV) and mean squared error (MSE). The progressive first-failure censored samples are produced from different censoring schemes of and various parameter values from the IWD by using the algorithm originally proposed by [17].
In general, we let , , and correspondingly the entropy is 1.172676. We use the ‘optim’ command in the R software (version 3.6.1, Lucent Technologies, Mary Hill, NJ, USA) to get the approximate MLEs of , , and entropy presented in Table 1. The Bayesian estimates under both asymmetric and symmetric loss functions are precisely computed by the Lindley method and Importance Samplings. For the Bayes estimation, we assign the value of hyperparameters as for Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7 and for Table 8 and Table 9. Under the LLF, we let and . Under the GELF, we choose and . We derive asymmetric intervals of parameters using the MLEs and log-transformed MLEs and HPD intervals. Pay attention that, for simplicity, the censoring schemes are presented by abbreviations such as represents and represents . Table 2, Table 3, Table 4, Table 5 and Table 6, Table 8 present the Bayes estimation of , , and entropy using the Lindley method. The Bayes estimation based on Importance Samplings is shown in Table 7 and Table 9. In Table 10, the interval estimation of entropy is presented.
Table 1.
Maximum likelihood estimates of , , and entropy when , , .
Table 2.
Bayes estimates under squared error loss function of , , and entropy based on the Lindley method when , , .
Table 3.
Bayes estimates under Linex loss function of , , and entropy based on Lindley method when , , , .
Table 4.
Bayes estimates under Linex loss function of , , and entropy based on Lindley method when , , , .
Table 5.
Bayes estimates under general entropy loss function of , , and entropy based on Lindley method when , , , .
Table 6.
Bayes estimates under general entropy loss function of , , and entropy based on Lindley method when , , , .
Table 7.
Bayes estimates of entropy using Importance Sampling when , , .
Table 8.
Bayes estimates under squared error loss function of , , and entropy based on Lindley method when , , H = 1.172676, a = 0, b = 0.
Table 9.
Bayes estimates of entropy using Importance Sampling procedure when , , , a = 0, b = 0.
Table 10.
Average length and coverage probability of asymptotic intervals/highest posterior density credible intervals of paramater and entropy when , , , k = 1.
As a whole, the EVs and MSEs of parameters and entropy all significantly decrease as the sample size n increases. In Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8 and Table 9, set m and n invariant, the EVs and MSEs of parameters and entropy both decrease as the group size k increases. Furthermore, set k and n invariant, the EVs and MSEs of parameters and entropy both decrease as m increases. Bayesian estimates with perform more precise than , which is so-called non-informative. Using MLE and Bayes estimation based on the Lindley method is better than the Importance Sampling procedure. Bayes estimation using the Lindley method is a little bit more precise than the MLE. For LLF, choosing seems to be better than . For GELF, competes as well as . In Table 7 and Table 9, we observe that the few censoring schemes such as and do not compete well.
In Table 10, the average length (AL) narrows down as the sample size n increases. Moreover, HPD intervals are more precise than confidence intervals based on AL. For confidence intervals, using log-transformed MLEs performs much better than MLEs. In almost all circumstances, the coverage probability (CP) of entropy derived here achieve their specified confidence intervals.
6. Real Data Analysis
We will analyze a real data set and apply the approaches put forward in the sections above. The data set was analyzed by [7,18]. The data show the surviving days of guinea pig injected with vairous species of tubercle bacilli. The quantity of regimen is the logarithmic of the quantity of bacillary units in 0.5 mL of the challenging solution. The sample size is 72 which are listed below: 12, 15, 22, 24, 24, 32, 32, 33, 34, 38, 38, 43, 44, 48, 52, 53, 54, 54, 55, 56, 57, 58, 58, 59, 60, 60, 60, 60, 61, 62, 63, 65, 65, 67, 68, 70, 70, 72, 73, 75, 76, 76, 81, 83, 84, 85, 87, 91, 95, 96, 98, 99, 109, 110, 121, 127, 129, 131, 143, 146, 146, 175, 175, 211, 233, 258, 258, 263, 297, 341, 341, 376 (unit: days).
Before analyzing the data, we want to test if the IWD matches the complete data well. To begin with, from [7], we conclude that the failure rate function of this data are unimodal, so it is scientific and reasonable to analyze the data using IWD. Then, we choose various approaches to analyze the goodness of fit of IWD using the MLE. We compute the and Kolmogorov–Smirnov (K–S) statistics with its associated p-value represented in Table 11. According to the p-value, the IWD fits the complete data well.
Table 11.
Summary for model fit using , K-S statistic and associated p-value.
Now, we can consider the censoring data to illustrate the previous approaches. To generate the first-failure censored sample, we randomly sort the given data into groups with identical units in each group, and we can get the first-failure censored sample: 12, 15, 22, 24, 32, 32, 33, 34, 38, 38, 43, 44, 48, 52, 54, 55, 56, 58, 58, 60, 60, 61, 63, 65, 65, 68, 70, 70, 73, 76, 84, 91, 109, 110, 129, 143. Then, we produce samples using three diffrent progressive first-failure censoring which are , and from the above sample with . The results are organized in Table 12.
Table 12.
Progressive first-failure censored sample in the given censoring scheme when , , .
In Table 13, for MLE, we calculate the EVs, MSEs, and confidence intervals of the parameters and entropy; for Bayes estimation, we obtain the EVs, MSEs, and HPD intervals of entropy and two parameters. The estimates of , , and entropy using the MLE and the Importance Sampling method are relatively close.
Table 13.
Point and interval estimation of parameters and entropy using MLE and Bayes methods.
7. Conclusions
In this article, the problem of statistical inference on the parameters and entropy of IWD under progressive first-failure censoring has been considered. Both the maximum likelikood estimation and Bayesian estimation are investigated. For Bayesian estimation, we apply the Lindley and Importance Sampling method to approximate the Bayesian estimates under both asymmetric and symmetric loss functions. We construct the approximate intervals based on MLEs and Log-transformed MLEs. In addition, we use the Importance Sampling method to derive the HPD intervals. Then, we compare the performance of estimates through EV and MSE. Although we have considered the estimation of entropy under progressive first-failure censoring scheme as much as possible, using a similar method, this censoring scheme can be widely extended to other more efficient and complex censoring schemes. This direction is still very promising and requires more attention and work.
Author Contributions
Methodology and Writing, J.Y.; Supervision, W.G.; Writing, Y.S.
Funding
This research was supported by Project 201910004093 which was supported by National Training Program of Innovation and Entrepreneurship for Undergraduates.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Wu, S.J.; Kuş, C. On estimation based on progressive first-failure-censored sampling. Comput. Stat. Data Anal. 2009, 53, 3659–3670. [Google Scholar] [CrossRef]
- Soliman, A.A.; Abou-Elheggag, N.A.; Ellah, A.H.A.; Modhesh, A.A. Bayesian and non-Bayesian inferences of the Burr-XII distribution for progressive first-failure censored data. Metron 2014, 70, 1–25. [Google Scholar] [CrossRef]
- Dube, M.; Krishna, H.; Garg, R. Generalized inverted exponential distribution under progressive first-failure censoring. J. Stat. Comput. Simul. 2015, 86, 1–20. [Google Scholar] [CrossRef]
- Singh, S.; Tripathi, Y.M. Reliability sampling plans for a lognormal distribution under progressive first-failure censoring with cost constraint. Stat. Pap. 2015, 56, 773–817. [Google Scholar] [CrossRef]
- Bakoban, R.A.; Abd-Elmougod, G.A. MCMC in analysis of progressively first failure censored competing risks data for gompertz model. J. Comput. Theor. Nanosci. 2016, 13, 6662–6670. [Google Scholar] [CrossRef]
- Ahmadi, M.V.; Doostparast, M. Pareto analysis for the lifetime performance index of products on the basis of progressively first-failure-censored batches under balanced symmetric and asymmetric loss functions. J. Appl. Stat. 2018, 46, 1–32. [Google Scholar] [CrossRef]
- Kundu, D.; Howlader, H. Bayesian inference and prediction of the inverse Weibull distribution for Type-II censored data. Comput. Stat. Data Anal. 2010, 54, 1547–1558. [Google Scholar] [CrossRef]
- Sultan, K.S.; Alsadat, N.H.; Kundu, D. Bayesian and maximum likelihood estimations of the inverse Weibull parameters under progressive type-II censoring. J. Stat. Comput. Simul. 2014, 84, 2248–2265. [Google Scholar] [CrossRef]
- Musleh, R.M.; Helu, A. Estimation of the inverse Weibull distribution based on progressively censored data: Comparative study. Reliab. Eng. Syst. Saf. 2014, 131, 216–227. [Google Scholar] [CrossRef]
- Singh, S.; Tripathi, Y.M. Estimating the parameters of an inverse Weibull distribution under progressive type-I interval censoring. Stat. Pap. 2018, 59, 21–56. [Google Scholar] [CrossRef]
- Mazucheli, J.; Menezes, A.F.B.; Dey, S. Bias-corrected maximum likelihood estimators of the parameters of the inverse Weibull distribution. Commun. Stat. Simul. Comput. 2018, 48, 2046–2055. [Google Scholar] [CrossRef]
- AboEleneen, Z.A. The Entropy of Progressively Censored Samples. Entropy 2011, 3, 437–449. [Google Scholar] [CrossRef]
- Kayal, S.; Kumar, S. Estimation of the Shannon’s entropy of several shifted exponential populations. Stat. Probab. Lett. 2013, 83, 1127–1135. [Google Scholar] [CrossRef]
- Cho, Y.; Sun, H.; Lee, K. An Estimation of the Entropy for a Rayleigh Distribution Based on Doubly-Generalized Type-II Hybrid Censored Samples. Entropy 2014, 16, 3655–3669. [Google Scholar] [CrossRef]
- Meeker, W.Q.; Escobar, L.A. 8—Maximum Likelihood Methods for Fitting Parametric Statistical Models. Methods Exp. Phys. 1994, 28, 211–244. [Google Scholar]
- Devroye, L. A simple algorithm for generating random variates with a log-concave density. Computing 1984, 33, 247–257. [Google Scholar] [CrossRef]
- Balakrishnan, N.; Sandhu, R.A. A Simple Simulational Algorithm for Generating Progressive Type-II Censored Samples. Am. Stat. 1995, 49, 229–230. [Google Scholar]
- Singh, S.K.; Singh, U.; Kumar, D. Bayesian estimation of parameters of inverse Weibull distribution. J. Appl. Stat. 2013, 40, 1597–1607. [Google Scholar] [CrossRef]
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
