Next Article in Journal
Deforming Gibbs Factor Using Tsallis q-Exponential with a Complex Parameter: An Ideal Bose Gas Case
Previous Article in Journal
Enhancing the Robustness of Dynamical Decoupling Sequences with Correlated Random Phases
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Statistical Inference for the Inverted Scale Family under General Progressive Type-II Censoring

Department of Mathematics, Beijing Jiaotong University, Beijing 100044, China
*
Author to whom correspondence should be addressed.
Symmetry 2020, 12(5), 731; https://doi.org/10.3390/sym12050731
Submission received: 31 March 2020 / Revised: 15 April 2020 / Accepted: 17 April 2020 / Published: 5 May 2020

Abstract

:
Two estimation problems are studied based on the general progressively censored samples, and the distributions from the inverted scale family (ISF) are considered as prospective life distributions. One is the exact interval estimation for the unknown parameter θ , which is achieved by constructing the pivotal quantity. Through Monte Carlo simulations, the average 90 % and 95 % confidence intervals are obtained, and the validity of the above interval estimation is illustrated with a numerical example. The other is the estimation of R = P ( Y < X ) in the case of ISF. The maximum likelihood estimator (MLE) as well as approximate maximum likelihood estimator (AMLE) is obtained, together with the corresponding R-symmetric asymptotic confidence intervals. With Bootstrap methods, we also propose two R-asymmetric confidence intervals, which have a good performance for small samples. Furthermore, assuming the scale parameters follow independent gamma priors, the Bayesian estimator as well as the HPD credible interval of R is thus acquired. Finally, we make an evaluation on the effectiveness of the proposed estimations through Monte Carlo simulations and provide an illustrative example of two real datasets.

1. Introduction

General progressive censoring plays a key role in the field of lifetime analysis. It generalizes the progressive censoring scheme, which allows removing surviving units from a life test stage-by-stage and to some extent, resolves the time and money constraints (Sen, [1]). Additionally, it contains two censoring schemes: left censoring and right censoring. Here considers the situation of general progressive Type-II right censoring (GPTIIC).
In a life-test, first randomly select N units as the test sample. Suppose that the first r failures are unobserved; for i = r + 1 , , n 1 , while the i t h failure is observed, R i surviving units are withdrawn from the test randomly; while the n t h failure is observed, all remaining units are withdrawn and the life-test is terminated, that is, R n = N n i = r + 1 n 1 R i . Denote the totally observed failure times as X r + 1 : n : N X n : n : N , which are also called the general progressively Type-II censored order statistics (GPTIICOS). For notational convenience, represent it with X r + 1 X n . Here, denote the corresponding censoring scheme as ( r , n , N , R r + 1 , , R n ) . Note that: (i) GPTIIC with ( 0 , n , N , 0 , , 0 , N n ) represents the conventional Type-II right censoring scheme. (ii) GPTIIC with ( 0 , n , N , R 1 , , R n ) represents the progressive Type-II right censoring scheme. The reader can find more information by consulting the book by Balakrishnan and Aggarwala [2], which has discussed some mathematical properties and inferential methods based on this general progressive censoring scheme.
Consider a class of distributions that have the inverted form of the standard probability density functions (PDF) and distribution functions (CDF) for the scale family. We call it the inverted scale family (ISF). It is said that a positive and continuous random variable X with CDF (Equation (1)) and PDF (Equation (2)) belongs to the inverted scale family (i.e., ISF( θ )) with the scale parameter θ .
F ( x , θ ) = 1 G ( θ x ) , θ > 0 , x > 0
f ( x , θ ) = θ x 2 g ( θ x ) , θ > 0 , x > 0
Note that a positive and continuous random variable Y with CDF G ( y ) and PDF g ( y ) is said to come from a standard scale distribution.
Distributions from ISF are essential for the reliability life data analysis. For example, the generalized inverse exponential distribution (GIED) from ISF, the case of the inverse exponential distribution (IED) included, is able to model the aging criteria of various shapes. In addition, a reliability model based on the inverse Weibull distribution (IWD) from ISF may be suitable when the empirical studies on a given data set indicate the probability of a unimodal hazard function. In light of the importance of ISF, statistical inference on the related distributions has appealed to many authors. For the inverted gamma distribution, Lin [3] acquired the MLE as well as the estimation of the corresponding reliability function. In the case of Type-II censoring, Kundu [4] dealt with the Bayesian estimation for IWD. Prakash [5] studied some properties for IED from a Bayesian point of view. Moreover, articles by Krishna [6], Dey [7], and Essam [8] are several recent works on the estimation for GIED models under the progressive censoring.
Although the scheme of GPTIIC and distributions from ISF both play a vital role in life researches, not much attention has been paid to the situation when the samples are from ISF under GPTIIC. So, in this case, we consider the following two estimation problems.
One is the interval estimation, which, as a classical estimation problem, has engaged the attention of many authors. Article by Viveros and Balakrishnan [9] derived the interval estimates for some life-time parameters. Balakrishnan et al. [10] obtained interval estimation for Normal distribution in the case of progressive Type-II censoring. Wang [11] conducted a study on the scale family under GPTIIC and proposed the exact interval estimation. Dey [7] derived interval estimations for GIED under progressive first censoring.
The other is the estimation of R = P ( Y < X ) , which denotes the stress-strength parameter. Suppose a random stress acts on a system; for constructing a stress-strength modeling, we denote the system strength as X and the value of the random stress as Y. The system will fail only when its strength is less than the received stress. Thus, R = P ( Y < X ) can reflect the system reliability. As R provides a wide application in many areas, it is necessary to discuss the related estimation problems in the statistical literature. Articles by Church and Harris [12], Downtown [13], Govidarajulu [14], Woodward and Kelley [15], and Owen et al. [16] dealt with the situation of X and Y coming from normal distributions. Raqab and Kundu [17], Kundu and Gupta, [18] and Kundu and Raqad [19] conducted studies on this estimation problem for the generalized Rayleigh distributions, Weibull distributions, and three-parameter Weibull distributions, respectively.
Here we conduct a study on the exact interval estimation of θ as well as the estimation of R for ISF under GPTIIC. First, the exact interval estimation of θ is achieved in Section 2 by constructing the pivotal quantity. Then we make an assessment on the interval estimation performance through Monte Carlo simulations and further illustrate the estimation process with a remission time dataset. Moreover, the estimation of R = P ( Y < X ) is achieved in Section 3. The MLE, AMLE, and Bayesian estimator of R are acquired, together with the corresponding interval estimations. In addition, with Bootstrap methods, we derive two interval estimates, suitable for small samples. Finally, we evaluate the effectiveness of the proposed estimations through Monte Carlo simulations and provide an illustrative example of two real datasets.

2. Exact Interval Estimation for θ

2.1. Interval Estimation with Pivotal Quantity Method

Here, a theorem (proposed by Aggarwala and Balakrishnan [2,20]) is introduced to construct the pivotal quantity.
Theorem 1.
Consider U r + 1 , U r + 2 , , U n are the GPTIICOS from the Uniform (0,1) distribution. Then
L r + i = 1 U n i + 1 1 U n i , i = 1 , , n r 1 , L n = 1 U r + 1 .
follow independent beta distributions given by Equation (3).
L r + i B e t a ( i + j = n i + 1 n R j , 1 ) , i = 1 , , n r 1 , L n B e t a ( N r , r + 1 ) .
Lemma 1.
Consider X r + 1 X n are the GPTIICOS from the inverted scale family. Then
L r + i = G ( θ X n i + 1 ) G ( θ X n i ) , i = 1 , , n r 1 , L n = G ( θ X r + 1 )
follow independent beta distributions given by Equation (3).
Proof. 
Since F ( X , θ ) , ranging from 0 to 1, increases as X increases, F ( X r + 1 , θ ) F ( X n , θ ) are the GPTIICOS from the Uniform (0,1) distribution. For i = r + 1 , , n , represent U i in the Theorem 1 with F ( X i , θ ) . The proof is thus completed. ☐
Denote the CDF of B e t a ( N r , r + 1 ) as H ( x ) . Then independent random variables
Q r + i = L r + i i + j = n i + 1 n R j , i = 1 , , n r 1 , Q n = H ( L n )
are from the Uniform (0,1) distribution. Here, L r + i is described as Equation (4). Thus
2 log ( Q r + 1 ) , 2 log ( Q r + 2 ) , , 2 log ( Q n )
are independent χ 2 ( 2 ) random variables. Here, χ 2 ( d f ) represents the χ 2 distribution with d f degrees of freedom. Then we construct the pivotal quantity as:
T ( θ ) = i = 1 n r [ 2 log ( Q r + i ) ] = 2 i = r + 2 n ( R i + 1 ) log [ 1 F ( X i , θ ) ] 2 log [ H ( 1 F ( X r + 1 , θ ) ) ] + 2 i = r + 2 n ( R i + 1 ) log [ 1 F ( X r + 1 , θ ) ] .
It is clear that T ( θ ) follows χ 2 ( 2 n 2 r ) . For the inverted scale family, when r = 0 , T ( θ ) reduces to
T ( θ ) = 2 i = 1 n ( R i + 1 ) log [ 1 F ( X i , θ ) ] = 2 i = 1 n ( R i + 1 ) log [ G ( θ X i ) ] .
As can be seen from Equation (6), for r = 0 , T ( θ ) strictly decreases as θ increases. We can also find that for any positive integer r < n , the exact confidence interval can be derived from Equation (5) easily when T ( θ ) is strictly monotonic as θ increases. Here we consider the failure rate function λ ( x , θ ) for the inverted scale family to determine the monotonicity of T ( θ ) . The following is a brief overview of λ ( x , θ ) .
Suppose a positive continuous random variable X from ISF( θ ) as the lifetime of some units in a life-test. Then, the failure rate function represents the probability intensity that an alive unit will fail at the time x, which is defined as:
λ ( x , θ ) = f ( x , θ ) [ 1 F ( x , θ ) ] , θ > 0 , x > 0 .
For the inverted scale family, it becomes
λ ( x , θ ) = θ x 2 g ( θ x ) G ( θ x ) .
Here, λ ( x , θ ) can be seen as a function of θ .
According to λ ( x , θ ) , a theorem is given below. Note that T ( θ ) (as shown in Equation (5)) is strictly decreasing on θ ( 0 , + ) if the condition given in Theorem 2 is satisfied.
Theorem 2.
Consider X r + 1 , , X n are the GPTIICOS from ISF(θ) with censoring scheme ( r , n , N , R r + 1 , , R n ) . The pivotal quantity T ( θ ) (as shown in Equation (5)) is strictly decreasing on ( 0 , + ) if the failure rate function λ ( x , θ ) is strictly decreasing on θ ( 0 , + ) .
Proof. 
Let h ( t ) = t g ( t ) G ( t ) and for i = r + 2 , , n , let y i ( θ ) = [ l o g ( 1 F ( X i , θ ) ) l o g ( 1 F ( X r + 1 , θ ) ) ] . The pivotal quantity T ( θ ) becomes
2 i = r + 2 n ( R i + 1 ) y i ( θ ) 2 l o g [ H ( G ( θ X r + 1 ) ) ] .
The failure rate function from Equation (7) can be rewritten as:
λ ( x , θ ) = 1 x h ( θ x ) .
According to Equation (8), since λ ( x , θ ) strictly decreases as θ increases, we can infer that h ( t ) is strictly decreasing on t ( 0 , + ) . Thus the monotonicity of y i ( θ ) can be determined. The derivative of y i ( θ ) is:
y i ( θ ) = g ( θ x r + 1 ) x r + 1 G ( θ x r + 1 ) g ( θ x i ) x i G ( θ x i ) = 1 θ [ h ( θ x i ) h ( θ x r + 1 ) ] < 0 .
From the above equation, it is clear that 2 i = r + 2 n ( R i + 1 ) y i ( θ ) is strictly decreasing on θ ( 0 , + ) . Considering the monotonicity of the CDF, we can find 2 l o g [ H ( G ( θ X r + 1 ) ) ] is also strictly decreasing on θ ( 0 , + ) . Thus, we complete the proof. ☐
The exact confidence interval of θ is given below.
Theorem 3.
For any ξ ( 0 , 1 ) , if T ( θ ) is strictly decreasing on ( 0 , + ) , thus
[ T 1 ( χ 1 ξ / 2 2 ( 2 n 2 r ) ) , T 1 ( χ ξ / 2 2 ( 2 n 2 r ) ) ] .
is an exact 1 ξ confidence interval for θ. Here, T 1 ( x ) denotes the solution to T ( θ ) = x for any x ( 0 , ) and χ ξ 2 ( d f ) denotes the ξ percentile of χ 2 ( d f ) .
Proof. 
Since T ( θ ) is a strictly decreasing function on ( 0 , + ) , the probability that θ belongs to the interval of Equation (10) is:
P ( T 1 ( χ 1 ξ / 2 2 ( 2 n 2 r ) ) θ T 1 ( χ ξ / 2 2 ( 2 n 2 r ) ) = P ( χ ξ / 2 2 ( 2 n 2 r ) T ( θ ) χ 1 ξ / 2 2 ( 2 n 2 r ) ) = 1 ξ .
 ☐
Remark 1.
For any ξ ( 0 , 1 ) , if T ( θ ) is strictly increasing on ( 0 , + ) , thus
[ T 1 ( χ ξ / 2 2 ( 2 n 2 r ) ) , T 1 ( χ 1 ξ / 2 2 ( 2 n 2 r ) ) ] .
is an exact 1 ξ confidence interval for θ. Here, T 1 ( x ) denotes the solution to T ( θ ) = x for x ( 0 , ) and χ ξ 2 ( d f ) denotes the ξ percentile of χ 2 ( d f ) .
Example 1.
The generalized inverse exponential distribution (i.e., GIED( θ , α ) ), which includes the inverse exponential distribution (i.e., IED(θ)) as a special case when α = 1 , is from the inverted scale family with
F ( x ; θ , α ) = 1 [ 1 e x p ( θ x ) ] α , θ , α > 0 , x > 0 , f ( x ; θ , α ) = α θ x 2 e x p ( θ x ) [ 1 e x p ( θ x ) ] α 1 , θ , α > 0 , x > 0 .
By Theorem 2, we have h ( t ) = α t e x p ( t ) 1 with the derivative h ( t ) = α [ e x p ( t ) t e x p ( t ) 1 ] [ e x p ( t ) 1 ] 2 . Let h 1 ( t ) = e x p ( t ) t e x p ( t ) 1 , t > 0 , then, a system of equations is given by
h 1 ( t ) = e x p ( t ) t e x p ( t ) 1 h 1 ( t ) = t e x p ( t ) < 0 h 1 ( 0 ) = 0
From Equation (12), when t ( 0 , + ) , we have h 1 ( t ) < 0 and thus h ( t ) < 0 , which proves that h ( t ) is strictly decreasing on t ( 0 , + ) . So λ ( x , θ ) = 1 x h ( θ x ) is strictly decreasing on θ ( 0 , + ) .
For the generalized inverse exponential distribution, from Equation (5), the pivotal quantity is developed as:
T ( θ ) = 2 α i = r + 2 n ( R i + 1 ) log [ 1 e x p ( θ X r + 1 ) 1 e x p ( θ X i ) ] 2 log [ H ( 1 e x p ( θ X r + 1 ) ) α ] .
When r = 0 , T ( θ ) reduces to
T ( θ ) = 2 α i = 1 n ( R i + 1 ) log [ 1 e x p ( θ X i ) ] .
Then applying Theorem 3, we achieve the exact interval estimation.
Example 2.
The inverse Weibull distribution (i.e., IWD( θ , α )), including the inverse exponential distribution (when α = 1 ) and the inverse Rayleigh distribution (IRD(θ), when α = 2 ) as special cases, is from the inverted scale family with
F ( x ; θ , α ) = e x p [ ( θ x ) α ] , θ , α > 0 , x > 0 , f ( x ; θ , α ) = α θ ( θ x ) α + 1 e x p [ ( θ x ) α ] , θ , α > 0 , x > 0 .
By Theorem 2, we have h ( t ) = α t α e x p ( t α ) 1 with the derivative h ( t ) = α 2 t α 1 [ e x p ( t α ) t α e x p ( t α ) 1 ] [ e x p ( t α ) 1 ] 2 . Let h 2 ( t ) = e x p ( t α ) t α e x p ( t α ) 1 , t > 0 , then, a system of equations is given by:
h 2 ( t ) = e x p ( t α ) t α e x p ( t α ) 1 h 2 ( t ) = α t 2 α 1 e x p ( t α ) < 0 h 2 ( 0 ) = 0
Clearly, when t ( 0 , + ) , we have h 2 ( t ) < 0 and thus h ( t ) < 0 , which proves that h ( t ) is strictly decreasing on t ( 0 , + ) . So λ ( x , θ ) = 1 x h ( θ x ) is strictly decreasing on θ ( 0 , + ) .
For the inverse Weibull distribution, from Equation (5), the pivotal quantity is developed as:
T ( θ ) = 2 i = r + 2 n ( R i + 1 ) log [ 1 e x p ( ( θ X r + 1 ) α ) 1 e x p ( ( θ X i ) α ) ] 2 log [ H ( 1 e x p ( ( θ X r + 1 ) α ) ) ] .
When r = 0 , T ( θ ) reduces to
T ( θ ) = 2 i = 1 n ( R i + 1 ) log [ 1 e x p ( ( θ X i ) α ) ] .
Then applying Theorem 3, we achieve the exact interval estimation.

2.2. Simulation Study

Here, the exact interval estimation for ISF under GPTIIC is implemented through Monte Carlo simulations. Take IED as a prospective life distribution (the CDF and PDF are given in Example 1). With interval coverage percentages, we evaluate the validity of the exact confidence interval derived in Theorem 3.
According to Balakrishnan [2], GPTIICOS from IED( θ ) can be generated by the following algorithm.
Step 1:
Generate a random variable L n from B e t a ( N r , r + 1 ) .
Step 2:
Generate n r 1 independent Uniform (0,1) random variables Z r + 1 , , Z n 1 .
Step 3:
Set L r + i = Z 1 a r + i where a r + i = i + j = n i + 1 n R j , i = 1 , , n r .
Step 4:
Set U r + i = 1 L n i + 1 L n i + 2 L n for i = 1 , , n r .
Step 5:
Set X r + i = θ log ( U r + i ) for i = 1 , , n r .
X r + 1 , , X n thus denote the GPTIICOS from IED( θ ) with the censoring scheme ( r , n , N , R r + 1 , , R n ) .
Here, suppose X r + 1 , , X n are from IED(5). For evaluating the performance of the exact confidence interval estimation, we consider the cases of N = 20 , N = 30 and N = 40 as well as four different ( r , n ) , including ( 1 , 10 ) , ( 3 , 12 ) , ( 1 , 16 ) , and ( 3 , 18 ) . In this case, we give the average confidence intervals for θ and the corresponding interval lengths and interval coverage percentages with 10,000 simulations. The results are shown in Table 1. Here, the considered confidence levels are 90 % and 95 % .
As can be seen from Table 1, the obtained interval coverage percentage is close to the corresponding confidence level. This indicates the proposed interval estimation has a good performance. Additionally, the simulation results show the relationship between the confidence interval length and the values of N , r , n :
(a) When n r (i.e., the number of completely observed failures) and N (i.e., the sample size) are fixed, the confidence interval length is, to some extent, related to r (that is, the number of the first unobserved failures), as shown in Table 1 that the average confidence interval is wider when r is larger.
(b) When the values of r and n are fixed, the interval length is related to N as it is larger when N is smaller.
(c) When the values of N and r are fixed, the interval length is related to n as it is smaller when n is larger.
Obviously, a narrow confidence interval can provide more information about the unknown parameter. We conclude that the estimation is more accurate when more samples are completely observed for the confidence interval length decreases as the number of the completely observed samples increases.

2.3. An Illustrated Example

A real dataset, reported by Gross and Declark [21], is provided below for illustrating the proposed estimation process in the previous subsections. It shows the remission time of 20 patients receiving analgesics (in minutes):
1.1 1.2 1.3 1.4 1.4 1.5 1.6 1.6 1.7 1.7
1.7 1.8 1.8 1.9 2.0 2.2 2.3 2.7 3.0 4.1
Here, we consider three reliability models based on IWD (as shown in Example 2), GIED (as shown in Example 1), and the Weibull distribution (i.e., WD, the CDF and PDF are given below), respectively.
F ( x ; θ , α ) = 1 e θ x α , θ , α > 0 , x > 0
f ( x ; θ , α ) = α θ x α 1 e θ x α , θ , α > 0 , x > 0
We use maximum likelihood estimation to construct the above three parametric models, and for each model, five measures are used to test the goodness-of-fit: (1) −ln, (2) K S distance, (3) p-value, (4) A I C , and (5) B I C . Here, for each model, −ln can be calculated with the data and MLEs of the corresponding parameters. The following briefly introduces A I C and B I C .
(i) A I C , proposed by Akaike [22], can reflect the fitting effect of a reliability model to the data. The equation is given below:
A I C = 2 m 2 l n .
Here, for an estimated parametric model, m denotes the number of parameters, and l n denotes the maximum value of the log-likelihood function. For a given data set, the best fit order of several competing models can be given from their A I C , and the one with lower A I C ranks higher.
(ii) B I C , proposed by Schwarz [23], is a selection criterion for choosing between several estimated statistical models with different numbers of parameters. The equation is given by:
B I C = m log ( n ) 2 l n .
Here, for an estimated parametric model, m denotes the number of parameters, n denotes the number of observed failures, and l n denotes the maximum value of the log-likelihood function. B I C has a close relationship with the A I C . When a data set is given, same as in A I C , the best fit order of several competing models can be given from their B I C , and the one with lower B I C ranks higher. However, compared with A I C , B I C has a greater penalty for additional parameters.
In the case of MLE, the values of the above five measures based on the three reliability models are listed in Table 2. Clearly, the best fit order for the given data set is:
B E S T I W D G I E D W D W O R S T
Since IWD has the lowest values of −ln, K S distance, A I C , B I C , and the highest p-value, it is considered the best of the three reliability models. Furthermore, the fitting effects of the three reliability models are intuitively shown in Figure 1. Hence, we fit the IWD model to the data set. It appears from Table 2 that the values of −ln, K S distance, p-value, A I C , and B I C in the case of MLE are 15.19896 , 0.1306 , 0.9187 , 34.39793 , and 36.17867 , respectively, indicating the IWD model fits well to the data when α takes the value of 3.777054 . Hence, for estimating the exact confidence interval of θ , we can assume that the sample follows IWD( θ ,3.777054). Set r = 2 , n = 16 , and then Table 3 shows the censoring scheme and the generated censored data.
We compute the exact confidence interval for θ with Equation (5). Here, the 90 % and 95 % confidence intervals are obtained to be ( 1.251123 , 1.817043 ) and ( 1.208818 , 1.874985 ) , respectively.

3. Estimation of R = P ( Y < X )

Consider the situation when the distributions of independent random variables X and Y are ISF( θ 1 ) and ISF( θ 2 ) (the CDF and PDF are shown in Section 1), respectively.
Then, the stress-strength parameter is described as:
R = P ( Y < X ) = 0 + d x 0 x θ 1 x 2 θ 2 y 2 g ( θ 1 x ) g ( θ 2 y ) d y = 0 + θ 1 x 2 g ( θ 1 x ) [ 1 G ( θ 2 x ) ] d x = R ( θ 1 , θ 2 ) .
Suppose X r + 1 X n are the GPTIICOS from X with censoring scheme ( r , n , N , R r + 1 , , R n ) ; Y r + 1 Y m are the GPTIICOS from Y with censoring scheme ( r , m , M , R r + 1 , , R m ) . Based on the above samples, we discuss the estimation of R.

3.1. Maximum Likelihood Estimator (MLE)

In order to get the MLE of R, the MLEs of θ 1 and θ 2 need to be derived. Here, denote the MLEs of θ 1 , θ 2 , R as θ 1 ^ , θ 2 ^ , R ^ for convenience.
The likelihood function is described as Equation (15) based on the above samples.
L θ 1 , θ 2 = c 1 F x r + 1 r i = r + 1 n f x i 1 F x i R i × c 2 F y r + 1 r j = r + 1 m f y j 1 F y j R j ,
where
c 1 = N r ( N r ) j = r + 2 n N i = r + 1 j 1 R i j + 1 ,
c 2 = M r ( M r ) j = r + 2 m M i = r + 1 j 1 R i j + 1 .
It seems impossible to get the MLEs of θ 1 and θ 2 explicitly from the likelihood Equations (16) and (17). However, we can solve Equations (16) and (17) with the R programming language to obtain θ 1 ^ and θ 2 ^ . Then R ^ can be expressed as R ^ = R ( θ 1 ^ , θ 2 ^ ) . In the following subsection, explicit solutions of θ 1 and θ 2 will be given with an approximative method.
ln L ( θ 1 , θ 2 ) θ 1 = n r θ 1 r g ( θ 1 x r + 1 ) 1 G ( θ 1 x r + 1 ) 1 x r + 1 + i = r + 1 n [ g ( θ 1 x i ) g ( θ 1 x i ) 1 x i + R i g ( θ 1 x i ) G ( θ 1 x i ) 1 x i ] = 0
ln L ( θ 1 , θ 2 ) θ 2 = m r θ 2 r g ( θ 2 y r + 1 ) 1 G ( θ 2 y r + 1 ) 1 y r + 1 + j = r + 1 m [ g ( θ 2 y j ) g ( θ 2 y i ) 1 y j + R j g ( θ 2 y j ) G ( θ 2 y j ) 1 y j ] = 0

3.2. Approximate Maximum Likelihood Estimator (AMLE)

From the above analysis, getting explicit solutions for θ 1 and θ 2 from the nonlinear Equations (16) and (17) is quite difficult. So we use Taylor expansions to acquire approximate estimators. Denote the AMLEs of θ 1 , θ 2 , R as θ 1 ˜ , θ 2 ˜ , R ˜ , respectively.
Considering the transformation Z = θ X , then its CDF and PDF are respectively given by:
M ( z ) = 1 G ( z ) , z > 0 ,
m ( z ) = g ( z ) , z > 0 .
Remark 2.
M ( z ) is decreasing on ( 0 , + ) .
Proof. 
According to the monotonicity of G ( x ) , the proof is thus completed. ☐
Let z i = θ 1 x i and w j = θ 2 y j . The likelihood equations of θ 1 and θ 2 can be rewritten by:
ln L * ( θ 1 , θ 2 ) θ 1 = n r θ 1 + r θ 1 m ( z r + 1 ) M ( z r + 1 ) z r + 1 + i = r + 1 n [ 1 θ 1 m ( z i ) m ( z i ) z i R i θ 1 m ( z i ) 1 M ( z i ) z i ] = 0 ,
ln L * ( θ 1 , θ 2 ) θ 2 = m r θ 2 + r θ 2 m ( w r + 1 ) M ( w r + 1 ) w r + 1 + j = r + 1 m [ 1 θ 2 m ( w j ) m ( w j ) w j R j θ 2 m ( w i ) 1 M ( w j ) w j ] = 0 .
Consider the following Taylor expansions that keep only the first two terms. Let p 1 ( z i ) = m ( z i ) M ( z i ) , p 2 ( z i ) = m ( z i ) m ( z i ) , p 3 ( z i ) = m ( z i ) 1 M ( z i ) and expand them around the point e i = E ( Z i ) ; let q 1 ( w j ) = m ( w j ) M ( w j ) , q 2 ( w j ) = m ( w j ) m ( w j ) , q 3 ( w j ) = m ( w j ) 1 M ( w j ) and expand them around the point e j = E ( W j ) . Note that for k { 1 , 2 , 3 } , p k ( z i ) and q k ( w j ) are the derivatives of p k ( z i ) and q k ( w j ) , respectively.
p 1 ( z i ) p 1 ( e i ) + p 1 ( e i ) ( z i e i ) = A i B i z i , w h e r e A i = p 1 ( e i ) e i p 1 ( e i ) , B i = p 1 ( e i ) , p 2 ( z i ) p 2 ( e i ) + p 2 ( e i ) ( z i e i ) = C i D i z i , w h e r e C i = p 2 ( e i ) e i p 2 ( e i ) , D i = p 2 ( e i ) , p 3 ( z i ) p 3 ( e i ) + p 3 ( e i ) ( z i e i ) = E i F i z i , w h e r e E i = p 3 ( e i ) e i p 3 ( e i ) , F i = p 3 ( e i ) , q 1 ( w j ) q 1 ( e j ) + q 1 ( e j ) ( w j e j ) = A j B j w j , w h e r e A j = q 1 ( e j ) e j q 1 ( e j ) , B j = q 1 ( e j ) , q 1 ( w j ) q 2 ( e j ) + q 2 ( e j ) ( w j e j ) = C j D j w j , w h e r e C j = q 2 ( e j ) e j q 2 ( e j ) , D j = q 2 ( e j ) , q 3 ( w j ) q 3 ( e j ) + q 3 ( e j ) ( w j e j ) = E j F j w j , w h e r e E j = q 3 ( e j ) e j q 3 ( e j ) , F j = q 3 ( e j ) .
Then we consider the expression for e i and e j . According to the transformation of Z = θ X and the monotonicity of M ( z ) , we have Z r + 1 Z n and M ( Z r + 1 , θ ) M ( Z n , θ ) . For i = r + 1 , , n , let M ( Z i ) = U i . Here U r + 1 U n denote the GPTIICOS from the Uniform (0,1) distribution. Then e i can be approximated as e i = E ( Z i ) M 1 ( E ( U i ) ) , where E ( U i ) can be obtained from the following equations (proposed by Balakrishnan [2]).
E ( U r + i ) = 1 j = n i + 1 n α j , i = 1 , , n r ,
where
α i = a i 1 + a i , i = r + 1 , , n 1 , α n = a n 1 + r + a n , a r + i = i + j = n i + 1 n R j , i = 1 , , n r .
Clearly, e j can be approximated in a similar way.
Combining the above Taylor expansions, the likelihood Equations (20) and (21) then reduce to
ln L * ( θ 1 , θ 2 ) θ 1 = 1 θ 1 { ( n r ) + r ( A r + 1 B r + 1 z r + 1 ) z r + 1 + i = r + 1 n [ ( C i D i z i ) z i R i ( E i F i z i ) z i ] } = 0 , ln L * ( θ 1 , θ 2 ) θ 2 = 1 θ 2 { ( m r ) + r ( A r + 1 B r + 1 w r + 1 ) w r + 1 + j = r + 1 m [ ( C j D j w j ) w j R j ( E j F j w j ) w j ] } = 0 .
Here, we show two theorems to give the explicit solutions for θ 1 and θ 2 .
Theorem 4.
[24] If ln g ( x ) is concave on ( a , b ) , then it is said that the function g ( x ) is log-concave, and the following conditions are satisfied: (i) g ( x ) g ( x ) is monotone decreasing on ( a , b ) . (ii) ( ln g ( x ) ) < 0 .
Remark 3.
It is easy to prove Theorem 4 holds in the case of a = 0 and b = + .
Theorem 5.
[25] Suppose g ( x ) is the density function on ( 0 , + ) and G ( x ) is the distribution function. If g ( x ) is log-concave on ( 0 , + ) , then (i) M ( x ) = 1 G ( x ) is log-concave, (ii) G ( x ) is also log-concave on ( 0 , + ) .
Lemma 2.
If the standard scale density function g ( x ) is log-concave on ( 0 , + ) , then the AMLEs of θ 1 and θ 2 are given in Equation (22) and R ˜ can be obtained with Equation (14).
θ 1 ˜ = α 2 α 2 2 4 ( n r ) α 1 2 α 1 , θ 2 ˜ = β 2 β 2 2 4 ( m r ) β 1 2 β 1 ,
where
α 1 = i = r + 1 n ( R i F i D i x i 2 ) r B r + 1 x r + 1 2 , α 2 = r A r + 1 x r + 1 + i = r + 1 n ( C i R i E i x i ) , β 1 = j = r + 1 m ( R j F j D j y j 2 ) r B r + 1 y r + 1 2 , β 2 = r A r + 1 y r + 1 + j = r + 1 m ( C j R j E j y j ) .
Proof. 
Substitute z i by θ 1 x i and w j by θ 2 y j and then Equation (22) are the solutions to the likelihood equations. We need to prove that θ 1 ˜ and θ 2 ˜ are positive. By Theorem 5, M ( x ) and G ( x ) are log-concave on ( 0 , + ) when g ( x ) is log-concave and we have p 1 ( e i ) , p 2 ( e i ) 0 and p 3 ( e i ) 0 applying Theorem 4. Thus, B i , D 1 0 and F i 0 . It is impossible that B i , D 1 , F i 0 for i = r + 1 , , n . So, α 1 is negative. Thus, from Equation (22), θ 1 ˜ is positive. Similarly, we can prove θ 2 ˜ is also positive.
p 1 ( e i ) = m ( e i ) M ( e i ) , p 2 ( e i ) = g ( e i ) g ( e i ) , p 3 ( e i ) = g ( e i ) G ( e i ) .
 ☐
Note that
(i) in the cases of GIED and IWD (given in Section 2.1, when α > 1 ), g ( x ) is log-concave on ( 0 , + ) ;
(ii) If g ( x ) is not log-concave on ( 0 , + ) , we can take the positive roots of the likelihood equations to get the AMLEs θ 1 ˜ and θ 2 ˜ .

3.3. R-Symmetric and R-Asymmetric Approximate Confidence Intervals

First, we obtain an asymptotic confidence interval, which is symmetric about R ^ , with the asymptotic distributions of θ 1 ^ , θ 2 ^ as well as R ^ . Then considering the obtained confidence interval may not work well for small samples, with Bootstrap methods, we further acquire two R-asymmetric interval estimates.

3.3.1. R-Symmetric Interval Estimation Based on Asymptotic Distributions

The asymptotic distribution of θ ^ = ( θ 1 ^ , θ 2 ^ ) can be derived with the Fisher information matrix of θ = ( θ 1 , θ 2 ) , which is described as follows.
I ( θ ) = E ( 2 l θ 1 2 ) E ( 2 l θ 1 θ 2 ) E ( 2 l θ 1 θ 2 ) E ( 2 l θ 2 2 ) = I 11 I 12 I 21 I 22 ,
where
I 11 = E ( 2 l θ 1 2 ) , I 22 = E ( 2 l θ 2 2 ) , I 12 = I 21 = 0 , 2 l θ 1 2 = n r θ 1 2 r x r + 1 2 g ( θ 1 x r + 1 ) [ 1 G ( θ 1 x r + 1 ) ] + ( g ( θ 1 x r + 1 ) ) 2 [ 1 G ( θ 1 x r + 1 ) ] 2 + i = r + 1 n [ g ( θ 1 x i ) g ( θ 1 x i ) ( g ( θ 1 x i ) ) 2 x i 2 ( g ( θ 1 x i ) ) 2 + R i x i 2 g ( θ 1 x i ) G ( θ 1 x i ) ( g ( θ 1 x i ) ) 2 ( G ( θ 1 x i ) ) 2 ] , 2 l θ 2 2 = m r θ 2 2 r y r + 1 2 g ( θ 2 y r + 1 ) [ 1 G ( θ 2 y r + 1 ) ] + ( g ( θ 2 y r + 1 ) ) 2 [ 1 G ( θ 2 y r + 1 ) ] 2 + j = r + 1 m [ g ( θ 2 y j ) g ( θ 2 y j ) ( g ( θ 2 y j ) ) 2 y j 2 ( g ( θ 2 y j ) ) 2 + R j y j 2 g ( θ 2 y j ) G ( θ 2 y j ) ( g ( θ 2 y j ) ) 2 ( G ( θ 2 y j ) ) 2 ] .
I ( θ ) can further be used to derive the approximate variance-covariance matrix. If all the regularity conditions are satisfied for the distribution from ISF, we give the following asymptotic distributions for θ ^ = ( θ 1 ^ , θ 2 ^ ) and R ^ based on the asymptotic properties of MLE.
Theorem 6.
For n r , m r , we have ( n r ( θ ^ 1 θ 1 ) , m r ( θ ^ 2 θ 2 ) ) N ( 0 , C 1 ( θ 1 , θ 2 ) ) in distribution where C ( θ 1 , θ 2 ) = c 11 0 0 c 22 with c 11 = I 11 n r , c 22 = I 22 m r .
Theorem 7.
For n r , m r , we have n r ( R ^ R ) N ( 0 , A ) where
A = b T C 1 b , C 1 ( θ 1 , θ 2 ) = n r I 11 0 0 m r I 22 , b = b 1 ( θ 1 , θ 2 ) b 2 ( θ 1 , θ 2 ) , b 1 ( θ 1 , θ 2 ) = 0 + [ 1 x 2 g ( θ 1 x ) + θ 1 x 3 g ( θ 1 x ) ] [ 1 G ( θ 2 x ) ] d x , b 2 ( θ 1 , θ 2 ) = 0 + θ 1 x 3 g ( θ 1 x ) g ( θ 2 x ) d x .
Proof. 
Applying the Delta method and Theorem 6, we can describe the asymptotic distribution of R ^ = R ( θ 1 ^ , θ 2 ^ ) as n r ( R ^ R ) N ( 0 , A ) , where
A = R ( θ 1 , θ 2 ) θ 1 R ( θ 1 , θ 2 ) θ 2 C 1 ( θ 1 , θ 2 ) R ( θ 1 , θ 2 ) θ 1 R ( θ 1 , θ 2 ) θ 2 .
It is easy to prove that b 1 ( θ 1 , θ 2 ) = R ( θ 1 , θ 2 ) θ 1 and b 2 ( θ 1 , θ 2 ) = R ( θ 1 , θ 2 ) θ 2 . Thus the proof is completed. ☐
From the results of Theorem 7, a 100 ( 1 ξ ) % asymptotic confidence interval can be determined as: [ R ^ A ^ n r z 1 ξ 2 , R ^ + A ^ n r z 1 ξ 2 ] , which is symmetric about R ^ . Here, A is estimated with θ 1 ^ and θ 2 ^ . z ξ is 100 ξ th percentile of N ( 0 , 1 ) . Similarly, by estimating matrix A with θ 1 ˜ and θ 2 ˜ , the R-symmetric asymptotic confidence interval based on R ˜ is thus acquired. Note that θ 1 ˜ and θ 2 ˜ is given in Section 3.2.

3.3.2. R-Asymmetric Interval Estimation Based on Bootstrap Methods

Since the obtained asymptotic distributions may not hold for small samples, two Bootstrap confidence intervals are thereby proposed, which are R-asymmetric and considered to have a better performance.
(a) Bootstrap-p method
Step 1:
Generate ( x r + 1 * , , x n * ) from ( x r + 1 , , x n ) and ( y r + 1 * , , y m * ) from ( y r + 1 , , y m ) , respectively. Based on the generated samples, calculate R * ^ with Equation (14);
Step 2:
Repeat step 1, N B O O T times;
Step 3:
Define the CDF of R * ^ as Φ 1 ( z ) = P ( R * ^ z ) . Let R ^ B P ( z ) = Φ 1 1 ( z ) for a given z. Thus an approximate 100 ( 1 ξ ) % confidence interval of R is ( R ^ B P ( ξ 2 ) , R ^ B P ( 1 ξ 2 ) ) .
(b) Bootstrap-t method
Step 1:
Calculate θ 1 ^ , θ 2 ^ and R ^ with ( x r + 1 , , x n ) and ( y r + 1 , , y m ) ;
Step 2:
Generate ( x r + 1 * , , x n * ) from ( x r + 1 , , x n ) and ( y r + 1 * , , y m * ) from ( y r + 1 , , y m ) , respectively;
Step 3:
Calculate R * ^ with Equation (14) and calculate B * with B * = n r ( R * ^ R ^ ) V a r ( R * ^ ) (Note that V a r ( R * ^ ) can be derived from Theorem 7);
Step 4:
Repeat steps 2 and 3, N B O O T times;
Step 5:
Define the CDF of B * as Φ 2 ( z ) = P ( B * z ) . Let R ^ B T ( z ) = R ^ + Φ 2 1 ( z ) V a r ( R ^ ) n r for a given z. Thus an approximate 100 ( 1 ξ ) % confidence interval of R is ( R ^ B T ( ξ 2 ) , R ^ B T ( 1 ξ 2 ) ) .

3.4. Bayesian Estimation

Since gamma distributions are frequently used as prior distributions for parameters from the inverted scale family according to the related articles [4,7,26,27,28], we assume independent gamma priors on θ 1 and θ 2 , that is, θ i G a m m a ( α i , β i ) for i { 1 , 2 } . The PDFs are given in Equation (24). Note that R can be derived in a similar way when θ follows other prior distributions.
π ( θ i | α i , β i ) = β i α i Γ ( α i ) θ i α i 1 e β i θ i , i = 1 , 2 , α i , β i > 0
Then, the joint posterior density function of θ 1 and θ 2 is:
L ( θ 1 , θ 2 | d a t a ) = L ( d a t a | θ 1 , θ 2 ) π ( θ 1 ) π ( θ 2 ) 0 0 L ( d a t a | θ 1 , θ 2 ) π ( θ 1 ) π ( θ 2 ) d θ 1 d θ 2 .
Equation (25) is found to be intractable, and therefore, getting the respective Bayesian estimator analytically is quite difficult. However, samples can be generated from the posterior distribution with Gibbs sampling so that the Bayesian estimator of R and HPD credible interval, proposed by Chen [29], can be obtained easily. Here give the posterior density functions of θ 1 and θ 2 :
f ( θ 1 | θ 2 , d a t a ) [ 1 G ( θ 1 x r + 1 ) ] r i = r + 1 n θ 1 x i 2 g ( θ 1 x i ) G ( θ 1 x i ) R j × θ 1 α 1 1 e β 1 θ 1 , f ( θ 2 | θ 1 , d a t a ) [ 1 G ( θ 2 y r + 1 ) ] r j = r + 1 m θ 2 y j 2 g ( θ 2 y j ) G ( θ 2 y j ) R j × θ 2 α 2 1 e β 2 θ 2 .
Both posterior density functions are found to be unknown. So the Metropolis-Hastings ( M H ) method is used to generate θ 1 ( k ) and θ 2 ( k ) from f ( θ 1 | θ 2 , d a t a ) and f ( θ 2 | θ 1 , d a t a ) , respectively. Gibbs sampling algorithm is as follows:
Step 1:
Determine the initial value ( θ 1 ( 0 ) , θ 2 ( 0 ) ) ;
Step 2:
Set k = 1 ;
Step 3:
Using the M H method, generate θ 1 ( k ) from f ( θ 1 | θ 2 , d a t a ) . Here, the proposal distribution is e x p ( θ 1 ( k 1 ) ) ;
Step 4:
Using the M H method, generate θ 2 ( k ) from f ( θ 2 | θ 1 , d a t a ) . Here, the proposal distribution is e x p ( θ 2 ( k 1 ) ) ;
Step 5:
Compute R ( k ) by Equation (14);
Step 6:
Set k = k + 1 ;
Step 7:
Repeat steps 3 to 6 K times.
Then the approximate posterior mean and posterior variance of R is:
E ^ ( R | d a t a ) = 1 K k = 1 K R ( k ) , V a r ^ ( R | d a t a ) = 1 K k = 1 K [ R ( k ) E ^ ( R | d a t a ) ] 2 .
The HPD credible interval at 100 ( 1 ξ ) % credible level can be given by:
( R [ ξ 2 K ] , R [ ( 1 ξ 2 ) K ] ) .
Here, R [ i ] denotes the i t h smallest of { R ( k ) ( k = 1 , 2 , , K ) }.

3.5. Simulation Study

Supposing both X and Y are from IED (as shown in Example 1), the obtained estimations for R can be achieved through Monte Carlo simulations. The GPTIICOS from X and Y are generated with the algorithm in Section 2.2.
Here three censoring schemes are considered, which are listed and numbered in Table 4, and six sets of parameter values ( θ 1 , θ 2 ) are used for simulations, including ( 1 , 1 ) , ( 1 , 2 ) , ( 1 , 3 ) , ( 2 , 2 ) , ( 2 , 3 ) , and ( 3 , 3 ) .
We make an evaluation on the validity of estimations obtained from each censoring scheme and each value of ( θ 1 , θ 2 ) . The evaluation process is as follows. First, the accuracy of the point estimates is compared in light of their mean squared errors (MSE) and biases. Then, with the average interval lengths and interval coverage percentages, we make an assessment on the different interval estimates.
The average point estimates for each parameter value and censoring scheme are obtained through 1000 simulations, which are shown in Table 5 together with the corresponding average biases and MSEs. Further, we give the heat maps of Table 5 in Figure 2 and Figure 3. The proposed intervals at 95 % confidence/credible level are also obtained. Table 6 gives the obtained average interval lengths and interval coverage percentages for different methods. In our simulations, the HPD credible intervals are acquired after 2000 samples while the Bootstrap confidence intervals are acquired after 250 resamples.
From Table 5, all the point estimates are quite effective as the corresponding MSEs and biases are pretty small. Additionally, we speculate from Figure 2 and Figure 3 that the accuracy of the estimators is somewhat related to the true value of ( θ 1 , θ 2 ) because the MSEs and biases of the Bayesian estimator show a great difference between the case of θ 1 = θ 2 and the others. We can find in the cases that ( θ 1 , θ 2 ) takes ( 1 , 3 ) and ( 2 , 3 ) , the MSEs and biases of the Bayesian estimator are much higher than those in the other cases. However, in the cases of θ 1 = θ 2 , Bayes estimates are extremely accurate.
For the interval estimations, we refer to Table 6. It can be seen that the interval length of the Bootstrap-t interval is wider than others for most cases and according to the interval coverage percentages, the performance of Bootstrap-t intervals is slightly better than the performance of asymptotic confidence intervals obtained from R ^ and R ˜ . Moreover, the interval coverage percentages of the R-symmetric asymptotic confidence intervals are more deviated from 95 % , which indicates that their performances are not good when compared with others. That is probably because small samples may affect the asymptotic distribution of R, which thereby may affect the accuracy of interval estimation. However, it is obvious that the HPD credible interval performs better when compared to the asymptotic intervals and Bootstrap intervals since the corresponding interval coverage percentages are closer to the nominal level. So, Bayes estimation is considered as the best method, and according to the interval coverage percentages, the ranking of the three methods is given below.
B E S T B a y e s B o o t s t r a p A s y m p t o t i c s W O R S T

3.6. An Illustrated Example

Here, an analysis of two real datasets, reported by Badar and Priest [30], is given below. Many reliability models, for example, the models based on the generalized Rayleigh distributions, Weibull distributions and three-parameter Weibull distributions, have been applied to the two datasets (see Raqab and Kundu [17], Kundu and Gupta [18], and Kundu and Raqad [19]).
Data Set 1.The data below shows the breaking strength for 63 carbon fibers of 10 mm gauge length in a tension test (unit: GPA).
1.901 2.132 2.203 2.228 2.257 2.350 2.361 2.396 2.397 2.445
2.454 2.474 2.518 2.522 2.525 2.532 2.575 2.614 2.616 2.618
2.624 2.659 2.675 2.738 2.740 2.856 2.917 2.928 2.937 2.937
2.977 2.996 3.030 3.125 3.139 3.145 3.220 3.223 3.235 3.243
3.264 3.272 3.294 3.332 3.346 3.377 3.408 3.435 3.493 3.501
3.537 3.554 3.562 3.628 3.852 3.871 3.886 3.971 4.024 4.027
4.225 4.395 5.020
Data Set 2.The data below shows the breaking strength for 69 carbon fibers of 20 mm gauge length in a tension test (unit: GPA).
1.312 1.314 1.479 1.552 1.700 1.803 1.861 1.865 1.944 1.958
1.966 1.997 2.006 2.021 2.027 2.055 2.063 2.098 2.140 2.179
2.224 2.240 2.253 2.270 2.272 2.274 2.301 2.301 2.359 2.382
2.382 2.426 2.434 2.435 2.478 2.490 2.511 2.514 2.535 2.554
2.566 2.570 2.586 2.629 2.633 2.642 2.648 2.684 2.697 2.726
2.770 2.773 2.800 2.809 2.818 2.821 2.848 2.880 2.954 3.012
3.067 3.084 3.090 3.096 3.128 3.233 3.433 3.585 3.585
Here, the fitting of three reliability models based on GIED (as shown in Example 1), WD (as shown in Section 2.3), IWD (as shown in Example 2) is considered. We use maximum likelihood estimation to construct the above three parametric models. Same as in Section 2.3, for each model, five measures are considered to test the goodness-of-fit: (1) −ln, (2) K S distance, (3) p-value, (4) A I C , and (5) B I C . In the case of MLE, these values based on the three reliability models are listed in Table 7. Clearly, the best fit order for both datasets is:
B E S T G I E D W D I W D W O R S T
Since GIED has not only the lowest values of −ln, A I C , and B I C but also the highest p-value, it is considered as the best model. Furthermore, the fitting effects of the three reliability models are intuitively shown in Figure 4. So GIED models are used to fit the two datasets separately. It appears from Table 7 that for Data Set 1, the values of −ln, K S distance, p-value, A I C , and B I C in the case of MLE are 57.7251 , 0.8601 , 0.7399 , 123.7367 , and 119.4504 , respectively, indicating the GIED model fits the Data Set 1 very well when α takes 175.2879 . Same as for Data Set 1, the GIED model fits Data Set 2 very well when α takes 205.87851 . Hence, it is reasonable to assume X follows GIED( θ 1 ,175.2879) and Y follows GIED( θ 2 ,205.87851). Set r = 3 . Table 8 shows the respective censoring schemes and the generated censored data.
With the R programming language, the proposed estimates of R are derived. As shown in Table 9, the MLE, AMLE, Bayesian estimator are 0.782, 0.794, and 0.782, respectively. The corresponding 95 % confidence/credible intervals are (0.680,0.885), (0.690,0.898), and (0.651,0.871). Additionally, The Bootstrap-p and Bootstrap-t confidence intervals are obtained to be (0.665,0.878) and (0.694,0.913), respectively.

4. Conclusions

Two estimation problems are solved in this paper. Under GPTIIC, we first achieve the exact interval estimation for ISF, and the Monte Carlo simulations indicate that it has a good performance. Then we derive the point and interval estimates of R in the case of ISF. The three point estimates are all accurate according to their biases and MSEs. However, according to the results of Bayes estimators, the accuracy of the estimators is thought to be related to the true values of the scale parameters. For interval estimates, the simulation results show that the HPD credible interval performs better since the corresponding interval coverage percentage is closer to the nominal level. But the R-symmetric asymptotic intervals based on R ^ and R ˜ are not good due to the small sample size. Moreover, the Bootstrap-t interval has the widest interval length in most cases. According to the interval percentages, the ranking of the three methods is: Bayes > Bootstrap > Asymptotics. For each estimation problem, we also demonstrate the estimation process using real datasets to show the feasibility of the proposed methods. Further work should focus on the two estimation problems for other distributions based on GPTIICS. In addition, under other censoring schemes other than GPTIIC, we can further consider the estimation problems for ISF.

Author Contributions

Methodology and Writing, J.G.; Writing, K.B.; Supervision, W.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Project 180170092 which was supported by National Training Program of Innovation and Entrepreneurship for Undergraduates.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sen, P.K. Progressive Censoring Schemes. In Encyclopedia of Statistical Sciences; Kotz, S., Johnson, N.L., Eds.; John Wiley & Sons: Hoboken, NJ, USA, 1986; Volume 7. [Google Scholar]
  2. Balakrishnan, N.; Aggarwala, R. Progressive Censoring: Theory, Methods, and Applications; Springer Science & Business Media: Berlin, Germany, 2000. [Google Scholar]
  3. Lin, C.T.; Duran, B.S.; Lewis, T.O. Inverted gamma as a life distribution. Microelectron. Reliab. 1989, 29, 619–626. [Google Scholar] [CrossRef]
  4. Kundu, D.; Howlader, H. Bayesian inference and prediction of the inverse Weibull distribution for Type-II censored data. Comput. Stat. Data Anal. 2010, 54, 1547–1558. [Google Scholar] [CrossRef]
  5. Prakash, G. Inverted Exponential distribution under a Bayesian viewpoint. J. Mod. Appl. Stat. Methods 2012, 11, 16. [Google Scholar] [CrossRef] [Green Version]
  6. Krishna, H.; Kumar, K. Reliability estimation in generalized inverted exponential distribution with progressively type II censored sample. J. Stat. Comput. Simul. 2013, 83, 1007–1019. [Google Scholar] [CrossRef]
  7. Dey, S.; Singh, S.; Tripathi, Y.M.; Asgharzadeh, A. Estimation and prediction for a progressively censored generalized inverted exponential distribution. Stat. Methodol. 2016, 32, 185–202. [Google Scholar] [CrossRef]
  8. Ahmed, E.A. Estimation and prediction for the generalized inverted exponential distribution based on progressively first-failure-censored data with application. J. Appl. Stat. 2017, 44, 1576–1608. [Google Scholar] [CrossRef]
  9. Viveros, R.; Balakrishnan, N. Interval estimation of parameters of life from progressively censored data. Technometrics 1994, 36, 84–91. [Google Scholar] [CrossRef]
  10. Balakrishnan, N.; Kannan, N.; Lin, C.T.; Ng, H.T. Point and interval estimation for Gaussian distribution, based on progressively Type-II censored samples. IEEE Trans. Reliab. 2003, 52, 90–95. [Google Scholar] [CrossRef]
  11. Wang, B.X. Exact interval estimation for the scale family under general progressive Type-II censoring. Commun. Stat. Theory Methods 2012, 41, 4444–4452. [Google Scholar] [CrossRef]
  12. Church, J.D.; Harris, B. The estimation of reliability from stress-strength relationships. Technometrics 1970, 12, 49–54. [Google Scholar] [CrossRef]
  13. Downton, F. The estimation of Pr (Y < X) in the normal case. Technometrics 1973, 15, 551–558. [Google Scholar]
  14. Govindarazulu, Z. Two-Sided confidence limits for P (X < Y) based on normal samples of X and Y. Sankhyā Indian J. Stat. Ser. B 1967, 29, 35–40. [Google Scholar]
  15. Woodward, W.A.; Kelley, G.D. Minimum variance unbiased estimation of P [Y < X] in the normal case. Technometrics 1977, 19, 95–98. [Google Scholar]
  16. Owen, D.B.; Craswell, K.J.; Hanson, D.L. Nonparametric Upper Confidence Bounds for Pr Y; Sandia Corporation: Albuquerque, NM, USA, 1964. [Google Scholar]
  17. Kundu, D.; Gupta, R.D. Estimation of P(Y < X) for the generalized exponential distribution. Metrika 2005, 61, 291–308. [Google Scholar]
  18. Kundu, D.; Gupta, R.D. Estimation of P(Y < X) for Weibull distribution. IEEE Trans. Reliab. 2006, 55, 270–280. [Google Scholar]
  19. Kundu, D.; Raqab, M.Z. Estimation of R = P (Y < X) for three-parameter Weibull distribution. Stat. Probab. Lett. 2009, 79, 1839–1846. [Google Scholar]
  20. Aggarwala, R.; Balakrishnan, N. Some properties of progressive censored order statistics from arbitrary and uniform distributions with applications to inference and simulation. J. Stat. Plan. Inference 1998, 70, 35–49. [Google Scholar] [CrossRef]
  21. Gross, H.P.; DeClark, R.P. Case report of radiation necrosis of the hip. S. D. J. Med. 1975, 28, 5–6. [Google Scholar] [PubMed]
  22. Akaike, H. A new look at the statistical model identification. IEEE Trans. Autom. Control 1974, 19, 716–723. [Google Scholar] [CrossRef]
  23. Schwarz, G. Estimating the dimension of a model. Ann. Stat. 1978, 6, 461–464. [Google Scholar] [CrossRef]
  24. Bagnoli, M.; Bergstrom, T. Log-Concave Probability and Its Applications; Springer: Berlin, Germany, 2005. [Google Scholar]
  25. Balakrishnan, N.; Mi, J. Existence and uniqueness of the mles for normal distribution based on general progressively type-ii censored samples. Stat. Probab. Lett. 2003, 64, 407–414. [Google Scholar] [CrossRef]
  26. Singh, S.; Tripathi, Y.M. Estimating the parameters of an inverse Weibull distribution under progressive type-I interval censoring. Stat. Pap. 2018, 59, 21–56. [Google Scholar] [CrossRef]
  27. Ali, S. Mixture of the inverse Rayleigh distribution: Properties and estimation in a Bayesian framework. Appl. Math. Model. 2015, 39, 515–530. [Google Scholar] [CrossRef]
  28. Dey, S.; Dey, T. Generalized inverted exponential distribution: Different methods of estimation. Am. J. Math. Manag. Sci. 2014, 33, 194–215. [Google Scholar] [CrossRef]
  29. Chen, M.H.; Shao, Q.M. Monte Carlo estimation of Bayesian credible and HPD intervals. J. Comput. Graph. Stat. 1999, 8, 69–92. [Google Scholar]
  30. Bader, M.G.; Priest, A.M. Statistical aspects of fibre and bundle strength in hybrid composites. Prog. Sci. Eng. Compos. 1982, 1129–1136. [Google Scholar] [CrossRef] [Green Version]
Figure 1. An empirical distribution function (CDF) and other CDFs based on different statistical models.
Figure 1. An empirical distribution function (CDF) and other CDFs based on different statistical models.
Symmetry 12 00731 g001
Figure 2. The bias heat map.
Figure 2. The bias heat map.
Symmetry 12 00731 g002
Figure 3. The MSE heat map.
Figure 3. The MSE heat map.
Symmetry 12 00731 g003
Figure 4. An empirical CDF and other CDFs based on different statistical models for the datasets.
Figure 4. An empirical CDF and other CDFs based on different statistical models for the datasets.
Symmetry 12 00731 g004
Table 1. Average confidence intervals (CI), interval lengths (IL), and interval coverage percentages (CP).
Table 1. Average confidence intervals (CI), interval lengths (IL), and interval coverage percentages (CP).
1 ξ = 0.90 1 ξ = 0.95
( N , r , n ) CIILCPCIILCP
( 20 , 1 , 10 ) ( 2.97 , 13.18 ) 10.21 0.904 ( 2.63 , 15.62 ) 12.99 0.953
( 20 , 3 , 12 ) ( 2.87 , 13.85 ) 10.98 0.900 ( 2.52 , 16.39 ) 13.87 0.951
( 20 , 1 , 16 ) ( 2.93 , 10.84 ) 7.91 0.901 ( 2.60 , 12.33 ) 9.73 0.951
( 20 , 3 , 18 ) ( 2.76 , 11.97 ) 9.21 0.902 ( 2.42 , 13.77 ) 11.35 0.948
( 30 , 1 , 10 ) ( 3.37 , 8.40 ) 5.03 0.897 ( 3.08 , 9.12 ) 6.04 0.952
( 30 , 3 , 12 ) ( 3.27 , 9.00 ) 5.73 0.896 ( 2.99 , 9.91 ) 6.92 0.952
( 30 , 1 , 16 ) ( 3.28 , 9.13 ) 5.85 0.897 ( 2.99 , 10.21 ) 7.22 0.947
( 30 , 3 , 18 ) ( 3.13 , 10.46 ) 7.33 0.895 ( 2.80 , 11.78 ) 8.98 0.951
( 40 , 1 , 10 ) ( 3.53 , 7.83 ) 4.30 0.900 ( 3.39 , 8.48 ) 5.09 0.949
( 40 , 3 , 12 ) ( 3.46 , 8.39 ) 4.93 0.903 ( 3.19 , 9.17 ) 5.98 0.950
( 40 , 1 , 16 ) ( 3.54 , 7.71 ) 4.17 0.900 ( 3.29 , 8.32 ) 5.03 0.950
( 40 , 3 , 18 ) ( 3.41 , 8.57 ) 5.16 0.897 ( 3.14 , 9.37 ) 6.23 0.950
Table 2. Fitting results for the three reliability models.
Table 2. Fitting results for the three reliability models.
Model α θ −ln K S p-Value AIC BIC
IWD 3.777054 1.58526 15.19896 0.1306 0.9187 34.39793 36.17867
GIED 19.519704 6.189745 16.22355 0.14001 0.8721 36.4471 38.22784
WD 2.785545 8.779702 19.06114 0.17819 0.6171 42.12228 43.90302
Table 3. Generated dataset and censoring scheme.
Table 3. Generated dataset and censoring scheme.
i , j 12345678910111213141516
X i 1.3 1.4 1.6 1.7 1.7 1.8 1.8 1.9 2.0 2.2 2.3 2.7 3.0 4.1
R i 11110000000000
Table 4. Censoring schemes ( C . S . ).
Table 4. Censoring schemes ( C . S . ).
C . S . ( r , n , N , R r + 1 , R r + 2 , , R n )
R 1 ( 2 , 10 , 29 , 3 , 1 , 2 , 3 , 1 , 2 , 3 , 4 )
R 2 ( 2 , 10 , 26 , 2 , 2 , 2 , 2 , 2 , 2 , 2 , 2 )
R 3 ( 2 , 10 , 26 , 4 , 2 , 4 , 2 , 1 , 1 , 1 , 1 )
Table 5. Average point estimates of R and the corresponding Biases (MSEs).
Table 5. Average point estimates of R and the corresponding Biases (MSEs).
MLE AMLE Bayes
( θ 1 , θ 2 ) C . S . R Bias ( Mse ) R Bias ( Mse ) R Bias ( Mse )
( 1 , 1 ) ( R 1 , R 1 ) 0.50085 0.00085 ( 0.00513 ) 0.50291 0.00291 ( 0.00481 ) 0.50016 0.00016 ( 0.00018 )
( R 1 , R 2 ) 0.50468 0.01468 ( 0.00479 ) 0.50699 0.00699 ( 0.00576 ) 0.50049 0.00049 ( 0.00017 )
( R 1 , R 3 ) 0.51370 0.01370 ( 0.00555 ) 0.52850 0.02850 ( 0.00649 ) 0.50050 0.00050 ( 0.00018 )
( R 2 , R 2 ) 0.50230 0.00230 ( 0.00515 ) 0.50231 0.00231 ( 0.00523 ) 0.50009 0.00009 ( 0.00016 )
( R 2 , R 3 ) 0.50833 0.00833 ( 0.00550 ) 0.51731 0.01731 ( 0.00634 ) 0.50078 0.00078 ( 0.00016 )
( R 3 , R 3 ) 0.50295 0.00295 ( 0.00592 ) 0.50262 0.00262 ( 0.00541 ) 0.50066 0.00066 ( 0.00017 )
( 1 , 2 ) ( R 1 , R 1 ) 0.33615 0.00315 ( 0.00425 ) 0.34790 0.01490 ( 0.00603 ) 0.33341 0.00041 ( 0.00007 )
( R 1 , R 2 ) 0.34654 0.01354 ( 0.00402 ) 0.34425 0.01125 ( 0.00434 ) 0.35512 0.02212 ( 0.00327 )
( R 1 , R 3 ) 0.34605 0.01305 ( 0.00473 ) 0.34831 0.01531 ( 0.00473 ) 0.35051 0.01751 ( 0.00289 )
( R 2 , R 2 ) 0.33812 0.00512 ( 0.00436 ) 0.33780 0.00480 ( 0.00494 ) 0.35882 0.02582 ( 0.00370 )
( R 2 , R 3 ) 0.33332 0.00032 ( 0.00464 ) 0.34141 0.00841 ( 0.00485 ) 0.35614 0.02314 ( 0.00352 )
( R 3 , R 3 ) 0.33670 0.00370 ( 0.00459 ) 0.33432 0.00132 ( 0.00454 ) 0.36062 0.02762 ( 0.00377 )
( 1 , 3 ) ( R 1 , R 1 ) 0.25153 0.00153 ( 0.00292 ) 0.25651 0.00651 ( 0.00308 ) 0.32660 0.07660 ( 0.01191 )
( R 1 , R 2 ) 0.25812 0.00812 ( 0.00290 ) 0.25621 0.00621 ( 0.00291 ) 0.32476 0.07476 ( 0.01080 )
( R 1 , R 3 ) 0.25847 0.00847 ( 0.00371 ) 0.27132 0.02132 ( 0.00353 ) 0.32228 0.07228 ( 0.01192 )
( R 2 , R 2 ) 0.25594 0.00594 ( 0.00354 ) 0.25403 0.00403 ( 0.00362 ) 0.32491 0.07491 ( 0.01116 )
( R 2 , R 3 ) 0.25292 0.00292 ( 0.00337 ) 0.25022 0.00022 ( 0.00341 ) 0.32323 0.07323 ( 0.01163 )
( R 3 , R 3 ) 0.25245 0.00245 ( 0.00301 ) 0.25808 0.00808 ( 0.00346 ) 0.32516 0.07516 ( 0.01053 )
( 2 , 2 ) ( R 1 , R 1 ) 0.50245 0.00245 ( 0.00495 ) 0.50019 0.00019 ( 0.00520 ) 0.50005 0.00005 ( 0.00007 )
( R 1 , R 2 ) 0.50341 0.00341 ( 0.00550 ) 0.51469 0.01469 ( 0.00575 ) 0.50044 0.00044 ( 0.00007 )
( R 1 , R 3 ) 0.50896 0.00896 ( 0.00552 ) 0.52607 0.02607 ( 0.00611 ) 0.50027 0.00027 ( 0.00006 )
( R 2 , R 2 ) 0.50033 0.00033 ( 0.00565 ) 0.50055 0.00055 ( 0.00542 ) 0.50013 0.00013 ( 0.00007 )
( R 2 , R 3 ) 0.50055 0.00055 ( 0.00618 ) 0.51098 0.01098 ( 0.00578 ) 0.50010 0.00010 ( 0.00008 )
( R 3 , R 3 ) 0.50365 0.00365 ( 0.00554 ) 0.50225 0.00255 ( 0.00596 ) 0.50079 0.00079 ( 0.00007 )
( 2 , 3 ) ( R 1 , R 1 ) 0.40467 0.00467 ( 0.00464 ) 0.40345 0.00345 ( 0.00465 ) 0.49350 0.09350 ( 0.00879 )
( R 1 , R 2 ) 0.41019 0.01019 ( 0.00485 ) 0.40838 0.00838 ( 0.00458 ) 0.49376 0.09376 ( 0.00885 )
( R 1 , R 3 ) 0.40929 0.00929 ( 0.00445 ) 0.42179 0.02179 ( 0.00531 ) 0.49361 0.09361 ( 0.00882 )
( R 2 , R 2 ) 0.40152 0.00152 ( 0.00449 ) 0.40775 0.00775 ( 0.00604 ) 0.49382 0.09382 ( 0.00885 )
( R 2 , R 3 ) 0.40044 0.00044 ( 0.00530 ) 0.41298 0.01298 ( 0.00560 ) 0.49362 0.09362 ( 0.00882 )
( R 3 , R 3 ) 0.40132 0.00132 ( 0.00526 ) 0.40099 0.00099 ( 0.00526 ) 0.49382 0.09382 ( 0.00886 )
( 3 , 3 ) ( R 1 , R 1 ) 0.50172 0.00172 ( 0.00522 ) 0.50209 0.00209 ( 0.00495 ) 0.50009 0.00009 ( 0.00005 )
( R 1 , R 2 ) 0.50577 0.00577 ( 0.00519 ) 0.51180 0.01180 ( 0.00581 ) 0.50010 0.00010 ( 0.00005 )
( R 1 , R 3 ) 0.50220 0.00220 ( 0.00598 ) 0.52334 0.02334 ( 0.00622 ) 0.50086 0.00086 ( 0.00005 )
( R 2 , R 2 ) 0.50062 0.00062 ( 0.00550 ) 0.50583 0.00583 ( 0.00590 ) 0.50003 0.00003 ( 0.00004 )
( R 2 , R 3 ) 0.50077 0.00077 ( 0.00545 ) 0.50499 0.00499 ( 0.00601 ) 0.50043 0.00043 ( 0.00005 )
( R 3 , R 3 ) 0.50363 0.00363 ( 0.00602 ) 0.50470 0.00470 ( 0.00629 ) 0.50020 0.00020 ( 0.00004 )
Table 6. Average interval length (interval coverage percentage) for different interval estimations.
Table 6. Average interval length (interval coverage percentage) for different interval estimations.
( θ 1 , θ 2 ) C . S . MLE AMLE Bayes Boot p Boot t
( 1 , 1 ) ( R 1 , R 1 ) 0.268 ( 0.943 ) 0.272 ( 0.925 ) 0.250 ( 0.954 ) 0.268 ( 0.922 ) 0.299 ( 0.948 )
( R 1 , R 2 ) 0.272 ( 0.899 ) 0.273 ( 0.911 ) 0.250 ( 0.938 ) 0.275 ( 0.924 ) 0.302 ( 0.943 )
( R 1 , R 3 ) 0.278 ( 0.914 ) 0.276 ( 0.875 ) 0.250 ( 0.930 ) 0.276 ( 0.931 ) 0.305 ( 0.954 )
( R 2 , R 2 ) 0.282 ( 0.912 ) 0.284 ( 0.924 ) 0.250 ( 0.947 ) 0.282 ( 0.914 ) 0.305 ( 0.926 )
( R 2 , R 3 ) 0.286 ( 0.919 ) 0.280 ( 0.907 ) 0.247 ( 0.946 ) 0.287 ( 0.930 ) 0.309 ( 0.935 )
( R 3 , R 3 ) 0.284 ( 0.935 ) 0.284 ( 0.925 ) 0.247 ( 0.955 ) 0.311 ( 0.954 ) 0.311 ( 0.946 )
( 1 , 2 ) ( R 1 , R 1 ) 0.237 ( 0.914 ) 0.239 ( 0.929 ) 0.249 ( 0.935 ) 0.241 ( 0.929 ) 0.266 ( 0.915 )
( R 1 , R 2 ) 0.248 ( 0.927 ) 0.245 ( 0.911 ) 0.249 ( 0.953 ) 0.253 ( 0.937 ) 0.270 ( 0.929 )
( R 1 , R 3 ) 0.255 ( 0.912 ) 0.247 ( 0.909 ) 0.249 ( 0.943 ) 0.258 ( 0.928 ) 0.272 ( 0.941 )
( R 2 , R 2 ) 0.250 ( 0.905 ) 0.251 ( 0.914 ) 0.250 ( 0.931 ) 0.255 ( 0.943 ) 0.276 ( 0.920 )
( R 2 , R 3 ) 0.256 ( 0.922 ) 0.253 ( 0.910 ) 0.249 ( 0.947 ) 0.261 ( 0.944 ) 0.277 ( 0.930 )
( R 3 , R 3 ) 0.257 ( 0.933 ) 0.257 ( 0.906 ) 0.250 ( 0.949 ) 0.261 ( 0.946 ) 0.279 ( 0.931 )
( 1 , 3 ) ( R 1 , R 1 ) 0.205 ( 0.924 ) 0.205 ( 0.934 ) 0.249 ( 0.930 ) 0.210 ( 0.939 ) 0.228 ( 0.928 )
( R 1 , R 2 ) 0.215 ( 0.910 ) 0.208 ( 0.913 ) 0.250 ( 0.951 ) 0.219 ( 0.936 ) 0.231 ( 0.919 )
( R 1 , R 3 ) 0.221 ( 0.924 ) 0.211 ( 0.896 ) 0.249 ( 0.947 ) 0.226 ( 0.917 ) 0.232 ( 0.902 )
( R 2 , R 2 ) 0.214 ( 0.914 ) 0.214 ( 0.919 ) 0.249 ( 0.940 ) 0.221 ( 0.938 ) 0.236 ( 0.926 )
( R 2 , R 3 ) 0.221 ( 0.909 ) 0.217 ( 0.905 ) 0.249 ( 0.943 ) 0.226 ( 0.937 ) 0.236 ( 0.910 )
( R 3 , R 3 ) 0.217 ( 0.931 ) 0.217 ( 0.909 ) 0.250 ( 0.957 ) 0.223 ( 0.957 ) 0.236 ( 0.927 )
( 2 , 2 ) ( R 1 , R 1 ) 0.269 ( 0.920 ) 0.265 ( 0.925 ) 0.250 ( 0.937 ) 0.268 ( 0.949 ) 0.298 ( 0.924 )
( R 1 , R 2 ) 0.274 ( 0.912 ) 0.275 ( 0.921 ) 0.250 ( 0.936 ) 0.276 ( 0.932 ) 0.303 ( 0.924 )
( R 1 , R 3 ) 0.274 ( 0.900 ) 0.276 ( 0.911 ) 0.249 ( 0.946 ) 0.277 ( 0.926 ) 0.306 ( 0.859 )
( R 2 , R 2 ) 0.281 ( 0.911 ) 0.279 ( 0.911 ) 0.249 ( 0.962 ) 0.282 ( 0.942 ) 0.307 ( 0.920 )
( R 2 , R 3 ) 0.281 ( 0.907 ) 0.281 ( 0.908 ) 0.250 ( 0.955 ) 0.283 ( 0.933 ) 0.308 ( 0.934 )
( R 3 , R 3 ) 0.286 ( 0.930 ) 0.284 ( 0.914 ) 0.251 ( 0.945 ) 0.287 ( 0.949 ) 0.309 ( 0.931 )
( 2 , 3 ) ( R 1 , R 1 ) 0.257 ( 0.909 ) 0.256 ( 0.900 ) 0.250 ( 0.943 ) 0.260 ( 0.925 ) 0.285 ( 0.911 )
( R 1 , R 2 ) 0.267 ( 0.926 ) 0.262 ( 0.911 ) 0.249 ( 0.956 ) 0.269 ( 0.952 ) 0.293 ( 0.934 )
( R 1 , R 3 ) 0.270 ( 0.912 ) 0.264 ( 0.896 ) 0.250 ( 0.934 ) 0.272 ( 0.929 ) 0.293 ( 0.922 )
( R 2 , R 2 ) 0.267 ( 0.907 ) 0.270 ( 0.916 ) 0.250 ( 0.947 ) 0.272 ( 0.939 ) 0.296 ( 0.918 )
( R 2 , R 3 ) 0.274 ( 0.921 ) 0.271 ( 0.910 ) 0.250 ( 0.940 ) 0.275 ( 0.943 ) 0.297 ( 0.929 )
( R 3 , R 3 ) 0.272 ( 0.912 ) 0.271 ( 0.907 ) 0.249 ( 0.952 ) 0.275 ( 0.944 ) 0.296 ( 0.933 )
( 3 , 3 ) ( R 1 , R 1 ) 0.267 ( 0.908 ) 0.267 ( 0.923 ) 0.249 ( 0.959 ) 0.269 ( 0.939 ) 0.296 ( 0.923 )
( R 1 , R 2 ) 0.274 ( 0.914 ) 0.275 ( 0.895 ) 0.249 ( 0.947 ) 0.275 ( 0.935 ) 0.302 ( 0.939 )
( R 1 , R 3 ) 0.276 ( 0.888 ) 0.276 ( 0.899 ) 0.250 ( 0.934 ) 0.276 ( 0.913 ) 0.305 ( 0.921 )
( R 2 , R 2 ) 0.281 ( 0.915 ) 0.282 ( 0.927 ) 0.250 ( 0.929 ) 0.283 ( 0.934 ) 0.308 ( 0.925 )
( R 2 , R 3 ) 0.283 ( 0.917 ) 0.280 ( 0.908 ) 0.250 ( 0.925 ) 0.284 ( 0.950 ) 0.308 ( 0.909 )
( R 3 , R 3 ) 0.285 ( 0.907 ) 0.283 ( 0.911 ) 0.251 ( 0.945 ) 0.285 ( 0.939 ) 0.307 ( 0.911 )
Table 7. Fitting results for the three reliability models.
Table 7. Fitting results for the three reliability models.
Data SetModel α θ −ln K S p-Value AIC BIC
1GIED 175.2879 16.8110 57.7251 0.8601 0.7399 119.4504 123.7367
WD 5.0494 424.5597 61.9570 0.0876 0.7192 127.914 132.2002
IWD 5.433788 2.721432 58.9022 0.1001 0.5528 121.8043 126.0906
2GIED 205.87851 13.88254 49.16796 0.041419 0.9998 102.3359 106.8041
WD 5.504847 214.130524 49.59614 0.056132 0.9816 103.1923 107.6605
IWD 4.126731 2.143723 63.62361 0.13363 0.1700 131.2472 135.7154
Table 8. Generated datasets and censoring schemes.
Table 8. Generated datasets and censoring schemes.
i , j 123456789101112131415
X i --- 2.228 2.397 2.522 2.616 2.738 2.937 3.125 3.235 3.332 3.493 3.628 4.024
R i ---444444444444
Y j --- 1.552 1.944 2.021 2.140 2.270 3.359 2.435 2.535 2.629 2.697 2.807 2.954
R j ---4444444444410
Table 9. Estimation results for the proposed estimators and confidence/credible intervals.
Table 9. Estimation results for the proposed estimators and confidence/credible intervals.
EstimationMLEAMLEBayesBootstrap-pBootstrap-t
Point Estimation 0.782 0.794 0.782
Interval Estimation ( 0.680 , 0.885 ) ( 0.690 , 0.898 ) ( 0.651 , 0.871 ) ( 0.665 , 0.878 ) ( 0.694 , 0.913 )

Share and Cite

MDPI and ACS Style

Gao, J.; Bai, K.; Gui, W. Statistical Inference for the Inverted Scale Family under General Progressive Type-II Censoring. Symmetry 2020, 12, 731. https://doi.org/10.3390/sym12050731

AMA Style

Gao J, Bai K, Gui W. Statistical Inference for the Inverted Scale Family under General Progressive Type-II Censoring. Symmetry. 2020; 12(5):731. https://doi.org/10.3390/sym12050731

Chicago/Turabian Style

Gao, Jing, Kehan Bai, and Wenhao Gui. 2020. "Statistical Inference for the Inverted Scale Family under General Progressive Type-II Censoring" Symmetry 12, no. 5: 731. https://doi.org/10.3390/sym12050731

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop