Next Article in Journal
IG-LSPIA: Least Squares Progressive Iterative Approximation for Isogeometric Collocation Method
Next Article in Special Issue
New Approaches on Parameter Estimation of the Gamma Distribution
Previous Article in Journal
Exponential Stability of Nonlinear Time-Varying Delay Differential Equations via Lyapunov–Razumikhin Technique
Previous Article in Special Issue
Design Efficiency of the Asymmetric Minimum Projection Uniform Designs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New One-Parameter Distribution for Right Censored Bayesian and Non-Bayesian Distributional Validation under Various Estimation Methods

1
Department of Statistics and Operations Research, Faculty of Science, King Saud University, P.O. Box 2455, Riyadh 11451, Saudi Arabia
2
Department of Mathematics, Laboratory of Probability and Statistics LaPS, Badji Mokhtar Annaba University, Annaba 23000, Algeria
3
Department of Mathematical Sciences, Ball State University, Muncie, IN 47306, USA
4
Department of Statistics, Mathematics and Insurance, Faculty of Commerce, Benha University, Benha 13518, Egypt
5
Department of Applied, Mathematical and Actuarial Statistics, Faculty of Commerce, Damietta University, Damietta 34517, Egypt
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(4), 897; https://doi.org/10.3390/math11040897
Submission received: 2 January 2023 / Revised: 1 February 2023 / Accepted: 3 February 2023 / Published: 10 February 2023
(This article belongs to the Special Issue Distribution Theory and Application)

Abstract

:
We propose a new extension of the exponential distribution for right censored Bayesian and non-Bayesian distributional validation. The parameter of the new distribution is estimated using several conventional methods, including the Bayesian method. The likelihood estimates and the Bayesian estimates are compared using Pitman’s closeness criteria. The Bayesian estimators are derived using three loss functions: the extended quadratic, the Linex, and the entropy functions. Through simulated experiments, all the estimating approaches offered have been assessed. The censored maximum likelihood method and the Bayesian approach are compared using the BB algorithm. The development of the Nikulin–Rao–Robson statistic for the new model in the uncensored situation is thoroughly discussed with the aid of two applications and a simulation exercise. For the novel model under the censored condition, two applications and the derivation of the Bagdonavičius and Nikulin statistic are also described.

1. Introduction

Exponential distribution is one of the often-encountered continuous probability distributions. It is frequently used to simulate the time between events. The mean and expected value of the exponential probability distribution, as well as its theoretical interpretation, will now be given. The exponential probability distribution, or Poisson point process, is the probability distribution of the interval between events in a process where events happen continuously and independently at a set average rate. In statistics and probability theory, this probability distribution is exploited. It is a unique instance of the gamma distribution. It is the continuous counterpart of the geometric distribution and possesses the key quality of being memoryless. In addition to the analysis of Poisson point processes, it is used in numerous other contexts. Exponential distribution is not the same as the class of exponential families of distributions, a major class of probability distributions that also includes the normal distribution, binomial distribution, gamma distribution, Poisson distribution, and many more probability distributions. One of the most often utilized continuous distributions is exponential probability distribution. It helps in figuring out how long there will be between events. It is used in many disciplines, such as physics, dependability theory, queuing theory, and others. Finding the height of different molecules in a gas at a stable temperature and pressure in a uniform gravitational field, as well as computing the monthly and annual highest values of consistent rainfall and river outflow volumes, are all tasks that can be accomplished using the exponential probability distribution. In statistical inference and dependability, selecting a suitable core model for new data analysis is becoming increasingly crucial. The results could be significantly affected by even a little departure from the core model. The difficulty of this work is increased by censorship. The chi-square test type is the one used most frequently to determine the goodness-of-fit. Numerous modifications to chi-square tests have been proposed by various researchers.
One of the aims of this study is to present a goodness-of-fit test for our parametric model, which is often used in survival analysis, social sciences, engineering, and dependability, in complete data scenarios, and in the presence of right censoring. For the one parameter Poisson-exponential (OPPE) model, we present the explicit forms of the quadratic test statistics (Nikulin–Rao–Robson test, Bagdonavičius and Nikulin test statistic, see Bagdonavičius et al. [1], Bagdonavičius and Nikulin [2], Bagdonavičius and Nikulin [3], Nikulin [4], Nikulin [5], Nikulin [6], and Rao and Robson [7]). Then, we use actual data to apply both tests. This study demonstrates how the Bayesian technique may enhance both the mean time between failure of the OPPE distribution and the maximum likelihood estimate of the parameter. By using a suitable loss function, we will also illustrate the utility of this adjustment. The OPPE distribution is distinguished by its ease of mathematical and scientific handling in statistical and mathematical modeling. As is common for many probability distribution researchers, we do not approach this new distribution in the normal method in this work. For instance, we do not focus as much on the conventional study of the new distribution, not because it is unimportant, but rather because we are more concerned with the practical applications of the mathematical and statistical modeling, as well as a significant portion of the distribution’s verification using censored data. To demonstrate the significance and flexibility of the new distribution and its wide range of applications in statistical and mathematical modeling, as well as the handling of controlled data, we omit several theoretical mathematical features, many algebraic derivations, and related theories. To be helpful to scholars in the field and assist them in providing additional similar and possibly more flexible distributions, we must briefly discuss the emergence of the new distribution in this context as well as how it was created and formed. The OPPE distribution’s cumulative distribution function (CDF) can be written as
Pr ( X δ x ) = F δ ( x ) = 1 1 e x p ( 1 ) { 1 e x p [ Δ ¯ δ ( x ) ] } ,
where x > 0 and δ > 0, Δ ¯ δ ( x ) = 1 Δ δ ( x ) and
Δ δ ( x ) = [ e x p ( δ x ) ] [ 1 e x p ( δ x ) ] .
The corresponding probability density function (PDF) can then be expressed as
f δ ( x ) = δ 1 e x p ( 1 ) Δ δ ( x ) e x p { [ x + Δ ¯ δ ( x ) ] } [ δ x + exp ( δ x ) 1 ] ,
for all x > 0 and δ > 0 . Many scholars have been interested in the exponential distribution and in providing new flexible extensions of it, and they have also been interested in the applications of these new extensions in various fields of science, such as engineering, insurance, medicine, reliability, actuarial science, and others. In this work, we propose to develop a modified chi-square type fit test for the OPPE model in the scenario when the parameter is unknown, and the data are complete. This test is based on the Nikulin–Rao–Robson (NKRR) statistic, which Nikulin [3], Nikulin [4], Nikulin [5], Nikulin [6], and Rao and Robson [7] separately proposed. This Y n 2 statistic is a logical refinement of the conventional chi-square test based on the maximum likelihood estimator on the original data. The original data were used in full for this statistic, which is a straightforward enough statistic to estimate the null hypothesis’s parameter. Second, this statistic entirely employed the Fisher data, which was to assess the information relative to a parameter contained in a model selection, based on the MLE. This provides a more unbiased evaluation of the chosen model of fit. Next, in the scenario when the parameters are unknown and the data are right-censored, we create a new goodness-of-fit test for this model. This modified chi-square Y n 2 test adjusts the NKRR statistic to account for both censoring and the unknown parameter. This study is a very significant numerical simulation study that we are conducting to demonstrate the invariance of the test statistic distribution on the one hand, and to test the null hypothesis H 0 that a sample originates from an OPPE model on the other. We compute Y n 2 the NKRR statistic as well as the Bagdonavičius and Nikulin statistics of various simulated samples, respectively. As a result, we can state that the suggested tests can, respectively, fit complete and censored data from the OPPE distribution. We conclude our contribution in this work with applications using real data, with the aim of demonstrating the applicability of the OPPE model across many scientific disciplines.
It is worth noting that the NKRR basic test, which is supported by complete data, is the most popular test in the last ten years. This is because it fits the complete truth data, and this is the case in most practical and applied cases in various fields. What distinguishes this statistical test also is the availability of ready-made statistical packages on the R program, whether for simulation or applications on actual data. However, the practical and experimental reality in many fields (such as the medical, chemical, engineering, etc.) necessitates that researchers deal with practical experiments that produce controlled censored data. This type of data, of course, needs certain tests dedicated to statistical dealing with it in the problem of distributional validation. The NKRR basic test is not the optimal choice in these cases. Hence, and based on this dilemma, was the primary and most important motive that prompted many researchers to think about introducing a new statistical test that fits the censored data. This new NKRR test, of course, is a modified test from the NKRR original test. According to the nature of the procedures of the two tests, both tests are not suitable for working with a type of data. One weakness of the original test, for instance, is that it is only applicable to entire data and cannot be used to deal with censored data. Therefore, this test may be a strong contender for doing statistical hypothesis tests if we are working with complete data. The modified test also has the drawback of only being appropriate for use with censored data; it is not designed for complete data and can only be used for censored data. Because it is designed for this type of data, this test will surely be a strong contender for statistical hypothesis tests if we are working with censored data.

2. Construction of NKRR Statistic for the OPPE Model

The NKRR statistic is a well-known variant of the traditional chi-squared tests in the situation of complete data (for more information, see Nikulin [3], Nikulin [4], Nikulin [5], Nikulin [6], and Rao and Robson [7]). The most common test to check whether a mathematical model is suitable for the data from observations is the chi-square statistic of Pearson. When the model’s parameters are unknown or data are censored, these tests, however, cannot be used. Natural adaptations of the Pearson statistic for the entire set of data were reported by Nikulin [3], Nikulin [4], Nikulin [5], Nikulin [6], and Rao and Robson [7] and are known as NKRRs. The chi-square distribution is used in this statistical test, which is a logical extension of the Pearson statistic. When the censoring is included in addition to the unknown parameter, the classical test is insufficient to support the null hypothesis. Bagdonavičius et al. [1], Bagdonavičius and Nikulin [2], Bagdonavičius and Nikulin [3], Nikulin [4], Nikulin [5], Nikulin [6], and Rao and Robson [7] suggested changing the NKRR statistic to take into account random right censoring. For the OPPE model in the current study, we recommend creating a modified chi-square test. The following NKRR statistic Y 2 was developed by Nikulin [3], Nikulin [4], Nikulin [5], Nikulin [6], and Rao and Robson [7] to test the hypothesis.
H 0     :     Pr { X δ x } = F δ ( x ) | x R ,
where in our case F δ ( x ) = F δ ( x ) | x > 0 ,   x 1 , x 2 , , x n , is a sample belong to a parametric family F δ ( x ) , and where
Y 2 ( δ _ ^ n ) = Z n 2 ( δ _ ^ n ) + n 1 L T ( δ _ ^ n ) ( Ι ( δ _ ^ n ) J ( δ _ ^ n ) ) 1 L ( δ _ ^ n ) ,
where
X n 2 ( δ _ ^ n ) = ( ς 1 n p 1 ( δ _ ^ n ) n p 1 ( δ _ ^ n ) , ς 2 n p 2 ( δ _ ^ n ) n p 2 ( δ _ ^ n ) , , ς n n p b ( δ _ ^ n ) n p b ( δ _ ^ n ) ) T
and J ( δ _ ^ n ) is the information (Inf-Mx) for the grouped data
J ( δ _ ^ n ) = ( δ _ ^ n ) T ( δ _ ^ n ) ,  
with
( δ _ ^ n ) = [ 1 p i p i ( δ _ ^ n ) δ ] r × s | ( i = 1 , 2 , , b   and   k = 1 , 2 , , s ) ,
and then
L ( δ ) = ( L 1 ( δ ) , , L s ( δ ) ) T   with   L k ( δ ) = i = 1 r ς i p i ( δ ) p i ( δ ) ,
where δ ^ is the estimated Fisher Inf-Mx and I ( δ ^ ) stands for the maximum likelihood estimator of the parameter vector. The Y 2 statistic has ( b 1 ) degrees of freedom and follows the χ b 1 2 distribution. Consider a set of observations x 1 , x 2 , , x n that are collected in I 1 , I 2 , , I b (these b subintervals are mutually disjoint:
I j = ( a j , b ( x ) 1 ; x j , b ( x ) )
The intervals I j ’s limits for a j , b ( x ) are determined as follows
p j ( δ ) = a j , b ( x ) 1 a j , b ( x ) f δ ( x ) d x | ( j = 1 , 2 , , b ) ,
and
a j , b ( x ) = F 1 ( j b ) | ( j = 1 , , b 1 ) .
The vector of frequencies is generated by grouping the data into I j intervals, ς j = ( ς 1 , ς 2 , , ς b ) T , where
ς j = i = 1 n 1 { x x I j }   | ( j = 1 , , b ) .
In order to determine if the utilized data are distributed in accordance with the OPPE model, in the situation of an unknown parameter δ , we design an NKRR test statistic in this study as a modified goodness-of-fit test. We use the estimated Fisher Inf-Mx to provide all the components of the Y 2 statistic of our model after computing the maximum likelihood estimator δ ^ of the unknown OPPE distribution’s parameter on the data set. For more applications under other different data sets, see Ibrahim et al. [8] and Yadav et al. [9].

3. Estimation and Inference

This section, via two subsections, discusses Bayesian and non-Bayesian estimating methods. Six non-Bayesian estimation methods are considered, including the maximum likelihood estimation (MLE), the Cramér–von Mises estimation (CVME), the ordinary least square estimation (OLSQE), the weighted-least square estimation (WLSQE), moment method, and the Kolmogorov estimation (KE).

3.1. Classical Estimation Methods

3.1.1. Maximum Likelihood Method

Using some observable data, maximum likelihood estimation (MLE), a statistical approach, can estimate the unknown parameter of a probability distribution. Consider the n-sample ( x 1 , x 2 , , x n ) and a fixed constant m , we assume that the m-sample ( x 1 , x 2 , , x m ) is generated from the OPPE distribution. The likelihood function of this sample is
L δ ( x ) = i = 1 m f δ ( x i ) [ 1 F δ ( x m ) ] n m
where N = n ! ( n m ) ! in both (1) and (2) we have
L δ ( x ) = N δ m δ m ( 1 e x p ( 1 ) ) n C δ ( x m ) n m i = 1 m A δ ( x i ) δ ( x i )
in which
A δ ( x i ) = Δ δ i ( x i ) [ δ x i 1 + e x p ( δ x i ) ] ,
δ ( x i ) = e x p ( δ x i ( 1 Δ δ ( x i ) ) ) ,
and
C δ ( x m ) = 1 e x p ( ( 1 Δ δ ( x m ) ) ) .
To obtain the maximum likelihood estimate (MLE) of δ , we have the log-likelihood ( ( δ ) ) function
( δ ) = l o g [ i = 1 n f δ ( x i , n ) ] ,
l δ ( x ) = l n N + m l n δ n l n ( 1 e x p ( 1 ) ) + ( n m ) l n C δ ( x m ) + i = 1 m l n A δ ( x i ) + i = 1 m l n δ ( x i ) .
The maximum likelihood estimator δ M L E of the parameter δ is the solution of the following non-linear eduation
l δ ( x ) δ = m δ + ( n m ) 1 C δ ( x m ) C δ ( x m ) δ + i = 1 m 1 A δ ( x i ) A δ ( x i ) δ + i = 1 m 1 δ ( x i ) δ ( x x ) δ = 0
where
δ Δ δ ( x i ) = [ x i x i e x p ( δ x i ) ( 1 δ x i ) ] e x p ( δ x i ( 1 e x p ( δ x i ) ) ,
δ A δ ( x i ) = ( δ Δ δ ( x i ) ) ( δ x i 1 + e x p ( δ x i ) ) + Δ δ ( x i ) ( x i + x i e x p ( δ x i ) ) ,
δ δ ( x i ) = [ ( δ Δ δ ( x i ) ) x i ] e x p ( δ x i ( 1 Δ δ ( x i ) ) ,
and
δ C δ ( x m ) = ( δ Δ δ ( x m ) ) e x p ( ( 1 Δ δ ( x m ) ) .

3.1.2. The CVME Method

The CVME of the parameter δ is obtained via minimizing the following expression with respect to δ , where
CVME ( δ ) = 1 12 n 1 + i = 1 n [ F δ ( x i , n ) l ( i , n ) [ 1 ] ] 2 | ( x i , n N ( 0 ) ) ,
and where l ( i , n ) [ 1 ] = 1 2 n ( 2 i 1 ) , and where
CVME ( δ ) = i = 1 n [ 1 e x p [ Δ ¯ δ ( x i , n ) ] 1 e x p ( 1 ) l ( i , n ) [ 1 ] ] 2 .
Then, CVME of the parameter δ is obtained by solving the following non-linear equation
0 = i = 1 n ( 1 e x p [ Δ ¯ δ ( x i , n ) ] 1 e x p ( 1 ) l ( i , n ) [ 1 ] ) ς ( δ ) ( x i , n , δ ) ,
where ς ( δ ) ( x i , n , δ ) = F δ ( x i , n ) / δ is the first derivatives of the CDF of OPPE distribution with respect to δ .

3.1.3. The OLSQ Method

Let F δ ( x i , n ) denote the CDF of the OPPE model and let x 1 < x 2 < < x n be the n ordered RS. The OLSQE ( O ( δ ) ) is obtained upon minimizing
O L S ( δ ) = i = 1 n [ F δ ( x i , n ) l ( i , n ) [ 2 ] ] 2 ,
where l ( i , n ) [ 2 ] = i n + 1 . Then, we have
O L S ( δ ) = i = 1 n [ 1 e x p [ Δ ¯ δ ( x i , n ) ] 1 e x p ( 1 ) l ( i , n ) [ 2 ] ] 2 .
The LSE is obtained via solving the following non-linear equation
0 = i = 1 n [ 1 e x p [ Δ ¯ δ ( x i , n ) ] 1 e x p ( 1 ) l ( i , n ) [ 2 ] ] ς ( δ ) ( x i , n , δ ) ,
where ς ( δ ) ( x i , n , δ ) is defined above.

3.1.4. The WLSQE Method

The WLSQE is obtained by minimizing the function W ( δ ) with respect to δ
W L S ( δ ) = i = 1 n d ( i , n ) [ 3 ] [ F δ ( x i , n ) l ( i , n ) [ 2 ] ] 2 ,
where l ( i , n ) [ 3 ] = [ ( 1 + n ) 2 ( 2 + n ) ] / [ i ( 1 + n i ) ] . The WLSQEs are obtained by solving
0 = i = 1 n l ( i , n ) [ 3 ] [ 1 e x p [ Δ ¯ δ ( x i , n ) ] 1 e x p ( 1 ) l ( i , n ) [ 2 ] ] ς ( δ ) ( x i , n , δ ) ,
where ς ( δ ) ( x i , n , δ ) is defined above.

3.1.5. Method of Moments

The method of moments is a technique used in statistics to estimate population parameters. To derive higher moments like skewness and kurtosis, the same method is applied. It begins by describing the population moments as functions of the important parameters (i.e., the anticipated powers of the random variable under discussion). The sample moments are then set to be equal to those. There are exactly as many of these equations as there are parameters that need to be estimated. Then, the equations are solved for the relevant parameters. Estimates of those parameters are used in the solutions. The moment estimation of the one-parameter of the OPPE distribution can be obtained by equating the first theoretical moment of (2) with the corresponding sample moments as follows
μ 1 = E ( X i , n ) = 1 n i = 1 n x i , n
where μ 1 can be evaluated from (2).

3.1.6. KE Method

The Kolmogorov estimate (KE) δ ^ of δ is obtained by minimizing the function
K E = K E ( δ ) = m a x 1 i n { 1 n i F δ ( x i , n ) , F δ ( x i , n ) 1 n ( i 1 ) } .
The KE δ ^ of δ is obtained by comparing [ 1 n i F δ ( x i , n ) ] | 1 i n and [ F δ ( x i , n ) 1 n ( i 1 ) ] | 1 i n and selecting the max one. However, for 1 i n , we are minimizing the whole function K ( δ ) . For more detail about the KE method.

3.2. Simulations and Assessment

For comparing the classical methods, some MCMC simulation studies are performed. The results are presented in: Table 1 (δ = 0.5|n = 50, 100, 200, 300, and 500); Table 2 (δ = 0.9|n = 50, 100, 200, 300, and 500) and Table 3 (δ = 1.5|n = 50, 100, 200, 300, and 500). The numerical assessments are performed depending on the mean squared errors (MSEs). First, we generate N = 1000 samples of the OPPE model. Based on Table 1, Table 2 and Table 3, it is noted that the performance of all estimation methods improves when n + . Despite the variety and abundance of the other classic methods, as demonstrated in Table 1, Table 2 and Table 3, the MLE approach is still the most efficient and reliable of the surviving classic methods. The MLE approach is generally mentioned as being advised for statistical modeling and applications. This evaluation is mostly based on a thorough simulation research, as displayed in Table 1, Table 2 and Table 3. Since the MSE for the MLE is the smallest for all n = 50, 100, 200, 300, and 500, this section uses simulation studies to evaluate rather than compare various estimating procedures. However, this does not exclude the use of simulation to compare various estimating approaches. Nevertheless, actual data are commonly used to assess different estimating approaches; for this reason, we will discuss a few examples specifically for this role. There are two more applications to compare the competing models.

3.3. Applications for Comparing Methods

Here, we will be very interested in comparing the different estimation methods, but this time through real data. The first data set called the failure time data or relief times (in minutes) of patients receiving an analgesic (see Gross and Clark [10]). The second data set is called the survival times (in days) of 72 guinea pigs infected with virulent tubercle bacilli, see Bjerkedal [11].
We discuss the skewness–kurtosis plot (or the Cullen and Frey plot) in these applications for examining initial fits of theoretical distributions such as normal, uniform, exponential, logistic, beta, lognormal, and Weibull. Plotting and bootstrapping are both applied for greater accuracy. The scattergram plots, the “nonparametric Kernel density estimation (NKDE)” method for examining the initial shape of the insurance claims density, the “Quantile-Quantile (Q-Q)” plot for examining the “normality” of the current data, the “total time in test (TTT)” plot for examining the initial shape of the empirical hazard rate function (HRF), and the “box plot” for identifying the extreme data were also presented.
Figure 1 gives the box plot for the failure time data (first row, the left panel), Q-Q plot for the failure time data (first row, the right panel), TTT plot for the failure time data (second row, the left panel), nonparametric Kernel density estimation plot for the failure time data (second row, the right panel), the Cullen and Frey plot for the failure time data (third row, the left panel), and scattergrams (third row, the right panel) for the failure time data.
Based on Figure 1 (first row), the relief data have only one extreme observation, based on Figure 1 (second row, the left panel), the HRF of the relief times is “monotonically increasing HRF”, based on Figure 1 (second row, the right panel), nonparametric Kernel density estimation is bimodal and right skewed with asymmetric shape, based on Figure 1 (third row, the left panel), the relief times data do not follow any of the theoretical distributions such as the normal, uniform, exponential, logistic, beta, lognormal, and Weibull. Figure 2 gives the box plot for the survival times (first row, the left panel), Q-Q plot for the survival times (first row, the right panel), TTT plot for the survival times (second row, the left panel), nonparametric Kernel density estimation plot (second row, the right panel), the Cullen and Frey plot (third row, the left panel), and scattergrams (third row, the right panel) for the survival times data. The survival data in Figure 2 (first row) has four extreme observations, the HRF of the survival times is “monotonically increasing HRF,” the nonparametric Kernel density estimation is bimodal and right skewed with an asymmetric shape, and the survival times data in Figure 2 (third row, left panel) do not follow any of the theoretical distributions, such as the normal or gamma distributions.
A statistical test called the Anderson–Darling test (ADT) is used to determine if a sample of data is taken from a particular probability distribution. The test is based on the assumption that the distribution under the test has no parameters that need to be estimated, in which case the test and its set of critical values are distribution-free. The test is most frequently applied when a family of distributions is being examined, in which case it is necessary to estimate the family’s parameters and take this into consideration when modifying the test statistic or its critical values. It is one of the most effective statistical strategies for detecting most deviations from normal distribution when used to determine whether a normal distribution adequately describes a collection of data.
The Cramér–von Mises criterion (CVMC) in statistics is a criterion used to compare two empirical distributions or to determine how well a cumulative distribution function fits an empirical distribution function. In other techniques, such as minimum distance, the CVMC and ADT tests are used in comparing methods. Table 4 gives the application results (CVMC and ADT) for comparing methods under the relief data. Table 5 lists the application CVMC and ADT for comparing methods under the survival data. Based on Table 4, it is seen that the moment method is the best with CVMC = 0.08439 and ADT = 0.499304, then the ML method with W CVMC = 0.085876 and ADT = 0.508074. Based on Table 4, it is seen that the moment method is the best with CVMC = 0.070712 and ADT = 0.463826, then the ML method with CVMC = 0.071247 and ADT = 0.465521.

3.4. Bayesian Analysis under Different Loss Functions

In this section, we will review some statistical aspects related to Bayesian estimations. Three types of Bayesian loss functions have been used. Additionally, I made a lot of useful comparisons using the new distribution. Many Bayesian statistical algorithms can be utilized, as per Muñoz-Gil et al. [12], and within their particular context, they can aid researchers in mathematics and statistical modeling tasks, particularly when applying Bayes’ theory. Muñoz-Gil et al. [12] offered the proposed representation of four full algorithms with their codes and implementations for this specific purpose, allowing the reader to follow the new algorithms and present more algorithms based on them.

3.4.1. Prior and Posterior Distributions

As prior distributions, we assume the parameter δ has an informative distribution as a prior, which is a Gamma distribution:
π ( δ ) = a b Γ ( b ) δ b 1 e x p ( a δ )   , a , b > 0 .
The posterior distribution of δ is
π ( δ | x ) = K δ m + b 1 [ 1 e x p ( 1 ) ] m e x p ( a δ ) i = 1 m P δ ( x i ) Δ δ ( x i ) Q δ ( x m ) n m ,
where
K = 0 + δ m + b 1 [ 1 e x p ( 1 ) ] m e x p ( a δ ) i = 1 m P δ ( x i ) Δ δ ( x i ) Q δ ( x m ) n m d δ ,
is the normalizing constant. In the following, we use the three loss functions, namely the generalized quadratic (GQ), the entropy, and the Linex loss functions to obtain the Bayesian estimators. In the following formulas, we consider γ, P, and r as integers.

3.4.2. Bayesian Estimators and Their Posterior Risk

The Bayesian estimator under the generalized quadratic loss function ( G Q ) is
δ G Q = I 0 + ( m + b 1 + γ , x m ) I 0 + ( m + b 2 + γ , x m ) ,
where
I 0 + ( m + b 1 + γ , x m ) = 0 + δ m + b 1 + γ [ 1 e x p ( 1 ) ] m e x p ( a δ ) i = 1 m Δ δ ( x i ) P δ ( x i ) Q δ ( x m ) n m d δ
and
I 0 + ( m + b 2 + γ , x m ) = 0 + δ m + b 2 + γ [ 1 e x p ( 1 ) ] m e x p ( a δ ) i = 1 m Δ δ ( x i ) P δ ( x i ) Q δ ( x m ) n m d δ .
Under the entropy loss function, we obtain the following estimator
δ E = [ K 0 + δ m + b 1 p [ 1 e x p ( 1 ) ] m e x p ( a δ ) i = 1 m Δ δ ( x i ) P δ ( x i ) Q δ ( x m ) n m d δ ] 1 p .
The corresponding posterior risk is
P R ( δ E ) = P E π ( l n ( δ ) l n ( δ E ) ) ,
finally, under the Linex loss function, the Bayesian estimator
δ L = K r l n [ 0 + δ m + b 1 [ 1 e x p ( 1 ) ] m e x p [ δ ( a + r ) ] i = 1 m P δ ( x i ) Δ δ ( x i ) Q δ ( x m ) n m d δ ] ,
and the corresponding posterior risk is
P R ( δ L ) = r ( δ G Q δ L ) .
Since it is unlikely possible to obtain all these estimators analytically, we suggest the use of the MCMC procedures to evaluate them. In this work, several Bayesian algorithms were developed recently for the analysis of tracer-diffusion single-particle tracking data. Following Muñoz-Gil et al. [12], many Bayesians statistical algorithms can be used, which within their specific framework, can help researchers in mathematical and statistical modeling operations, especially when using Bayes’ theory. For this specific purpose, Muñoz-Gil et al. [12] presented the proposed visualization of four complete algorithms with their codes and applications, and the reader can track the new algorithms and other algorithms can be presented based on them.

3.4.3. Comparing the Likelihood Estimation and the Bayesian Estimation Using Pitman’s Closeness Criterion

We are going to compare the performance of the proposed Bayesian estimators with the MLEs, for that purpose, we perform a MCMC simulation method with δ = 1.5 and a = 3, b = 2. We generate N = 1000 type II censored samples following the OPPE model, we use different sample sizes n = 30, 100, 200 while m = 10, 40, 160, respectively, and we obtain the following results. Table 6 contains the values of the estimators produced by the function BB algorithm. Here, we observe that, especially as sample size n is raised, the estimated values are fairly close to the parameter’s genuine values. Table 7, Table 8 and Table 9 present the Bayesian estimators and PR (in brackets) under the generalized quadratic loss function, the entropy loss function, and the Linex loss function, respectively. Table 10 lists the Bayesian estimators and PR (in brackets) for each of the three loss functions.
In Table 7, the estimation under the generalized quadratic loss function, we remark that the value γ = 1 gives the best posterior risk. Additionally, we obtain the smallest suitable posterior risk when n is long. In the estimation under the entropy loss function, we obtain Table 8 where we can notice that the value p = 0.5 when n = 200 provides the best posterior risk. We can notice clearly that the value r = 1.5 provides the best PR. In conclusion, it is evident from a brief comparison of the three loss functions that the quadratic loss function yields the best outcomes; the results are further illustrated in Table 10. The best Bayesian estimators and maximum likelihood estimators should be compared, as we suggest. We employ the Pitman closeness criterion for this purpose (more information can be found in Pitman [13], Fuller [14], and Jozani [15]).
Definition 1.
An estimator  ϑ 1  of a parameter  ϑ  dominates another estimator  ϑ 2  in the sense of Pitman’s closeness criterion if, for all  ϑ Θ ,
P ϑ [ | ϑ 1 ϑ | < | ϑ 2 ϑ | ] > 0.5 .
We show the Pitman probability’ values in Table 11 so that we may contrast the Bayesian estimators with the MLE estimator when using the three loss functions when γ = −1, p = −0.5, and r = −1.5. Then, we notice that, according to this criterion, the Bayesian estimators of the parameter is better than the MLE. Additionally, the Linex loss function has the best values in comparison with the other two loss functions with the probability 0.734, δ(n=200,m=160).

4. Distributional Validation

We utilize the statistic type test based on a version of the NKRR statistic suggested by Bagdonavičius et al. [1], Bagdonavičius and Nikulin [2], Bagdonavičius and Nikulin [3], Nikulin [4], Nikulin [5], Nikulin [6], and Rao and Robson [7] to confirm the sufficiency of the OPPE model when the parameter is unknown, and the data are censored. The failure rate x i follows an OPPE distribution, hence we modify this test for an OPPE model. Let’= us think about the null hypothesis:
H 0 : F ( x ) F 0 = F 0 , δ ( x ) | x R ,
The OPPE distribution’s survival function (SrF) and cumulative hazard function are as follows:
S δ ( x ) = 1 F δ ( x ) = 1 [ 1 1 e x p ( 1 ) ( 1 e x p { [ 1 Δ δ ( x ) ] } ) ] ,
where
Δ δ ( x ) = [ e x p ( δ x ) ] [ 1 e x p ( δ x ) ] .
and
V δ ( x ) = ln [ S δ ( x ) ] ,
For all j , we have a constant value of e j , X = E k / k under this choice of intervals. Since the inverse hazard function of the OPPE distribution lacks an explicit form, intervals can be estimated iteratively. Let us divide a finite time period [ 0 , τ ] into k > s smaller periods I j = ( a j 1 ( x ) , a j , b ( x ) ] and
0 = < a 0 , b < a 1 , b < a k 1 , b < a k , b = + .
If x ( x ) is the ith element in the ordered statistics ( x ( 1 ) , , x ( n ) ) , and if [ V δ ( x ) ] 1 is the inverse of the cumulative hazard function V δ ( x ) , then the estimated value of  a ^ j , b ( x ) where
E k = E k ( x ) = l = 1 i 1 V δ ( x ) ,
and a j , b ( x ) are random data functions, such as the k selected intervals have equal expected numbers of failures e j , X . The test for hypothesis H 0 can be based on the statistic
Y n , r 1 , ε 2 ( δ ^ ) = Z T S ^ Z
where Z = ( Z 1 , Z 2 , , Z k ) T , and
Z j = 1 n ( O j , X e j , X ) | ( j = 1 , 2 , , k ) ,
and O j , X reflect the total number of failures that have been noticed throughout these times. Due to Bagdonavičius et al. [1], Bagdonavičius and Nikulin [2], Bagdonavičius and Nikulin [3], Nikulin [4], Nikulin [5], Nikulin [6], and Rao and Robson [7], the test statistic is expressed as:
Y n , r 1 , ε 2 ( δ ^ ) = j = 1 k 1 O j , X ( O j , X e j , X ) 2 + V W , G ,
where V W , G and many other details are given in Bagdonavičius et al. [1], Bagdonavičius and Nikulin [2], Bagdonavičius and Nikulin [3], Nikulin [4], Nikulin [5], Nikulin [6], and Rao and Robson [7]. We compute each component of the Y n , r 1 , ε 2 ( δ ^ ) statistic for the OPPE model. The statistic Y n , r 1 , ε 2 ( δ ^ ) has a chi-square limit distribution, and its degree of freedom is d f = r a n k ( S ) = t r a c e ( S 1 S ) where d f = k . The estimated significance threshold is rejected if Y n , r 1 , ε 2 ( δ ^ ) > χ ε 2 ( d f ) (where χ ε 2 ( d f ) is the quantile of chi-square with r = d f = r a n k ( S ) ), then the approximate significance level ε is rejected Hypothesis, where P l j ^ is the Fisher information.

4.1. Uncensored Simulation Study under the NKRR Statistics Y 2 ( ε ; δ ^ )

We conducted an extensive study using numerical simulation to confirm the arguments of this work. Therefore, we generated the N statistics of 15,000 simulated samples with sizes of n = 25, n = 50, n = 150, n = 350, and n = 600, in order to test the null hypothesis H 0 that the sample belongs to the OPPE model. We determine the average of the non-rejection numbers of the null hypothesis Y 2 ( ε ; δ ^ )   χ ε 2 ( b 1 ) for various theoretical levels (ε = 0.01, 0.02, 0.05, 0.1). The corresponding empirical and theoretical levels are represented in Table 12. It can be seen that the calculated empirical level value is very close to its corresponding theoretical level value. Therefore, we conclude that the recommended test is very suitable for the OPPE distribution.

4.2. Uncensored Applications under the NKRR Statistics Y 2 ( ε ; δ ^ )

4.2.1. Strengths of Glass Fibers

In this example, 100 carbon fiber fracture stresses (in Gba) are included in this data set given by Nicholase and Padgett [16]. Using the BB algorithm, we can get the MLE value of the parameter δ, assuming that our OPPE model can fit the strength data of 1.5 cm glass fiber: δ ^ = 2.14068 . We can compute and provide the Fisher Inf-Mx as follows using the value:
I ( δ ^ ) = 1.648771 .
The critical values for the NKRR statistical test were: Y 2 ( ε ; δ ^ ) = 11.855642 and χ 0.05 2 ( 6 ) = 12.59159 , so the OPPE distribution can effectively simulate and model the 1.5 cm glass fiber data.

4.2.2. Heat Exchanger Tube Crack

The crack data, which includes examinations carried out at eight chosen intervals until fractures emerged in 167 similar turbine parts, was acquired from the book by Meeker and Escobar [17].
Time   of   inspection 186 606 902 1077 1209 1377 1592 1932 Number   of   fans   found   to   have   cracks 5 16 12 18 18 2 6 17
We test the null hypothesis that these data are modified by our OPPE distribution using previously acquired NKRR statistics. We calculate the MLE δ ^ = 2.475019 using R programming and the BB method (see Ravi (2009)). The estimated Fisher Inf-Mx at that time is:
I ( δ ^ ) = 1.592466
So, we arrive with Y 2 ( ε ; δ ^ ) = 19.84927 as the answer. The crucial value for significance level ε = 0.05 and χ 0.01 2 ( 12 ) = 21.02607 . This model’s NKRR statistic ( Y 2 ( ε ; δ ^ ) ) is less than the essential value, allowing us to conclude that the data correctly fit the OPPE model.

4.3. Censored Simulation Study under the NKRR Statistics Y n , r 1 , ε 2 ( δ ^ )

In the censored simulation study under the NKRR statistics Y 2 , it is expected that the produced sample ( N = 14000 ) is censored at 25 % with d f = 5 . We determine the average value of the non-rejection numbers of the null hypothesis for various theoretical levels ( ε = 0.01 ,   0.02 ,   0.05 ,   0.1 ) , where Y n , r 1 , ε 2 ( δ ^ ) χ ε 2 ( r 1 ) . The matching theoretical and empirical levels are presented in Table 13, demonstrating how closely the determined empirical level value resembles the associated theoretical level value. Therefore, we draw the conclusion that the custom test is ideally suited to the OPPE model. These findings lead us to the conclusion that the theoretical level of the chi-square distribution on degrees of freedom corresponds to the empirical significance level of the Y n , r 1 , ε 2 ( δ ^ ) statistics at which it is statistically significant. It may be inferred from this that the suggested test can accurately fit the censored data from the OPPE distribution. Returning to the defects of the NKRR basic test, the statistical literature contains many statistical hypothesis tests that fit the complete data, and for this reason, the original test has many competing tests in this field, and the fact is that the NKRR basic test actually has many alternatives. As for the modified NKRR test, it is considered an individual statistical test of its kind in the field of statistical tests due to its importance and the nature of the data that it can deal with, which are the data subject to censorship.

4.4. Censored Applications under the NKRR Statistics Y n , x 1 , ε 2 ( δ ^ )

4.4.1. Censored Lung Cancer Data Set

The lung cancer data provided by Loprinzi et al. [18] from the North Central Cancer Treatment Group investigated the survival in patients with advanced lung cancer (n = 228 and censored items = 63), and their performance scores rate how well the patient can do typical daily activities. If we assume that the data are distributed according to the OPPE distribution, we can estimate the vector parameter δ ^ by applying the maximum likelihood estimation approach as: δ ^ = 3.004756 . As a number of classes, we employ d f = 8 . The following is how the test statistic Y n , r 1 , ε 2 ( δ ^ ) items are presented:
a j , b ( X ) ^ 92.138171.694216.037283.012355.086456.277685.2611022.3174
O j , X ^ 2930353132252818
e j , X 6.554386.554386.554386.554386.554386.554386.554386.55438
The estimated matrix P l j ^ ( X ) and the estimated information (E-Inf) I ( δ ^ ) via Fisher are as follows:
P l j ^ ( X ) −0.6827−0.43680.70150.76840.50940.30070.8769−0.2884
and
I ( δ ^ ) = 3.0134022 .
The chi-squared test has a critical value of χ 0.05 2 ( d f = 8 ) = 15.50731 . We find that the estimated statistic for the suggested test is Y n , r 1 , ε 2 ( δ ^ ) = 14.616535 using the earlier findings. We may state that our hypothesis H 0 is accepted because the tabulated value of the Y n , r 1 , ε 2 ( δ ^ ) statistic is higher than the computed value. This leads us to the conclusion that there is a 5% chance that the lung cancer data will deviate from the OPPE distribution.

4.4.2. Censored Capacitor Data Reliability Data Set

A set of data for basic reliability assessments data on glass capacitor longevity as a function of voltage and operating temperature from a factorial experiment (see Meeker and Escobar [17]). Each combination of temperature and voltage has eight capacitors. Testing was stopped after the fourth failure at each combination, n = 64 and censored items = 32. The parameter vector’s maximum likelihood estimator δ ^ , assuming that the data are distributed according to the OPPE distribution, is: δ ^ = 1.680225 . We pick a few classes with d f = 8 . The components of the statistical test Y n , r 1 , ε 2 ( δ ^ ) are as follows:
a j , b ( X ) ^ 346.1573469.502587.113679.0171078.8341089.1091102.1671106.444
O j , X ^ 11156106565
e j , X 6.910426.910426.910426.910426.910426.910426.910426.91042
The estimated matrix P l j ^ ( X ) and Fisher’s E-Inf I ( δ ^ ) are:
P l j ^ ( X ) 0.48965−0.73465−0.400120.29784−0.29467−0.951130.603780.28564
and
I ( δ ^ ) = 4.1675894
The value of the statistical test Y n , r 1 , ε 2 ( δ ^ ) = 13.84577 is then assessed. χ 0.05 2 ( 8 ) = 15.50731 > Y n , r 1 , ε 2 ( δ ^ ) is the crucial value. We may conclude that the OPPE model is used to update the life statistics for glass capacitors.

5. Conclusions

The one parameter Poisson-exponential (OPPE) model, a new adaptable variation of the exponentiated exponential model, is introduced and studied in this article. Six well-known estimation techniques are investigated, discussed, and used: Cramer–von Mises, ordinary least square, L-moments, maximum likelihood, Kolmogorov, and weighted-least square. The Bayesian estimation under the squared error loss function was also devised and investigated. The use of two real data and two simulated data allows for a complete evaluation of all available techniques. The utility and adaptability of the OPPE distribution are illustrated using two real data sets. The new model outperforms many other competing models in modeling relief times and survival times, according to the Akaike Information Criterion, Consistent Akaike Information Criterion, Hannan–Quinn Information Criterion, Bayesian Information Criterion, Cramér–von Mises, and Anderson–Darling statistics. However, there are more results the reader can read by examining the paper and the applications, the following results can be specifically highlighted:
  • The maximum likelihood method is still the most effective and trustworthy of the remaining classic approaches. Both the Bayesian technique and the Maximum Likelihood method are recommended for statistical modeling and applications.
  • Under different loss functions, the Bayesian estimation is provided. Three loss functions, the generalized quadratic, the Linex, and the entropy, are used to produce the Bayesian estimators, and many useful details are provided.
  • All of the provided estimation methods have been evaluated through simulation tests with particular parameter and controls (these simulation studies are all stated in the paper at the appropriate places).
  • The BB algorithm for process estimation under censored samples is used to compare the Bayesian technique and the censored maximum likelihood method.
  • It is shown in detail how the NKRR statistic is created for the OPPE model in the unfiltered instance. The results of a simulation research show that the OPPE model and the NKRR test are a good fit.
  • A simulation study for evaluating statistics is described, along with the construction of the Bagdonavičius and Nikulin statistic for the novel model under the censored situation. Two real data applications are examined in a censored scenario; the first data are reliability information on capacitors, and the second data are information about lung cancer (medical data). We deduce from these applications that the suggested technique can successfully fit censored data from the OPPE distribution.
  • For the uncensored strengths of glass fibers data, the critical values for the NKRR statistical test were: Y 2 ( ε ; δ ^ )   = 11.855642 and χ 0.05 2 ( 6 ) = 12.59159 , so the OPPE distribution can effectively simulate and model the uncensored 1.5 cm glass fiber data.
  • For the uncensored heat exchanger tube crack data, the critical values for the NKRR statistical test were: Y 2 ( ε ; δ ^ )   = 19.84927 and χ 0.05 2 ( 6 ) = 12.59159 , so the OPPE distribution can effectively simulate and model the uncensored heat exchanger tube crack data.
  • For the censored lung cancer data set: the value of the statistical test Y n , r 1 , ε 2 ( δ ^ ) = 13.84577 , where χ 0.05 2 ( 8 ) = 15.50731 > Y n , r 1 , ε 2 ( δ ^ ) = 14.616535 . Then, we conclude that the OPPE model can be used for modeling the censored lung cancer data set.
  • For censored capacitor data reliability data set: the value of the statistical test Y n , r 1 , ε 2 ( δ ^ ) = 13.84577 , where χ 0.05 2 ( 8 ) = 15.50731 > Y n , r 1 , ε 2 ( δ ^ ) = 13.84577 . Then, we conclude that the OPPE model can be used for modeling the censored capacitor data reliability data set.
  • It is worth noting that NKRR basic test, which is supported by complete data, is the most popular test in the last ten years. This is because it fits the complete truth data, and this is the case in most practical and applied cases in various fields. What distinguishes this statistical test also is the availability of ready-made statistical packages on the R program, whether for simulation or applications on actual data.
  • However, the practical and experimental reality in many fields (such as the medical, chemical, engineering, etc.) necessitates that researchers deal with practical experiments that produce controlled censored data. This type of data, of course, needs certain tests dedicated to statistical dealing with it in the problem of distributional validation. The NKRR basic test is not the optimal choice in these cases. Hence, and based on this dilemma, was the primary and most important motive that prompted many researchers to think about introducing a new statistical test that fits the censored data. This new NKRR test, of course, is a modified test from the NKRR original test.
  • According to the nature of the procedures of the two tests, both tests are not suitable for working with a type of data. For example, one of the drawbacks of the original test is that it is not suitable for dealing with censored data, it is only for complete data and cannot be applied to censored data. Therefore, if we are dealing with complete data, then this test may be a strong candidate for performing statistical hypothesis tests.
  • Moreover, a disadvantage of the modified test is that it is only suitable for dealing with censored data, it is not intended for complete data and can only be applied to censored data. Therefore, if we are dealing with censored data, this test will undoubtedly be a strong candidate for statistical hypothesis tests, because it is intended for this type of data.
  • Returning to the defects of the NKRR basic test, the statistical literature contains many statistical hypothesis tests that fit the complete data, and for this reason, the original test has many competing tests in this field, and the fact is that the NKRR basic test actually has many alternatives.
  • As for the modified NKRR test, it is considered an individual statistical test of its kind in the field of statistical tests due to its importance and the nature of the data that it can deal with, which are the data subject to censorship.

Author Contributions

W.E.: validation, writing the original draft preparation, conceptualization, data curation, formal analysis, software. Y.T.: funding acquisition methodology, formal analysis, conceptualization, software. H.G.: validation, formal analysis, data duration, conceptualization, software. T.H.: validation, formal analysis, data duration, conceptualization, formal analysis, software. A.H.: validation, data duration, conceptualization, software. M.M.A.: review and editing, conceptualization, formal analysis, supervision. H.M.Y.: review and editing, software, validation, writing the original draft preparation, conceptualization, supervision. M.I.: review and editing, formal analysis, software, validation, writing the original draft preparation, conceptualization. All authors have read and agreed to the published version of the manuscript.

Funding

The study was funded by Researchers Supporting Project number (RSP2023R488), King Saud University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset can be provided upon requested.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

CDFCumulative distribution function
OPPEOne parameter Poisson-exponential
SrFSurvival function
ADTAnderson-Darling test
CVMECramér–von Mises estimation
MLEsMaximum likelihood estimations
HRFHazard rate function
Inf-MxInformation matrix
E- InfEstimated information
OLSQEOrdinary least squares estimation
KEKolmogorov estimation
WLSQEWeighted least squares estimation
MSEMean square error
( δ )   Log-likelihood
CVMCCramér–von Mises criterion
dfDegrees of freedom
P–PProbability-probability
TTTTotal time in test
Q-QQuantile-quantile
NKRRNikulin–Rao–Robson

References

  1. Bagdonavičius, V.B.; Levuliene, R.J.; Nikulin, M.S. Chi-Squared Goodness-of-Fit Tests for Parametric Accelerated Failure Time Models. Commun. Stat.-Theory Methods 2013, 42, 2768–2785. [Google Scholar]
  2. Bagdonavičius, V.; Nikulin, M. Chi-squared goodness-of-fit test for right censored data. Int. J. Appl. Math. Stat. 2011, 24, 30–50. [Google Scholar]
  3. Bagdonavičius, V.; Nikulin, M. Chi-squared tests for general composite hypotheses from censored samples. Comptes Rendus Math. 2011, 349, 219–223. [Google Scholar]
  4. Nikulin., M.S. Chi-squared test for normality. In Proceedings of the International Vilnius Conference on Probability Theory and Mathematical Statistics; Vilnius University: Vilnius, Lithuania, 1973; Volume 2, pp. 119–122. [Google Scholar]
  5. Nikulin, M.S. Chi-squared test for continuous distributions with shift and scale parameter. Theory Probab. Its Appl. 1973, 18, 559–568. [Google Scholar] [CrossRef]
  6. Nikulin, M.S. On a Chi-squared test for continuous distributions. Theory Probab. Its Appl. 1973, 19, 638–639. [Google Scholar]
  7. Rao, K.C.; Robson, D.S. A Chi-Square Statistic for Goodness-of-Fit Tests Within the Exponential Family. Commun. Stat.-Simul. Comput. 1974, 3, 1139–1153. [Google Scholar]
  8. Ibrahim, M.; Hamedani, G.G.; Butt, N.S.; Yousof, H.M. Expanding the Nadarajah Haghighi Model: Copula, Censored and Uncensored Validation, Characterizations and Applications. Pak. J. Stat. Oper. Res. 2022, 18, 537–553. [Google Scholar] [CrossRef]
  9. Yadav, A.S.; Shukla, S.; Goual, H.; Saha, M.; Yousof, H.M. Validation of xgamma exponential model via Nikulin-Rao-Robson goodness-of- fit test under complete and censored sample with different methods of estimation. Stat. Optim. Inf. Comput. 2022, 10, 457–483. [Google Scholar]
  10. Gross, J.; Clark, V.A. Survival Distributions: Reliability Applications in the Biometrical Sciences; John Wiley: New York, NY, USA, 1975. [Google Scholar]
  11. Bjerkedal, T. Acquisition of resistance in guinea pigs infected with different doses of virulent tubercle bacilli. Am. J. Epidemiol. 1960, 72, 130–148. [Google Scholar]
  12. Muñoz-Gil, G.; Volpe, G.; Garcia-March, M.A.; Aghion, E.; Argun, A.; Hong, C.B.; Manzo, C. Objective comparison of methods to decode anomalous diffusion. Nat. Commun. 2021, 12, 6253. [Google Scholar] [CrossRef]
  13. Pitman, E. The closest estimates of statistical parameter. In Mathematical Proceeding of the Cambridge Philosophical Society; Cambridge University Press: Cambridge, UK, 1937; Volume 33. [Google Scholar]
  14. Fuller, W.A. Closeness of estimators. In Encyclopedia of Statistical Sciences; Wiley: New York, NY, USA, 1982; Volume 2. [Google Scholar]
  15. Jozani, M.J.; Davies, K.F.; Balakrishnan, N. Pitman closeness results concerning ranked set sampling. Stat. Probab. Lett. 2012, 82, 2260–2269. [Google Scholar] [CrossRef]
  16. Nicholase, M.D.; Padgett, W.J. A bootstrap control chart for Weibull percentiles. Qual. Reliab. Eng. Int. 2006, 22, 141–151. [Google Scholar] [CrossRef]
  17. Meeker, W.Q.; Escobar, L.A.; Lu, C.J. Accelerated degradation tests: Modeling and analysis. Technometrics 1998, 40, 89–99. [Google Scholar] [CrossRef]
  18. Loprinzi, C.L.; ALaurie, J.; Wieand, H.S.; EKrook, J.; Novotny, P.J.; Kugler, J.W.; Bartel, J.; Law, M.; Bateman, M.; EKlatt, N. Prospective evaluation of prognostic variables from patient-completed questionnaires. North Central Cancer Treatment Group. J. Clin. Oncol. 1994, 12, 601–607. [Google Scholar] [CrossRef]
Figure 1. Box plot, Q-Q plot, TTT plot, nonparametric Kernel density estimation, Cullen and Frey, and scattergrams for the relief times.
Figure 1. Box plot, Q-Q plot, TTT plot, nonparametric Kernel density estimation, Cullen and Frey, and scattergrams for the relief times.
Mathematics 11 00897 g001
Figure 2. Box plot, Q-Q plot, TTT plot, nonparametric Kernel density estimation, Cullen and Frey, and scattergrams for the survival times.
Figure 2. Box plot, Q-Q plot, TTT plot, nonparametric Kernel density estimation, Cullen and Frey, and scattergrams for the survival times.
Mathematics 11 00897 g002
Table 1. Simulation results for δ = 0.5.
Table 1. Simulation results for δ = 0.5.
nMLEOLSQWLSQCVMMomentKE
500.003050.003340.003140.003550.003290.00366
1000.001480.001630.001630.001610.001630.00179
2000.000730.000810.000810.000800.000820.00088
3000.000470.000540.000510.000540.000490.00055
5000.000270.000310.000310.000310.000300.00032
Table 2. Simulation results for δ = 0.9.
Table 2. Simulation results for δ = 0.9.
nMLEOLSQWLSQCVMMomentKE
500.009810.010730.010350.009390.011950.01187
1000.004890.005330.005110.005560.005390.00616
2000.002480.002860.002780.002820.002390.00285
3000.001550.001770.001770.001780.001740.00180
5000.000920.000980.0009870.000990.000950.00108
Table 3. Simulation results for δ = 1.5.
Table 3. Simulation results for δ = 1.5.
nMLEOLSQWLSQCVMMomentKE
500.025800.031890.028470.028880.030390.03448
1000.012960.015910.014140.015090.014080.01582
2000.006330.007490.007320.007440.007400.00785
3000.004600.004660.004980.004800.004940.00504
5000.002710.002990.002910.002980.003020.00301
Table 4. p-values for comparing methods under the relief data.
Table 4. p-values for comparing methods under the relief data.
Method δCVMC ADT
ML0.5032840.0858760.508074
LS0.4709170.0869230.514258
WLSQ0.4287570.0884660.523383
CVM0.4709550.0869210.514251
Moment0.5576760.084390.499304
KE0.4679100.0870260.514869
Table 5. p-values for comparing methods under the survival data.
Table 5. p-values for comparing methods under the survival data.
Method δCVMCADT
ML0.5677830.0712470.465521
LS0.5463700.0716900.467139
WLSQ0.5513500.0715810.466728
CVM0.5465760.0716850.467122
Moment0.5992470.0707120.463826
KE0.5279650.0721240.468851
Table 6. The MLE of the parameter with quadratic error (in brackets).
Table 6. The MLE of the parameter with quadratic error (in brackets).
N = 5000n = 30n = 100n = 200
m1040160
1.4905 1.465431.5104
δ→(0.0125)(0.0321)(0.0023)
Table 7. Bayesian estimators and PR (in brackets) under GQ loss function.
Table 7. Bayesian estimators and PR (in brackets) under GQ loss function.
γ ↓ N = 5000, mn = 30, 10n = 100, 40n = 200, 160
−21.6490 (0.0076)1.6825 (0.0053)1.6432 (0.0023)
−1.51.7990 (0.0087)1.0825 (0.0061)1.2127 (0.0016)
−11.6182 (0.0005)0.9739 (0.0001)2.0018 (0.0001)
−0.51.4994 (0.0071)1.4888 (0.0060)1.5138 (0.0011)
0.51.9760 (0.0085)1.7926 (0.0077)1.3439 (0.0013)
11.7516 (0.0087)1.2972 (0.0058)1.7208 (0.0035)
1.51.6743 (0.0045)1.5632 (0.0065)1.3275 (0.0032)
21.4768 (0.1241)1.4191 (0.1181)1.7158 (0.0033)
Table 8. Bayesian estimators and PR (in brackets) under the entropy loss function.
Table 8. Bayesian estimators and PR (in brackets) under the entropy loss function.
N = 5000n = 30n = 100n = 200
p ↓ m1040160
−21.6037 (0.0009)0.7757 (0.3190)1.8493 (0.0308)
−1.51.3990 (0.1644)1.2144 (0.0019)1.0942 (0.0080)
−11.0886 (0.0031)0.7654 (0.1173)1.7697 (0.0099)
−0.51.6001 (0.0065)1.542 (0.00020)1.5638 (0.0071)
0.51.483 (0.00413)1.4899 (0.0023)1.5121 (0.0012)
11.4579 (0.0997)1.4354 (0.0944)1.4571 (0.0014)
1.51.6981 (0.0038)0.4830 (0.0733)1.2148 (0.0009)
21.356 (0.16440)1.0942 (0.0080)1.2144 (0.0019)
Table 9. Bayesian estimators and PR (in brackets) under Linex loss function.
Table 9. Bayesian estimators and PR (in brackets) under Linex loss function.
N = 5000n = 30n = 100n = 200
r↓m1040160
−2 1.3815 (0.0182)1.5434 (0.0183)1.3815 (0.0001)
−1.5 1.2080 (0.0014)1.2609 (0.0199)1.4011 (0.0004)
−1 1.6234 (0.0108)0.9900 (0.0103)1.2819 (0.0065)
−0.5 1.1915 (0.0111)1.3187 (0.0195)1.4091 (0.0032)
0.5 1.3909 (0.0231)1.4560 (0.0183)1.3324 (0.0032)
1 1.3815 (0.0183)1.3815 (0.0183)1.3815 (0.0183)
1.5 0.5155 (0.05190.5315 (0.0183)1.5109 (0,0057)
2 1.4193 (0.0131)0.9547 (0.1041)1.4045 (0.0004)
Table 10. Bayesian estimators and PR (in brackets) under the three loss functions.
Table 10. Bayesian estimators and PR (in brackets) under the three loss functions.
N = 5000n1 = 10n2 = 50n3 = 200
m1040160
G Q | γ = 1 1.4994 (0.00710)1.4888 (0.0060)1.5138 (0.0011)
E n t r o p y | p = 0.5 1.4830 (0.00413)1.4899 (0.0023)1.5121 (0.0012)
L i n e x | r = 1.5 0.5155 (0.05190)0.5315 (0.0183)1.5109 (0.0057)
Table 11. Pitman comparison of the estimators.
Table 11. Pitman comparison of the estimators.
N = 5000n = 10n = 50n = 200
m840160
G Q | γ = 1 0.6780.6540.674
E n t r o p y | p = 0.5 0.5340.5870.632
L i n e x | r = 1.5 0.5870.57890.734
Table 12. Empirical levels and corresponding theoretical levels (ε = 0.01, 0.02, 0.05, 0.1) and N = 15,000.
Table 12. Empirical levels and corresponding theoretical levels (ε = 0.01, 0.02, 0.05, 0.1) and N = 15,000.
n↓ & ε→ε1 = 0.01ε2 = 0.02ε3 = 0.05ε4 = 0.1
n1 = 250.99410.98290.95220.9031
n2 = 500.99360.98200.95130.9024
n3 = 1500.99220.98120.95100.9012
n5 = 3000.99090.98070.95070.9008
n6 = 6000.99050.98040.95030.9004
Table 13. Empirical levels and corresponding theoretical levels (ε = 0.01, 0.02, 0.05, 0.1) and N = 14,000.
Table 13. Empirical levels and corresponding theoretical levels (ε = 0.01, 0.02, 0.05, 0.1) and N = 14,000.
n ↓&ε→ε1 = 0.01ε2 = 0.02ε3 = 0.05ε4 = 0.1
n 1 = 25 0.99300.98290.95330.9025
n 2 = 50 0.99250.98180.95240.9014
n 3 = 150 0.99120.98120.95160.9009
n 5 = 300 0.99060.98070.95080.9004
n 6 = 600 0.99040.98030.95020.9001
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Emam, W.; Tashkandy, Y.; Goual, H.; Hamida, T.; Hiba, A.; Ali, M.M.; Yousof, H.M.; Ibrahim, M. A New One-Parameter Distribution for Right Censored Bayesian and Non-Bayesian Distributional Validation under Various Estimation Methods. Mathematics 2023, 11, 897. https://doi.org/10.3390/math11040897

AMA Style

Emam W, Tashkandy Y, Goual H, Hamida T, Hiba A, Ali MM, Yousof HM, Ibrahim M. A New One-Parameter Distribution for Right Censored Bayesian and Non-Bayesian Distributional Validation under Various Estimation Methods. Mathematics. 2023; 11(4):897. https://doi.org/10.3390/math11040897

Chicago/Turabian Style

Emam, Walid, Yusra Tashkandy, Hafida Goual, Talhi Hamida, Aiachi Hiba, M. Masoom Ali, Haitham M. Yousof, and Mohamed Ibrahim. 2023. "A New One-Parameter Distribution for Right Censored Bayesian and Non-Bayesian Distributional Validation under Various Estimation Methods" Mathematics 11, no. 4: 897. https://doi.org/10.3390/math11040897

APA Style

Emam, W., Tashkandy, Y., Goual, H., Hamida, T., Hiba, A., Ali, M. M., Yousof, H. M., & Ibrahim, M. (2023). A New One-Parameter Distribution for Right Censored Bayesian and Non-Bayesian Distributional Validation under Various Estimation Methods. Mathematics, 11(4), 897. https://doi.org/10.3390/math11040897

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop