1. Introduction
Exponential distribution is one of the often-encountered continuous probability distributions. It is frequently used to simulate the time between events. The mean and expected value of the exponential probability distribution, as well as its theoretical interpretation, will now be given. The exponential probability distribution, or Poisson point process, is the probability distribution of the interval between events in a process where events happen continuously and independently at a set average rate. In statistics and probability theory, this probability distribution is exploited. It is a unique instance of the gamma distribution. It is the continuous counterpart of the geometric distribution and possesses the key quality of being memoryless. In addition to the analysis of Poisson point processes, it is used in numerous other contexts. Exponential distribution is not the same as the class of exponential families of distributions, a major class of probability distributions that also includes the normal distribution, binomial distribution, gamma distribution, Poisson distribution, and many more probability distributions. One of the most often utilized continuous distributions is exponential probability distribution. It helps in figuring out how long there will be between events. It is used in many disciplines, such as physics, dependability theory, queuing theory, and others. Finding the height of different molecules in a gas at a stable temperature and pressure in a uniform gravitational field, as well as computing the monthly and annual highest values of consistent rainfall and river outflow volumes, are all tasks that can be accomplished using the exponential probability distribution. In statistical inference and dependability, selecting a suitable core model for new data analysis is becoming increasingly crucial. The results could be significantly affected by even a little departure from the core model. The difficulty of this work is increased by censorship. The chi-square test type is the one used most frequently to determine the goodness-of-fit. Numerous modifications to chi-square tests have been proposed by various researchers.
One of the aims of this study is to present a goodness-of-fit test for our parametric model, which is often used in survival analysis, social sciences, engineering, and dependability, in complete data scenarios, and in the presence of right censoring. For the one parameter Poisson-exponential (OPPE) model, we present the explicit forms of the quadratic test statistics (Nikulin–Rao–Robson test, Bagdonavičius and Nikulin test statistic, see Bagdonavičius et al. [
1], Bagdonavičius and Nikulin [
2], Bagdonavičius and Nikulin [
3], Nikulin [
4], Nikulin [
5], Nikulin [
6], and Rao and Robson [
7]). Then, we use actual data to apply both tests. This study demonstrates how the Bayesian technique may enhance both the mean time between failure of the OPPE distribution and the maximum likelihood estimate of the parameter. By using a suitable loss function, we will also illustrate the utility of this adjustment. The OPPE distribution is distinguished by its ease of mathematical and scientific handling in statistical and mathematical modeling. As is common for many probability distribution researchers, we do not approach this new distribution in the normal method in this work. For instance, we do not focus as much on the conventional study of the new distribution, not because it is unimportant, but rather because we are more concerned with the practical applications of the mathematical and statistical modeling, as well as a significant portion of the distribution’s verification using censored data. To demonstrate the significance and flexibility of the new distribution and its wide range of applications in statistical and mathematical modeling, as well as the handling of controlled data, we omit several theoretical mathematical features, many algebraic derivations, and related theories. To be helpful to scholars in the field and assist them in providing additional similar and possibly more flexible distributions, we must briefly discuss the emergence of the new distribution in this context as well as how it was created and formed. The OPPE distribution’s cumulative distribution function (CDF) can be written as
where
x > 0 and
δ > 0,
and
The corresponding probability density function (PDF) can then be expressed as
for all
and
. Many scholars have been interested in the exponential distribution and in providing new flexible extensions of it, and they have also been interested in the applications of these new extensions in various fields of science, such as engineering, insurance, medicine, reliability, actuarial science, and others. In this work, we propose to develop a modified chi-square type fit test for the OPPE model in the scenario when the parameter is unknown, and the data are complete. This test is based on the Nikulin–Rao–Robson (NKRR) statistic, which Nikulin [
3], Nikulin [
4], Nikulin [
5], Nikulin [
6], and Rao and Robson [
7] separately proposed. This
statistic is a logical refinement of the conventional chi-square test based on the maximum likelihood estimator on the original data. The original data were used in full for this statistic, which is a straightforward enough statistic to estimate the null hypothesis’s parameter. Second, this statistic entirely employed the Fisher data, which was to assess the information relative to a parameter contained in a model selection, based on the MLE. This provides a more unbiased evaluation of the chosen model of fit. Next, in the scenario when the parameters are unknown and the data are right-censored, we create a new goodness-of-fit test for this model. This modified chi-square
test adjusts the NKRR statistic to account for both censoring and the unknown parameter. This study is a very significant numerical simulation study that we are conducting to demonstrate the invariance of the test statistic distribution on the one hand, and to test the null hypothesis
that a sample originates from an OPPE model on the other. We compute
the NKRR statistic as well as the Bagdonavičius and Nikulin statistics of various simulated samples, respectively. As a result, we can state that the suggested tests can, respectively, fit complete and censored data from the OPPE distribution. We conclude our contribution in this work with applications using real data, with the aim of demonstrating the applicability of the OPPE model across many scientific disciplines.
It is worth noting that the NKRR basic test, which is supported by complete data, is the most popular test in the last ten years. This is because it fits the complete truth data, and this is the case in most practical and applied cases in various fields. What distinguishes this statistical test also is the availability of ready-made statistical packages on the R program, whether for simulation or applications on actual data. However, the practical and experimental reality in many fields (such as the medical, chemical, engineering, etc.) necessitates that researchers deal with practical experiments that produce controlled censored data. This type of data, of course, needs certain tests dedicated to statistical dealing with it in the problem of distributional validation. The NKRR basic test is not the optimal choice in these cases. Hence, and based on this dilemma, was the primary and most important motive that prompted many researchers to think about introducing a new statistical test that fits the censored data. This new NKRR test, of course, is a modified test from the NKRR original test. According to the nature of the procedures of the two tests, both tests are not suitable for working with a type of data. One weakness of the original test, for instance, is that it is only applicable to entire data and cannot be used to deal with censored data. Therefore, this test may be a strong contender for doing statistical hypothesis tests if we are working with complete data. The modified test also has the drawback of only being appropriate for use with censored data; it is not designed for complete data and can only be used for censored data. Because it is designed for this type of data, this test will surely be a strong contender for statistical hypothesis tests if we are working with censored data.
2. Construction of NKRR Statistic for the OPPE Model
The NKRR statistic is a well-known variant of the traditional chi-squared tests in the situation of complete data (for more information, see Nikulin [
3], Nikulin [
4], Nikulin [
5], Nikulin [
6], and Rao and Robson [
7]). The most common test to check whether a mathematical model is suitable for the data from observations is the chi-square statistic of Pearson. When the model’s parameters are unknown or data are censored, these tests, however, cannot be used. Natural adaptations of the Pearson statistic for the entire set of data were reported by Nikulin [
3], Nikulin [
4], Nikulin [
5], Nikulin [
6], and Rao and Robson [
7] and are known as NKRRs. The chi-square distribution is used in this statistical test, which is a logical extension of the Pearson statistic. When the censoring is included in addition to the unknown parameter, the classical test is insufficient to support the null hypothesis. Bagdonavičius et al. [
1], Bagdonavičius and Nikulin [
2], Bagdonavičius and Nikulin [
3], Nikulin [
4], Nikulin [
5], Nikulin [
6], and Rao and Robson [
7] suggested changing the NKRR statistic to take into account random right censoring. For the OPPE model in the current study, we recommend creating a modified chi-square test. The following NKRR statistic
was developed by Nikulin [
3], Nikulin [
4], Nikulin [
5], Nikulin [
6], and Rao and Robson [
7] to test the hypothesis.
where in our case
, is a sample belong to a parametric family
, and where
where
and
is the information (Inf-Mx) for the grouped data
with
and then
where
is the estimated Fisher Inf-Mx and
stands for the maximum likelihood estimator of the parameter vector. The
statistic has
degrees of freedom and follows the
distribution. Consider a set of observations
that are collected in
(these
subintervals are mutually disjoint:
The intervals
’s limits for
are determined as follows
and
The vector of frequencies is generated by grouping the data into
intervals,
, where
In order to determine if the utilized data are distributed in accordance with the OPPE model, in the situation of an unknown parameter
, we design an NKRR test statistic in this study as a modified goodness-of-fit test. We use the estimated Fisher Inf-Mx to provide all the components of the
statistic of our model after computing the maximum likelihood estimator
of the unknown OPPE distribution’s parameter on the data set. For more applications under other different data sets, see Ibrahim et al. [
8] and Yadav et al. [
9].
3. Estimation and Inference
This section, via two subsections, discusses Bayesian and non-Bayesian estimating methods. Six non-Bayesian estimation methods are considered, including the maximum likelihood estimation (MLE), the Cramér–von Mises estimation (CVME), the ordinary least square estimation (OLSQE), the weighted-least square estimation (WLSQE), moment method, and the Kolmogorov estimation (KE).
3.1. Classical Estimation Methods
3.1.1. Maximum Likelihood Method
Using some observable data, maximum likelihood estimation (MLE), a statistical approach, can estimate the unknown parameter of a probability distribution. Consider the n-sample
and a fixed constant
, we assume that the m-sample
is generated from the OPPE distribution. The likelihood function of this sample is
where
in both (1) and (2) we have
in which
and
To obtain the maximum likelihood estimate (MLE) of
, we have the log-likelihood (
) function
The maximum likelihood estimator
of the parameter
is the solution of the following non-linear eduation
where
and
3.1.2. The CVME Method
The CVME of the parameter
is obtained via minimizing the following expression with respect to
, where
and where
and where
Then, CVME of the parameter
is obtained by solving the following non-linear equation
where
is the first derivatives of the CDF of OPPE distribution with respect to
.
3.1.3. The OLSQ Method
Let
denote the CDF of the OPPE model and let
be the
ordered RS. The OLSQE (
) is obtained upon minimizing
where
. Then, we have
The LSE is obtained via solving the following non-linear equation
where
is defined above.
3.1.4. The WLSQE Method
The WLSQE is obtained by minimizing the function
with respect to
where
The WLSQEs are obtained by solving
where
is defined above.
3.1.5. Method of Moments
The method of moments is a technique used in statistics to estimate population parameters. To derive higher moments like skewness and kurtosis, the same method is applied. It begins by describing the population moments as functions of the important parameters (i.e., the anticipated powers of the random variable under discussion). The sample moments are then set to be equal to those. There are exactly as many of these equations as there are parameters that need to be estimated. Then, the equations are solved for the relevant parameters. Estimates of those parameters are used in the solutions. The moment estimation of the one-parameter of the OPPE distribution can be obtained by equating the first theoretical moment of (2) with the corresponding sample moments as follows
where
can be evaluated from (2).
3.1.6. KE Method
The Kolmogorov estimate (KE)
of
is obtained by minimizing the function
The KE of is obtained by comparing and and selecting the max one. However, for , we are minimizing the whole function For more detail about the KE method.
3.2. Simulations and Assessment
For comparing the classical methods, some MCMC simulation studies are performed. The results are presented in:
Table 1 (
δ = 0.5|
n = 50, 100, 200, 300, and 500);
Table 2 (
δ = 0.9|
n = 50, 100, 200, 300, and 500) and
Table 3 (
δ = 1.5|
n = 50, 100, 200, 300, and 500). The numerical assessments are performed depending on the mean squared errors (MSEs). First, we generate N = 1000 samples of the OPPE model. Based on
Table 1,
Table 2 and
Table 3, it is noted that the performance of all estimation methods improves when
. Despite the variety and abundance of the other classic methods, as demonstrated in
Table 1,
Table 2 and
Table 3, the MLE approach is still the most efficient and reliable of the surviving classic methods. The MLE approach is generally mentioned as being advised for statistical modeling and applications. This evaluation is mostly based on a thorough simulation research, as displayed in
Table 1,
Table 2 and
Table 3. Since the MSE for the MLE is the smallest for all
n = 50, 100, 200, 300, and 500, this section uses simulation studies to evaluate rather than compare various estimating procedures. However, this does not exclude the use of simulation to compare various estimating approaches. Nevertheless, actual data are commonly used to assess different estimating approaches; for this reason, we will discuss a few examples specifically for this role. There are two more applications to compare the competing models.
3.3. Applications for Comparing Methods
Here, we will be very interested in comparing the different estimation methods, but this time through real data. The first data set called the failure time data or relief times (in minutes) of patients receiving an analgesic (see Gross and Clark [
10]). The second data set is called the survival times (in days) of 72 guinea pigs infected with virulent tubercle bacilli, see Bjerkedal [
11].
We discuss the skewness–kurtosis plot (or the Cullen and Frey plot) in these applications for examining initial fits of theoretical distributions such as normal, uniform, exponential, logistic, beta, lognormal, and Weibull. Plotting and bootstrapping are both applied for greater accuracy. The scattergram plots, the “nonparametric Kernel density estimation (NKDE)” method for examining the initial shape of the insurance claims density, the “Quantile-Quantile (Q-Q)” plot for examining the “normality” of the current data, the “total time in test (TTT)” plot for examining the initial shape of the empirical hazard rate function (HRF), and the “box plot” for identifying the extreme data were also presented.
Figure 1 gives the box plot for the failure time data (first row, the left panel), Q-Q plot for the failure time data (first row, the right panel), TTT plot for the failure time data (second row, the left panel), nonparametric Kernel density estimation plot for the failure time data (second row, the right panel), the Cullen and Frey plot for the failure time data (third row, the left panel), and scattergrams (third row, the right panel) for the failure time data.
Based on
Figure 1 (first row), the relief data have only one extreme observation, based on
Figure 1 (second row, the left panel), the HRF of the relief times is “monotonically increasing HRF”, based on
Figure 1 (second row, the right panel), nonparametric Kernel density estimation is bimodal and right skewed with asymmetric shape, based on
Figure 1 (third row, the left panel), the relief times data do not follow any of the theoretical distributions such as the normal, uniform, exponential, logistic, beta, lognormal, and Weibull.
Figure 2 gives the box plot for the survival times (first row, the left panel), Q-Q plot for the survival times (first row, the right panel), TTT plot for the survival times (second row, the left panel), nonparametric Kernel density estimation plot (second row, the right panel), the Cullen and Frey plot (third row, the left panel), and scattergrams (third row, the right panel) for the survival times data. The survival data in
Figure 2 (first row) has four extreme observations, the HRF of the survival times is “monotonically increasing HRF,” the nonparametric Kernel density estimation is bimodal and right skewed with an asymmetric shape, and the survival times data in
Figure 2 (third row, left panel) do not follow any of the theoretical distributions, such as the normal or gamma distributions.
A statistical test called the Anderson–Darling test (ADT) is used to determine if a sample of data is taken from a particular probability distribution. The test is based on the assumption that the distribution under the test has no parameters that need to be estimated, in which case the test and its set of critical values are distribution-free. The test is most frequently applied when a family of distributions is being examined, in which case it is necessary to estimate the family’s parameters and take this into consideration when modifying the test statistic or its critical values. It is one of the most effective statistical strategies for detecting most deviations from normal distribution when used to determine whether a normal distribution adequately describes a collection of data.
The Cramér–von Mises criterion (CVMC) in statistics is a criterion used to compare two empirical distributions or to determine how well a cumulative distribution function fits an empirical distribution function. In other techniques, such as minimum distance, the CVMC and ADT tests are used in comparing methods.
Table 4 gives the application results (CVMC and ADT) for comparing methods under the relief data.
Table 5 lists the application CVMC and ADT for comparing methods under the survival data. Based on
Table 4, it is seen that the moment method is the best with CVMC = 0.08439 and ADT = 0.499304, then the ML method with W CVMC = 0.085876 and ADT = 0.508074. Based on
Table 4, it is seen that the moment method is the best with CVMC = 0.070712 and ADT = 0.463826, then the ML method with CVMC = 0.071247 and ADT = 0.465521.
3.4. Bayesian Analysis under Different Loss Functions
In this section, we will review some statistical aspects related to Bayesian estimations. Three types of Bayesian loss functions have been used. Additionally, I made a lot of useful comparisons using the new distribution. Many Bayesian statistical algorithms can be utilized, as per Muñoz-Gil et al. [
12], and within their particular context, they can aid researchers in mathematics and statistical modeling tasks, particularly when applying Bayes’ theory. Muñoz-Gil et al. [
12] offered the proposed representation of four full algorithms with their codes and implementations for this specific purpose, allowing the reader to follow the new algorithms and present more algorithms based on them.
3.4.1. Prior and Posterior Distributions
As prior distributions, we assume the parameter
has an informative distribution as a prior, which is a Gamma distribution:
The posterior distribution of
is
where
is the normalizing constant. In the following, we use the three loss functions, namely the generalized quadratic (GQ), the entropy, and the Linex loss functions to obtain the Bayesian estimators. In the following formulas, we consider
γ,
P, and
r as integers.
3.4.2. Bayesian Estimators and Their Posterior Risk
The Bayesian estimator under the generalized quadratic loss function
is
where
and
Under the entropy loss function, we obtain the following estimator
The corresponding posterior risk is
finally, under the Linex loss function, the Bayesian estimator
and the corresponding posterior risk is
Since it is unlikely possible to obtain all these estimators analytically, we suggest the use of the MCMC procedures to evaluate them. In this work, several Bayesian algorithms were developed recently for the analysis of tracer-diffusion single-particle tracking data. Following Muñoz-Gil et al. [
12], many Bayesians statistical algorithms can be used, which within their specific framework, can help researchers in mathematical and statistical modeling operations, especially when using Bayes’ theory. For this specific purpose, Muñoz-Gil et al. [
12] presented the proposed visualization of four complete algorithms with their codes and applications, and the reader can track the new algorithms and other algorithms can be presented based on them.
3.4.3. Comparing the Likelihood Estimation and the Bayesian Estimation Using Pitman’s Closeness Criterion
We are going to compare the performance of the proposed Bayesian estimators with the MLEs, for that purpose, we perform a MCMC simulation method with
δ = 1.5 and
a = 3,
b = 2. We generate
N = 1000 type II censored samples following the OPPE model, we use different sample sizes
n = 30, 100, 200 while
m = 10, 40, 160, respectively, and we obtain the following results.
Table 6 contains the values of the estimators produced by the function BB algorithm. Here, we observe that, especially as sample size n is raised, the estimated values are fairly close to the parameter’s genuine values.
Table 7,
Table 8 and
Table 9 present the Bayesian estimators and PR (in brackets) under the generalized quadratic loss function, the entropy loss function, and the Linex loss function, respectively.
Table 10 lists the Bayesian estimators and PR (in brackets) for each of the three loss functions.
In
Table 7, the estimation under the generalized quadratic loss function, we remark that the value
gives the best posterior risk. Additionally, we obtain the smallest suitable posterior risk when
n is long. In the estimation under the entropy loss function, we obtain
Table 8 where we can notice that the value
when
provides the best posterior risk. We can notice clearly that the value
provides the best PR. In conclusion, it is evident from a brief comparison of the three loss functions that the quadratic loss function yields the best outcomes; the results are further illustrated in
Table 10. The best Bayesian estimators and maximum likelihood estimators should be compared, as we suggest. We employ the Pitman closeness criterion for this purpose (more information can be found in Pitman [
13], Fuller [
14], and Jozani [
15]).
Definition 1. An estimator of a parameter dominates another estimator in the sense of Pitman’s closeness criterion if, for all ,
We show the Pitman probability’ values in Table 11 so that we may contrast the Bayesian estimators with the MLE estimator when using the three loss functions when γ = −1, p = −0.5, and r = −1.5. Then, we notice that, according to this criterion, the Bayesian estimators of the parameter is better than the MLE. Additionally, the Linex loss function has the best values in comparison with the other two loss functions with the probability 0.734, δ(n=200,m=160). 4. Distributional Validation
We utilize the statistic type test based on a version of the NKRR statistic suggested by Bagdonavičius et al. [
1], Bagdonavičius and Nikulin [
2], Bagdonavičius and Nikulin [
3], Nikulin [
4], Nikulin [
5], Nikulin [
6], and Rao and Robson [
7] to confirm the sufficiency of the OPPE model when the parameter is unknown, and the data are censored. The failure rate
follows an OPPE distribution, hence we modify this test for an OPPE model. Let’= us think about the null hypothesis:
The OPPE distribution’s survival function (SrF) and cumulative hazard function are as follows:
where
and
For all
, we have a constant value of
under this choice of intervals. Since the inverse hazard function of the OPPE distribution lacks an explicit form, intervals can be estimated iteratively. Let us divide a finite time period
into
smaller periods
and
If
is the
ith element in the ordered statistics
, and if
is the inverse of the cumulative hazard function
, then the estimated value of
where
and
are random data functions, such as the
selected intervals have equal expected numbers of failures
. The test for hypothesis
can be based on the statistic
where
and
and
reflect the total number of failures that have been noticed throughout these times. Due to Bagdonavičius et al. [
1], Bagdonavičius and Nikulin [
2], Bagdonavičius and Nikulin [
3], Nikulin [
4], Nikulin [
5], Nikulin [
6], and Rao and Robson [
7], the test statistic is expressed as:
where
and many other details are given in Bagdonavičius et al. [
1], Bagdonavičius and Nikulin [
2], Bagdonavičius and Nikulin [
3], Nikulin [
4], Nikulin [
5], Nikulin [
6], and Rao and Robson [
7]. We compute each component of the
statistic for the OPPE model. The statistic
has a chi-square limit distribution, and its degree of freedom is
where
. The estimated significance threshold is rejected if
(where
is the quantile of chi-square with
), then the approximate significance level
is rejected Hypothesis, where
is the Fisher information.
4.1. Uncensored Simulation Study under the NKRR Statistics
We conducted an extensive study using numerical simulation to confirm the arguments of this work. Therefore, we generated the N statistics of 15,000 simulated samples with sizes of
n = 25,
n = 50,
n = 150,
n = 350, and
n = 600, in order to test the null hypothesis
that the sample belongs to the OPPE model. We determine the average of the non-rejection numbers of the null hypothesis
for various theoretical levels (
ε = 0.01, 0.02, 0.05, 0.1). The corresponding empirical and theoretical levels are represented in
Table 12. It can be seen that the calculated empirical level value is very close to its corresponding theoretical level value. Therefore, we conclude that the recommended test is very suitable for the OPPE distribution.
4.2. Uncensored Applications under the NKRR Statistics
4.2.1. Strengths of Glass Fibers
In this example, 100 carbon fiber fracture stresses (in Gba) are included in this data set given by Nicholase and Padgett [
16]. Using the BB algorithm, we can get the MLE value of the parameter
δ, assuming that our OPPE model can fit the strength data of 1.5 cm glass fiber:
. We can compute and provide the Fisher Inf-Mx as follows using the value:
The critical values for the NKRR statistical test were: and , so the OPPE distribution can effectively simulate and model the 1.5 cm glass fiber data.
4.2.2. Heat Exchanger Tube Crack
The crack data, which includes examinations carried out at eight chosen intervals until fractures emerged in 167 similar turbine parts, was acquired from the book by Meeker and Escobar [
17].
We test the null hypothesis that these data are modified by our OPPE distribution using previously acquired NKRR statistics. We calculate the MLE
using R programming and the BB method (see Ravi (2009)). The estimated Fisher Inf-Mx at that time is:
So, we arrive with as the answer. The crucial value for significance level and . This model’s NKRR statistic () is less than the essential value, allowing us to conclude that the data correctly fit the OPPE model.
4.3. Censored Simulation Study under the NKRR Statistics
In the censored simulation study under the NKRR statistics
, it is expected that the produced sample
is censored at
with
. We determine the average value of the non-rejection numbers of the null hypothesis for various theoretical levels
, where
≤
. The matching theoretical and empirical levels are presented in
Table 13, demonstrating how closely the determined empirical level value resembles the associated theoretical level value. Therefore, we draw the conclusion that the custom test is ideally suited to the OPPE model. These findings lead us to the conclusion that the theoretical level of the chi-square distribution on degrees of freedom corresponds to the empirical significance level of the
statistics at which it is statistically significant. It may be inferred from this that the suggested test can accurately fit the censored data from the OPPE distribution. Returning to the defects of the NKRR basic test, the statistical literature contains many statistical hypothesis tests that fit the complete data, and for this reason, the original test has many competing tests in this field, and the fact is that the NKRR basic test actually has many alternatives. As for the modified NKRR test, it is considered an individual statistical test of its kind in the field of statistical tests due to its importance and the nature of the data that it can deal with, which are the data subject to censorship.
4.4. Censored Applications under the NKRR Statistics
4.4.1. Censored Lung Cancer Data Set
The lung cancer data provided by Loprinzi et al. [
18] from the North Central Cancer Treatment Group investigated the survival in patients with advanced lung cancer (
n = 228 and censored items = 63), and their performance scores rate how well the patient can do typical daily activities. If we assume that the data are distributed according to the OPPE distribution, we can estimate the vector parameter
by applying the maximum likelihood estimation approach as:
. As a number of classes, we employ
. The following is how the test statistic
items are presented:
| 92.138 | 171.694 | 216.037 | 283.012 | 355.086 | 456.277 | 685.261 | 1022.3174 |
| 29 | 30 | 35 | 31 | 32 | 25 | 28 | 18 |
| 6.55438 | 6.55438 | 6.55438 | 6.55438 | 6.55438 | 6.55438 | 6.55438 | 6.55438 |
The estimated matrix
and the estimated information (E-Inf)
via Fisher are as follows:
| −0.6827 | −0.4368 | 0.7015 | 0.7684 | 0.5094 | 0.3007 | 0.8769 | −0.2884 |
and
The chi-squared test has a critical value of We find that the estimated statistic for the suggested test is using the earlier findings. We may state that our hypothesis is accepted because the tabulated value of the statistic is higher than the computed value. This leads us to the conclusion that there is a 5% chance that the lung cancer data will deviate from the OPPE distribution.
4.4.2. Censored Capacitor Data Reliability Data Set
A set of data for basic reliability assessments data on glass capacitor longevity as a function of voltage and operating temperature from a factorial experiment (see Meeker and Escobar [
17]). Each combination of temperature and voltage has eight capacitors. Testing was stopped after the fourth failure at each combination,
n = 64 and censored items = 32. The parameter vector’s maximum likelihood estimator
, assuming that the data are distributed according to the OPPE distribution, is:
. We pick a few classes with
. The components of the statistical test
are as follows:
| 346.1573 | 469.502 | 587.113 | 679.017 | 1078.834 | 1089.109 | 1102.167 | 1106.444 |
| 11 | 15 | 6 | 10 | 6 | 5 | 6 | 5 |
| 6.91042 | 6.91042 | 6.91042 | 6.91042 | 6.91042 | 6.91042 | 6.91042 | 6.91042 |
The estimated matrix
and Fisher’s E-Inf
are:
| 0.48965 | −0.73465 | −0.40012 | 0.29784 | −0.29467 | −0.95113 | 0.60378 | 0.28564 |
and
The value of the statistical test is then assessed. is the crucial value. We may conclude that the OPPE model is used to update the life statistics for glass capacitors.
5. Conclusions
The one parameter Poisson-exponential (OPPE) model, a new adaptable variation of the exponentiated exponential model, is introduced and studied in this article. Six well-known estimation techniques are investigated, discussed, and used: Cramer–von Mises, ordinary least square, L-moments, maximum likelihood, Kolmogorov, and weighted-least square. The Bayesian estimation under the squared error loss function was also devised and investigated. The use of two real data and two simulated data allows for a complete evaluation of all available techniques. The utility and adaptability of the OPPE distribution are illustrated using two real data sets. The new model outperforms many other competing models in modeling relief times and survival times, according to the Akaike Information Criterion, Consistent Akaike Information Criterion, Hannan–Quinn Information Criterion, Bayesian Information Criterion, Cramér–von Mises, and Anderson–Darling statistics. However, there are more results the reader can read by examining the paper and the applications, the following results can be specifically highlighted:
The maximum likelihood method is still the most effective and trustworthy of the remaining classic approaches. Both the Bayesian technique and the Maximum Likelihood method are recommended for statistical modeling and applications.
Under different loss functions, the Bayesian estimation is provided. Three loss functions, the generalized quadratic, the Linex, and the entropy, are used to produce the Bayesian estimators, and many useful details are provided.
All of the provided estimation methods have been evaluated through simulation tests with particular parameter and controls (these simulation studies are all stated in the paper at the appropriate places).
The BB algorithm for process estimation under censored samples is used to compare the Bayesian technique and the censored maximum likelihood method.
It is shown in detail how the NKRR statistic is created for the OPPE model in the unfiltered instance. The results of a simulation research show that the OPPE model and the NKRR test are a good fit.
A simulation study for evaluating statistics is described, along with the construction of the Bagdonavičius and Nikulin statistic for the novel model under the censored situation. Two real data applications are examined in a censored scenario; the first data are reliability information on capacitors, and the second data are information about lung cancer (medical data). We deduce from these applications that the suggested technique can successfully fit censored data from the OPPE distribution.
For the uncensored strengths of glass fibers data, the critical values for the NKRR statistical test were: and , so the OPPE distribution can effectively simulate and model the uncensored 1.5 cm glass fiber data.
For the uncensored heat exchanger tube crack data, the critical values for the NKRR statistical test were: and , so the OPPE distribution can effectively simulate and model the uncensored heat exchanger tube crack data.
For the censored lung cancer data set: the value of the statistical test , where . Then, we conclude that the OPPE model can be used for modeling the censored lung cancer data set.
For censored capacitor data reliability data set: the value of the statistical test , where . Then, we conclude that the OPPE model can be used for modeling the censored capacitor data reliability data set.
It is worth noting that NKRR basic test, which is supported by complete data, is the most popular test in the last ten years. This is because it fits the complete truth data, and this is the case in most practical and applied cases in various fields. What distinguishes this statistical test also is the availability of ready-made statistical packages on the R program, whether for simulation or applications on actual data.
However, the practical and experimental reality in many fields (such as the medical, chemical, engineering, etc.) necessitates that researchers deal with practical experiments that produce controlled censored data. This type of data, of course, needs certain tests dedicated to statistical dealing with it in the problem of distributional validation. The NKRR basic test is not the optimal choice in these cases. Hence, and based on this dilemma, was the primary and most important motive that prompted many researchers to think about introducing a new statistical test that fits the censored data. This new NKRR test, of course, is a modified test from the NKRR original test.
According to the nature of the procedures of the two tests, both tests are not suitable for working with a type of data. For example, one of the drawbacks of the original test is that it is not suitable for dealing with censored data, it is only for complete data and cannot be applied to censored data. Therefore, if we are dealing with complete data, then this test may be a strong candidate for performing statistical hypothesis tests.
Moreover, a disadvantage of the modified test is that it is only suitable for dealing with censored data, it is not intended for complete data and can only be applied to censored data. Therefore, if we are dealing with censored data, this test will undoubtedly be a strong candidate for statistical hypothesis tests, because it is intended for this type of data.
Returning to the defects of the NKRR basic test, the statistical literature contains many statistical hypothesis tests that fit the complete data, and for this reason, the original test has many competing tests in this field, and the fact is that the NKRR basic test actually has many alternatives.
As for the modified NKRR test, it is considered an individual statistical test of its kind in the field of statistical tests due to its importance and the nature of the data that it can deal with, which are the data subject to censorship.