1. Introduction
In many lifetime experiments, data often seem continuous; however, they are actually discrete observations. This discrepancy drives the development of techniques that transform continuous distributions into discrete versions, offering a suitable match to the observed data. Discrete distributions in statistical modeling are widely encouraged for various compelling reasons. They address limitations in data collection methods, measurement intervals, or the inherent characteristics of observed phenomena, such as electronic device failures or maintenance cycles. Discrete distributions represent data with countable or finite values, such as the number of items tested, people in a queue, coin toss outcomes, or defective products in manufacturing. These distributions are intuitive and easy to interpret, corresponding to datasets with a defined range of possible values.
Moreover, many discrete distributions provide explicit expressions for their probability generating function and/or probability mass function (PMF), simplifying analytical operations and permitting effective moment and probability calculations without requiring integration. They are highly adaptable, making them suitable for modeling diverse real-world phenomena, including ecosystem species distribution, genetic diversity, and network traffic patterns. Their computational efficiency and versatility make them invaluable tools in statistical modeling.
Recently, a variety of discrete models have been developed, with applications in certain fields such as medicine, engineering, reliability studies, and survival analysis. For a deeper understanding and practical applications of discrete distributions, one can explore references Roy and Gupta [
1] and Roy and Ghosh [
2], among others. However, while these studies have advanced discrete reliability theory and other application areas, they primarily focus on continuous formulations or general discretization techniques that do not specifically address the nuances of the Hjorth distribution.
As a result, many scholars have significantly advanced discrete reliability theory, offering innovative approaches and perspectives. Transforming continuous models into their discrete counterparts involves various discretization techniques, which are extensively discussed and documented in academic research. These methods are designed to create discrete distributions that closely reflect their continuous origins. A range of approaches to discretization have been explored in the literature; one may refer to Bracquemond and Gaudoin [
3], Lai [
4], and Chakraborty [
5].
A prominent approach in the development of discrete distributions involves the application of the survival function as a discretization technique. Notable contributions in this area include the derivation of discrete analogs for normal and Rayleigh distributions, presented by Roy [
6,
7], respectively, both employing the survival function methodology. Extending this framework, the discrete formulation of the Burr-II distribution was systematically examined by Al-Huniti and AL-Dayian [
8]. Additionally, Bebbington et al. [
9] provided a comprehensive study on the discrete additive counterpart of the Weibull distribution. Further advancements and examples of discrete transformations across various distributions are detailed by Haj Ahmad and Almetwally [
10] and Chesneau et al. [
11], among others.
This study employs the survival discretization method to transform the continuous Hjorth distribution into its discrete counterpart. Although the Hjorth distribution is highly flexible in modeling various hazard rate shapes in a continuous framework, practical scenarios often involve data recorded in discrete intervals or subject to censoring. Discretizing the Hjorth distribution using the survival function approach not only preserves key statistical features—like percentiles and quantiles—but also facilitates the analysis of data that are subject to constraints such as Type-II censoring. This discrete formulation directly addresses the practical challenges in reliability and survival analysis, where capturing the precise timing of failure events in a discrete manner is crucial for accurate modeling and inference. Unlike other discretization techniques, this method retains the intrinsic properties of the continuous model while providing a framework that is more robust to data irregularities.
Admittedly, the DH model’s computational intensity and reliance on numerical techniques are limitations that we address through rigorous simulation studies and methodological refinements. While the DH model has demonstrated superior performance in our comparative analysis, it may exhibit sensitivity to certain types of data distributions, particularly when the failure rate does not follow the bathtub or monotonic hazard rate structures that the model is designed to accommodate.
Consider a scenario where n items undergo a life-testing experiment, and only the first r failure times, denoted by , are observed. The set is called a Type-II censored sample. The remaining items are censored and known only to have more than failure times.
Various statistical characteristics, including the quantile function and order statistics, are derived, and statistical inference methods are explored. These methods utilize the maximum likelihood estimation approach and the Bayes framework. To assess the estimators, the two estimation techniques for the newly developed discrete distribution are compared. Simulation studies are conducted using numerical methodologies. To the best of our knowledge, the existing literature does not address the use of censored data to analyze discrete data following the Hjorth distribution. Therefore, this study aims to bridge this gap in distribution theory. Two real-world datasets are analyzed using goodness-of-fit tests to evaluate the effectiveness of the proposed model compared to other discrete alternatives. The first example discusses the failure times of some electronic devices, and the second considers ball-bearing failure analysis.
The rest of this paper is organized as follows:
Section 2 introduces the discrete Hjorth distribution, while
Section 3 presents some statistical functions.
Section 4 and
Section 5 calculate statistical inference, including maximum likelihood and Bayesian estimation, respectively. Monte Carlo results are presented and commented on in
Section 6.
Section 7 studies real-world data illustration examples.
Section 8 explores the findings and concluding remarks.
2. Discrete Hjorth Distribution
The Hjorth distribution (by Hjorth [
12]) is a continuous distribution that extends the Rayleigh, exponential, and linear failure rate distributions. Later, this distribution received some attention in reliability analysis and lifetime experiments. Hence, its statistical properties, inferential statistics, and reliability analysis have been studied by Guess et al. [
13]. Yadav et al. [
14] used progressive censoring and estimated the parameters of the Hjorth model in addition to the hazard rate and the reliability functions. Pushkarna et al. [
15] studied some recurrence relations for the Hjorth moment model under progressive censoring. Pandey et al. [
16] used Bayesian inference to estimate Hjorth parameters under a generalized Type-I progressive censoring sample. Korkmaz et al. [
17] presented a new regression model using the Hjorth model. It is noted for its effectiveness in modeling datasets with smaller values rather than larger ones. Elshahhat and Nassar [
18] examined survival analysis of the Hjorth model by adaptive Type-II progressive hybrid censoring.
We are motivated to use the Hjorth model because it has increasing, decreasing, constant, upside-down, unimodal, and bathtub hazard rates. It can also be considered a suitable model for fitting the bimodal, unimodal, U-shaped, and other-shaped data. Hence, it was observed in the literature that the Hjorth model outperforms several well-known lifetime distributions. Its continuous form, however, limits its use for datasets that are inherently discrete. By converting the Hjorth model into a discrete form, a new distribution is derived that accommodates count data while preserving Hjorth’s ability to model tail behavior.
The probability density function of the continuous Hjorth model is expressed as
where
The survival function is written as
in which
and
are positive shape parameters, and
is a scale parameter.
Here,
controls the rate of decay in the exponential component. Hence, higher values of
result in a faster decay of the density. The parameter
influences the shape of the distribution through the polynomial term
. It affects the tail behavior and the curvature of the density function. A larger
tends to increase the weight in the tail, which affects how quickly the probability mass of the distribution decreases as
x increases.
adjusts the baseline level of the polynomial component. It plays a crucial role in determining the heaviness of the tail and the overall curvature near the origin, thereby influencing the skewness and the rate at which the distribution transitions from its peak to the tail. The hazard rate function for the Hjorth model is
This study aims to develop a novel discrete form of the Hjorth model and perform the statistical analysis under Type-II censored data for time and cost constraints. This counterpart, termed the discrete Hjorth (DH) model, is constructed using the survival discretization approach. In this section, the PMF, cumulative distribution function (CDF), and associated properties of DH model are outlined.
Roy [
6,
7] used the survival function to define the PMF for the new discrete distribution, which is presented in the following form:
where
is provided in Equation (
2); hence, the DH PMF is provided by
where
,
, and
.
The DH using the survival discretization method has the next CDF:
The hazard rate function for the DH model is written as
Figure 1 illustrates various shapes of the PMF and HRF for the DH distribution, generated using different parameter selections. The PMF of the DH model is shown to exhibit right-skewed or symmetric patterns, while the HRF displays diverse forms, including decreasing, increasing, and bathtub-shaped patterns. Additionally, the shapes presented in the figure highlight that the DH model can naturally represent forward or backward recurrence time patterns within renewal processes.
In the next section, some statistical properties, functions, and characteristics of the DH model are derived, including quantiles, moments, and ordered statistics.
3. Statistical Functions
This section discusses statistical functions for the DH distribution, including the quantile function, moments, skewness, kurtosis, and order statistics.
3.1. Quantile Function and Moments
The quantile function (say,
) for a discrete distribution is the inverse of its CDF. It is used basically to generate random samples for simulation purposes. To determine the quantile function, we solve for
k in terms of
p:
Rearranging Equation (
8), we obtain
Substituting
into (
9), we have
and its natural logarithm becomes
This equation must be solved numerically for k as an analytical closed-form expression is generally not feasible due to the combination of logarithmic and exponential terms. The complexity of the DH’s CDF makes direct algebraic inversion infeasible for deriving the quantile function, particularly due to the presence of exponential and power terms. Extracting k from the expression for poses significant analytical difficulties and is unlikely to yield an exact solution because of the intricate structure and exponential decay term. As an alternative to an exact analytical form, approximations or numerical techniques can be employed for practical applications. Moments are a crucial statistical tool, offering detailed insights into the shape and properties of a probability distribution. They are widely applied in fields such as quality control, risk assessment, and environmental analysis.
To compute the moments for the DH model, consider a non-negative random variable
l∼
. The
sth moment, say
, can be expressed as follows:
An exact expression for the moment is not obtainable; thus, numerical methods are necessary to evaluate the moment. Specifically, take and several choices from and .
Table 1 illustrates the behavior with respect to the mean (
), variance (
), index of dispersion (
), coefficient of variation (
), skewness (
), and kurtosis (
) for the DH model. As a result, we summarize the following points:
As grows (for fixed and ), the values of , , , and decrease while those for and increase.
As grows (for fixed and ), the values of and increase while those for , , , and decrease.
As grows (for fixed and ), the values of and decrease while those for , , , and increase.
The DH model is well suited for modeling both under- and over-dispersed data as its variance can be smaller than the mean for certain parameter values, while the variance is greater than the mean in other choices regarding parameter values.
The DH model exhibits flexible dispersion characteristics depending on parameter choices.
For reliability modeling, higher dispersion ( > 1) suggests suitability for lifetime data with higher variability.
The positive skewness values indicate that this distribution is right-skewed. Additionally, a skewness value approaching zero suggests the possibility of a symmetric curve for the PMF.
Elevated kurtosis signifies greater tail risk and the likelihood of outliers relative to a normal distribution. Changes in the distribution can be observed by adjusting , , and .
3.2. Ordered Statistics
Consider
as a random sample following the DH model, and assume that
represent the related order statistics; hence, the
order statistics have a CDF at
w, which is
The negative Binomial theorem can be used as a series representation for the
, which can be written as
The
ith order statistic under the DH model has PMF that can be expressed as follows
Hence, the
moments of
can be written as follows:
4. Maximum Likelihood Estimation
In this section, we estimate the undetermined parameters , , and for the DH model by applying the maximum likelihood estimation (MLE) approach. The MLE estimates are obtained based on the Type-II censoring scheme, and, by using the missing information principle, the variance–covariance matrix (VCM) of , , and is obtained. This matrix is then utilized to construct asymptotic confidence intervals for , , and . Let represent a Type-II censored sampling from DH model.
Referring to the PMF in Equation (
5) and the CDF in Equation (7), the likelihood and the log-likelihood functions are explored, respectively, as
and
The parameters’ MLEs are obtained by deriving the likelihood function (
16) partially for each parameter such as
and
where
and
The obtained system of nonlinear Equations (
17)–(
19) is solved by using numerical techniques to determine the parameter estimates. These calculations can be complex and computationally intensive, especially when dealing with sums and exponentials. For actual applications, numerical methods or software packages capable of symbolic differentiation may be beneficial for accurately computing these derivatives, especially when optimizing parameters in practical scenarios. Various numerical techniques have been explored in the literature; in this study, the Newton–Raphson method is employed. This method was selected primarily due to its quadratic convergence rate that enables rapid refinement of estimates when the initial guess is reasonably close to the true solution. The results of the analysis are discussed in
Section 6.
To construct the VCM of the DH model parameters , we need to use the Fisher Information Matrix (FIM), which is derived from the second-order partial derivatives of the log-likelihood function . The FIM (say, ) is a symmetric matrix where each entry corresponds to a second-order partial derivative. The inverse of the FIM provides the VCM (denoted by ), which is then used to build asymptotic confidence intervals for the parameters.
The FIM and VCM (at
) are expressed, respectively, as
and
The second-order partial derivatives of the log-likelihood function
ℓ are developed as follows:
and
where
and
and
Hence, the asymptotic confidence interval for is provided by , where is the critical value from the standard normal distribution for a given th confidence level.
5. Bayesian Inference
This section deals with the Bayesian estimation, which is applied to estimate the unknown parameters of the DH model. This method treats the parameters as random variables following a certain model, referred to as the prior distribution. Since prior information is often unavailable, it becomes necessary to choose a suitable prior.
A joint conjugate prior distribution is selected for parameters
,
, and
, with each parameter assumed to follow a gamma distribution. Consequently,
,
, and
, where
and
(for
) represent non-negative hyper-parameters of the specified distributions. Therefore, the prior distributions for
,
, and
are defined as
and
Therefore, the joint prior function for
,
, and
is
The joint posterior (say,
) of
,
, and
with data availability is
where
is the normalizing factor.
The parameter estimation for the DH model has been analyzed using the squared error (SE) loss function. Later, to evaluate the efficiency of the estimation methods and to examine the impact of parameter values on these techniques, a simulation analysis is conducted to assess the efficiency of the estimators based on several metrics, namely average point estimate, mean absolute bias, mean square error, average interval length, and coverage probability.
Under the SE loss function, Bayesian estimation of a parameter
(for example) is defined as the expected value for the joint posterior distribution as
It is necessary to employ numerical approaches for calculating the triple integration mentioned in Equation (
31). To achieve this, we adopted the Markov Chain Monte Carlo (MCMC) approach. A suitable
code was developed to execute this process.
Applying the survival discretization method leads to the development of the DH model, whose PMF is defined in Equation (
5). The joint posterior density under a Type-II censored sample is expressed as follows:
Bayesian analysis for the parameters
,
, and
under the SE loss function is beyond developing their conditional posterior functions, respectively, as
and
The Bayes estimators of
,
, and
, based on Equations (
33)–(
35), respectively, cannot follow any known statistical distribution. To handle this problem, the Metropolis–Hastings (M–H) sampler, the most important term of the MCMC family of Bayes calculation techniques, is recommended to evaluate the Bayes point estimates and their Bayesian credible intervals (BCIs). It is useful to remember that the BCI represents a range of values where an unknown parameter is likely to reside with a specified probability based on the obtained data and prior beliefs regarding the parameters.
Now, the implementation of the MCMC approach through the M–H sampler is outlined as listed below:
- Step 1:
Start by assuming a basic value
- Step 2:
Set
- Step 3:
From (
33), create
from
, and then follow the next steps (a)–(d):
- (a)
Find .
- (b)
Obtain .
- (c)
Generate from .
- (d)
If , adopt the proposal and set ; otherwise, set .
- Step 4:
Redo Step 3 for and .
- Step 5:
Set .
- Step 6:
Obtain () for by redoing Steps 2 to 4 N times.
- Step 7:
Discard the first iterations (say,
) as burn-in, and obtain the Bayes estimates of
,
, and
(say,
) as
where
- Step 8:
Sort
in ascending order and obtain the
BCIs for
as
6. Numerical Comparisons
In this part, to test the efficiency of the acquired estimators of DH’s parameters , , and , different comparisons via Monte Carlo experiments are performed. After that, some comments on the outcomes of the simulation are provided. The numerical part was performed using the R (64) software package.
6.1. Simulation Scenarios
We now suggest the following steps to gather a Type-II censored dataset from the proposed model:
- Step 1:
Fix the number of replications as 2000.
- Step 2:
Fix the values of ; that is, (a) Set-1:(0.5,0.8,1) and (b) Set-2:(1,1.5,2).
- Step 3:
Set sample size n as .
- Step 4:
Obtain for as an independent observation from a uniform distribution .
- Step 5:
Obtain pseudo-random values of size
n from the DH distribution as
- Step 6:
Sort the outputs in Step 5; then, for a failure percentage (FP%), determine the value of r such as , 80, and 100%. It is important to note that, when FP%, censored (incomplete) sampling reverts to completed sampling.
- Step 7:
Fix the informative prior sets of and for to develop the Bayes point and interval estimations, such as
- (i)
For Set-1:
Prior-A: and for ;
Prior-B: and for .
- (ii)
For Set-2:
Prior-A: and for ;
Prior-B: and for .
- Step 8:
Compute the average point estimate (APE) of
(as an example):
where
is the offered estimate of
at
ith sample.
- Step 9:
Compute the mean squared error (MSE) and mean absolute bias (MAB) of
:
and
- Step 10:
Compute the average interval length (AIL) and coverage provability (CP) of
:
and
where (
,
) refers to the estimated interval limits and
denotes the indicator operator.
The chosen prior distributions reflect reasonable assumptions about parameters , , and based on prior research and empirical observations in reliability and survival studies. Prior-A represents a more informative prior with moderate variance, assuming prior knowledge about the expected parameter range, whereas Prior-B has greater uncertainty by enabling higher variance, making it suitable for scenarios where less prior knowledge is available.
6.2. Simulation Results and Discussion
In
Table 2 and
Table 3, the APEs, MSEs, and MABs of
,
, and
can be found in the first, second, and third columns, respectively. Additionally, in
Table 4 and
Table 5, the AILs and CPs of
,
, and
can be found in the first and second columns, respectively.
From the facts presented in
Table 2 and
Table 5 and the plots in
Figure 2, the statistical behavior of the proposed estimators for parameters
,
, and
under various experimental conditions is summarized as follows:
As the sample size n increases, the performance of all the estimators improves significantly. Specifically, lower MSE, MAB, and AIL values are observed, along with higher CP values. These results demonstrate the asymptotic consistency of the proposed estimation methods, reinforcing their reliability for larger datasets.
Increasing r (or, equivalently, increasing FP%) enhances the precision of the calculated estimators. This improvement is reflected in reduced MSE, MAB, and AIL values, while CP values exhibit an increasing trend, indicating stronger inferential accuracy.
As we anticipate, when FP%, the precision of all simulation results of , , or behaves better under a complete sampling situation than others.
When comparing point estimates of , , and , the Bayesian estimation approach consistently outperforms the likelihood approach in terms of lower simulated MSE and MAB values. This suggests that the Bayesian framework provides more stable and efficient estimators in finite samples.
Regarding interval estimation, the BCI method demonstrates superior performance over the ACI method. Specifically, BCI-derived intervals exhibit shorter AIL values while maintaining higher CP values, emphasizing their greater informativeness and accuracy.
The choice regarding the prior distribution significantly influences the Bayesian estimation results. For all three parameters, , , and , estimates obtained using Prior-B outperform those derived from Prior-A. This can be attributed to the smaller variance of Prior-B, which leads to more precise and concentrated posterior distributions.
The overall estimation accuracy varies depending on the data structure. The results from Set-1 yield more precise estimates of , , and compared to Set-2. Moreover, as the values of these parameters increase, a deterioration in estimation precision is observed, characterized by higher MSE, MAB, and AIL values and a corresponding decline in CP values.
Finally, for analyzing discrete Hjorth data in the presence of censored observations, the Bayesian framework is strongly recommended due to its robust inferential properties, particularly in handling incomplete data efficiently.
Using the simulated values of MSEs and AILs (for example) corresponding to DH parameters
,
, and
,
Figure 2 provides clearer insights into the simulation results and confirms all the data presented in
Table 2,
Table 3,
Table 4 and
Table 5.
7. Real-World Data Analysis
This section analyzes two applications using two separate actual datasets to (i) assess the offered model’s adaptation and effectiveness to real problems; (ii) show how the inferential outcomes can be applied to a real-world scenario; and (iii) assess if the suggested model is more appropriate than eight other discrete models in the literature. Now, we consider the following applications:
The first application (say, App [A]) examines the failure times of eighteen electronic devices. These data were first introduced by Wang [
19]; see
Table 6.
The second application (say, App [B]) analyzes how many millions of revolutions each of 22 ball bearings will make before they fail; see Caroni [
20]. For computational purposes, we divide each revolution by ten and list the newly transformed ball-bearing data in
Table 6.
Briefly, in
Table 7, several statistics for applications [A] and [B], namely minimum, maximum, quartiles (1st, 2nd, and 3rd), mean, mode, standard deviation (St.D.), skewness, and kurtosis, are evaluated.
Aside from three well-known discrete models, namely negative binomial (NB), geometric (Geom), and Poisson (Pois), based on the above two applications, the DH model will be compared with eight other comparable models from the literature to demonstrate the reliability and advantage of the newly developed model. The competitive distributions include the following:
Discrete exponentiated-Chen (DEC
) by Alotaibi et al. [
21];
Exponentiated discrete Weibull (EDW
) by Nekoukhou and Bidram [
22];
Discrete modified Weibull (DMW
) by Almalki and Nadarajah [
23];
Discrete Burr Type-XII (DB
) by Krishna and Pundir [
24];
Discrete Perks (DP
) by Tyagi et al. [
25];
Discrete generalized-exponential (DGE
) by Nekoukhou et al. [
26];
Discrete gamma (DG
) by Chakraborty and Chakravarty [
27];
Discrete Burr–Hatke (DBH
) by El-Morshedy et al. [
28].
To judge the best model, several metrics are utilized, namely negative log-likelihood (N-LogL), Akaike (Ak.), consistent-Akaike (C-Ak.), Bayesian (Bayes), Hannan–Quinn (H–Q), and Kolmogorov–Smirnov (K–S) statistic (with its
p-value). From both applications [A] and [B], the fitted values corresponding to these criteria as well as the MLEs (along with their standard errors (Std.Ers)) of
,
, and
are obtained; see
Table 8. It is clear from
Table 8 for both analyzed applications that the fitted DH distribution produces the lowest values for all fitted metrics except the highest
p-value among all other fitted competing models. As a result, the DH probability model is better than others.
Several goodness-of-fit visualization tools are employed, including (i) histogram of data with fitted probability lines, (ii) fitted reliability lines, (iii) probability–probability (PP), and (iv) total-time-test (TTT) plots for the DH and its competing distributions, as displayed in
Figure 3. Clearly, in
Figure 3, we only compared the DH with models with
p-values greater than 5%. It exhibits that the newly proposed DH distribution provides the best fit for both applications in terms of estimated PMF, RF, and PP curves.
The estimated TTT lines shown in
Figure 3 indicate that the given datasets employed in applications [A] and [B] provide bathtub and increasing failure rates, respectively. These failure rate shapes support the same shapes of the DH model represented in
Figure 1. Additionally, all facts shown in
Figure 3 support the same numerical findings reported in
Table 8. We can therefore infer from the results presented in
Figure 3 (or
Table 8) that, compared to other traditional (or modern) statistical models, the new DH distribution provides a significantly superior fit.
Type-II censoring is a powerful tool in engineering reliability studies. It enables cost-effective, efficient, and statistically robust failure data analysis, ensuring better product design, maintenance, and quality control. As a result, it provides reliable failure estimates faster and cheaper than complete sampling.
To assess the obtained frequentist and Bayes estimators of
,
, and
, three artificial Type-II censored samples are obtained for each dataset based on varying the values of
r; refer to
Table 9. In every sample, the MLEs (along with the 95% ACIs) as well as the Bayes estimation (along with the 95% BCIs) are obtained. All Bayes evaluations are conducted by running the MCMC sampler 50,000 times and discarding the first 10,000 iterations as burn-in. Due to the lack of prior information about the DH’s parameters, we set
for
In
Table 10, the maximum likelihood and Bayes estimates (with their Std.Ers) as well as 95% ACI/BCI bounds (with their interval widths (ILs)) of
,
, and
are listed. It indicates that the point and interval estimation results developed from the Bayes method outperformed those developed from the likelihood method. When
r is increased, all estimates of
,
, or
behave well because their Std.Er and IL values decreased.
Based on the last 40,000 outputs for
,
, and
obtained from the two applications when
and 10 (for example), respectively,
Figure 4 shows the trace and density plots along with their sample averages (in solid lines) and 95% two-bound BCIs (in dashed lines). It shows that the MCMC sampler converges highly effectively and that the calculated posteriors for all DH parameters are reasonably symmetric.
Additionally, to explore the convergence and blending of Markovian chains, the acceptance rates of the M–H proposals are calculated, such as
For App. [A]: The acceptance rates from are 90.12, 93.40, and 95.38%;
For App. [B]: The acceptance rates from are 89.63, 92.78, and 97.47%.
As a consequence, the estimated acceptance rates in App. [A] (or App. [B]) confirm the same facts revealed in
Figure 4; that is, the percentage of iterations in which proposals were approved is significantly higher.
The numerical findings from the two applications employed, among other things, demonstrate that the suggested model outperformed the other models. As a result, the methods for applying the newly discrete Hjorth probability model to electrical devices or ball-bearing datasets provide an adequate explanation.
8. Conclusions
Discrete distributions are well suited for representing data confined to a finite or countably infinite set of data. The straightforward structure, availability of closed-form expressions, and capacity to capture real-world scenarios make them an ideal option. Additionally, they are computationally efficient and effectively handle categorical data modeling. This paper introduced a discrete version of the Hjorth distribution using the survival discretization method, addressing the gap in the literature regarding the use of censored data for analyzing discrete models. The proposed distribution preserved the essential statistical features of the original Hjorth distribution, such as its median, percentiles, and general structure, while providing a flexible framework for modeling real-world discrete data. The statistical properties of the DH model, including the quantile function and order statistics, were thoroughly derived. Maximum likelihood inference and Bayesian methods were employed for parameter estimation, and their performance was evaluated through simulation studies. Type-II censored samples were utilized to demonstrate the practical applicability of the model, with variance–covariance matrices enabling the construction of asymptotic confidence intervals for the parameters. The effectiveness of the DH model was further validated using two real-world datasets: one concerning failure times of electronic devices and another focusing on ball-bearing failures. The goodness-of-fit measures highlighted the superior performance of the proposed model compared to the existing discrete alternatives. Overall, the DH model proved to be a versatile and robust tool for statistical modeling, particularly in reliability and survival analysis. Its ability to adapt to different data patterns, such as decreasing, increasing, and bathtub-shaped hazard rates, underscores its potential for broader applications in various fields, including engineering, medicine, and beyond.
While the proposed DH model demonstrated strong flexibility and effectiveness in capturing various hazard rate shapes, certain limitations can be acknowledged. The model’s performance may be less reliable with very small sample sizes due to increased variability in parameter estimates, necessitating techniques such as Bayesian estimation with informative priors. Additionally, while the DH model demonstrated flexibility across various hazard rate trends, its fit may be affected when dealing with datasets exhibiting extreme variance or strong skewness.
Future research could explore refined model selection criteria to guide practitioners in choosing between the DH model and alternative approaches based on dataset characteristics. Future work may explore extending this framework to other distributions, different censoring schemes, and refining the computational methods for enhanced efficiency.