Next Article in Journal
On Chaos, Tipping and Delayed Dynamical Transitions in a Hassell-Type Population Model with an Allee Effect
Next Article in Special Issue
Fairness in Healthcare Services for Italian Older People: A Convolution-Based Evaluation to Support Policy Decision Makers
Previous Article in Journal
Advances in Clinical Trial Design: Employing Adaptive Multiple Testing and Neyman Allocation for Unequal Samples
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robustness and Efficiency Considerations When Testing Process Reliability with a Limit of Detection

Department of Statistics and Actuarial Science, University of Waterloo, 200 University Ave. W., Waterloo, ON N2L 3G1, Canada
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(8), 1274; https://doi.org/10.3390/math13081274
Submission received: 10 March 2025 / Revised: 10 April 2025 / Accepted: 11 April 2025 / Published: 12 April 2025
(This article belongs to the Special Issue Improved Mathematical Methods in Decision Making Models)

Abstract

:
Processes in biotechnology are considered reliable if they produce samples satisfying regulatory benchmarks. For example, laboratories may be required to show that levels of an undesirable analyte rarely (e.g., in less than 5% of samples) exceed a tolerance threshold. This can be challenging when measurement systems feature a lower limit of detection, rendering some observations left-censored. We investigate the implications of detection limits on location-scale model-based inference in reliability studies, including their impact on large and finite sample properties of various estimators and the sensitivity of results to model misspecification. To address the need for robust methods, we introduce a flexible weakly parametric model in which the right tail of the response distribution is approximated using a piecewise-constant hazard model. Simulation studies are reported that investigate the performance of the established and proposed methods, and an illustrative application is given to a study of drinking can weights. We conclude with a discussion of areas warranting future work.

1. Introduction

Testing and manufacturing processes in biotechnology are considered reliable if they produce samples satisfying regulatory benchmarks. For example, laboratories separating blood into transfusable components are required to show that levels of an undesirable analyte in the blood units produced rarely (e.g., in less than 5% of samples) exceed a tolerance threshold [1]. Likewise, the United States Food and Drug Administration stipulates maximum allowable levels of residual chemicals and bacterial toxins for medical devices labelled as sterile [2], and of various ingredients for food products labelled as allergen- or gluten-free [3]. Similar situations arise in environmental science, where sites such as nuclear and waste disposal plants must demonstrate that the concentration of a contaminant in the surrounding soil, air, or water is sufficiently low [4].
In these scientific settings, measurement systems often feature a lower limit of detection (LLD), rendering some observations left-censored. Two common ad hoc approaches for dealing with such data are “deletion”, in which samples yielding values below the LLD are discarded, and “substitution”, in which values are taken to be zero, the LLD itself, or some intermediate value (such as half the LLD) [5,6]. While these methods continue to be widely used [7], much of the literature in environmental science has focused on characterizing their limitations and investigating improved techniques for left-censored data (e.g., [8,9,10,11,12,13,14]). These techniques include Kaplan–Meier estimation [15], maximum likelihood estimation for parametric models [16], and “regression on order statistics” [11], the latter of which involves conducting least squares regression on the quantiles of the log-transformed observations. The targets of inference in this work have mostly been central features of the measurement distribution, such as the mean, variance, median, and interquartile range. Our interest lies in the impact of an LLD in studies aiming to demonstrate process reliability, which inherently aims at inferences regarding the upper tail of the distribution. For example, demonstrating that the levels of an analyte are “low enough” (i.e., below some specified threshold) in at least 95% of blood samples requires inferences regarding the 95th percentile of the response distribution. Huynh et al. [17] investigated a fully parametric Bayesian approach to estimating various quantities but found that this method performed comparably to or worse than substitution methods when estimating the 95th percentile. The impact of an LLD on frequentist inference in the upper tail of a distribution has been less studied and motivates our work.
In Section 2, we introduce some notation, specify the hypothesis test of interest, and discuss analysis in the setting without an LLD, and in Section 3, we examine the large sample impacts of an LLD on estimation and inferences from a (possibly misspecified) normal model. Tests of reliability based on misspecified models will often not control rejection rates at the desired levels. Moreover, the power of goodness of fit tests for detecting model misspecification is generally low for typical sample sizes, particularly in the presence of an LLD. In Section 4, we address these issues by introducing a flexible weakly parametric method in which the right tail of the measurement distribution is approximated using a piecewise-constant hazard model. In Section 5, we use simulation studies to investigate finite sample settings with sample sizes commonly encountered in laboratory research. We also report on findings from simulation studies which can help guide piecewise-constant hazard-based model selection as well as sample size specification under various models, and an R function is available for custom investigations. An application to a study of drinking can weights is given for illustration in Section 6. We conclude in Section 7 with some recommendations for the design of future studies and discuss areas for future work.

2. Background on Methods

Let Y denote the real-valued response with cumulative distribution function (c.d.f.) F and density function f. Suppose n independent observations y 1 , , y n are available from the distribution. For i = 1 , , n , suppose observation y i is subject to LLD d i such that the measurement y i is available if y i > d i , but otherwise, we only know I ( y i < d i ) . One could consider settings with different LLDs for different subsets of observations; this may be suitable in studies involving multiple laboratories or measurement devices. One may also view the LLD as a quantity that varies randomly depending on environmental factors such as ambient temperature or humidity. For simplicity, we consider the setting of a study with a single measurement device and set d i = d for all i.
We consider the goal of demonstrating that a process produces satisfactory measurements defined by a desired percentage of values below some threshold τ . We therefore aim to show F ( τ ) = P ( Y < τ ) > p 0 at the α significance level (i.e., with “100 ( 1 α ) % confidence”) for some probability p 0 . Letting p 1 = F ( τ ) denote the true value of the c.d.f. at τ , this entails a test of reliability:
H 0 : p 1 = p 0 vs . H A : p 1 > p 0 .
To carry out a test of this null hypothesis, we focus on the broadly applicable and mathematically convenient family of parametric location-scale models [18]. By convention, we work with the real-valued forms of these distributions (e.g., normal, logistic, extreme value) rather than the non-negative analogs (e.g., log-normal, log-logistic, Weibull); as a result, when measurements take non-negative values (e.g., concentrations), it may be necessary to re-map the original measurements (and the original threshold), say using the log transform, in order to obtain real-valued Y and τ .
We now consider the hypothesis test of interest in (1) based on parametric location-scale models. Suppose Y is in the location-scale family with location parameter θ 1 and scale parameter θ 2 with θ = ( θ 1 , θ 2 ) . We write
p 1 = F ( τ ; θ ) = F 0 τ θ 1 θ 2
where F 0 ( · ) is the c.d.f. for the standardized random variable ( Y θ 1 ) / θ 2 , so that the hypotheses of interest can be expressed equivalently as
H 0 : F 0 τ θ 1 θ 2 = p 0 vs . H A : F 0 τ θ 1 θ 2 > p 0 .
With a lower limit of detection, the data for a sample of n independent observations are { ( x i , δ i ) , i = 1 , , n } , where x i = max ( y i , d ) and δ i = I ( x i = y i ) , i = 1 , , n . The central idea in constructing the testing procedure is to exploit the assumption of the parametric model to learn about the shape of the distribution and enhance inference (i.e., increase the precision and power of tests) regarding the upper tail of the distribution. Under a working model defined by the error distribution F 0 ( · ) , the likelihood is
L ( θ ; x , δ ) = i = 1 n f 0 x i θ 1 θ 2 δ i F 0 x i θ 1 θ 2 1 δ i
where x = ( x 1 , , x n ) and δ = ( δ 1 , , δ n ) . Such a model can be fitted to left-censored data via maximum likelihood [16] using the survreg function in R version 4.3.2 [19] or using Parametric Distribution Analysis (Arbitrary Censoring) in Minitab® version 21.4.0. To get a 100 p % confidence interval (CI) for
F ( τ ; θ ) = F 0 τ θ 1 θ 2 = F 0 ( ψ 1 + ψ 2 τ )
where ψ 1 = θ 1 / θ 2 and ψ 2 = 1 / θ 2 , we first use the delta method [20] to obtain a 100 p % CI for ψ 1 + ψ 2 τ , then apply F 0 to the endpoints of this CI. We reject H 0 if the lower limit of the CI exceeds p 0 and fail to reject H 0 otherwise.

3. Inference Under Model Misspecification

The normal distribution is often the default model for responses in applications, so we next consider properties of estimators and inference procedures when one assumes Y 1 , , Y n are (independently) generated from a normal distribution with mean θ 1 a and variance θ 2 a 2 and the data-generating distribution may differ. As illustrations, we consider the cases where Y is generated from (i) a logistic distribution with c.d.f.
F T ( y ; θ ) = 1 + exp y θ 1 θ 2 1 ,
and (ii) an extreme value distribution with c.d.f.
F T ( y ; θ ) = 1 exp exp y θ 1 θ 2 .
Here, we use subscript T to emphasize that F T is the true (not assumed) c.d.f. of Y.
Now let X i = max ( Y i , d ) with Δ i = I ( X i = Y i ) , and fix the left-censoring rate q = P ( Y i d ) = E { 1 Δ i } . Then, the observed data likelihood for the assumed normal model is
L ( θ a ; x , δ ) = i = 1 n 1 θ 2 a exp 1 2 x i θ 1 a θ 2 a 2 δ i F a , 0 x i θ 1 a θ 2 a 1 δ i ,
or, equivalently,
L ( ψ a ; x , δ ) = i = 1 n ψ 2 a exp 1 2 ( ψ 1 a + ψ 2 a x i ) 2 δ i F a , 0 ( ψ 1 a + ψ 2 a x i ) 1 δ i
where F a , 0 denotes the c.d.f. of N ( 0 , 1 ) , and ψ a = ( ψ 1 a , ψ 2 a ) = ( θ 1 a / θ 2 a , 1 / θ 2 a ) such that
ψ 1 a + ψ 2 a x = x θ 1 a θ 2 a .
Since δ i = I ( x i = y i ) , we can also write this as
L ( ψ a ; y , δ ) = i = 1 n ψ 2 a exp 1 2 ( ψ 1 a + ψ 2 a y i ) 2 δ i F a , 0 ( ψ 1 a + ψ 2 a d ) 1 δ i .
As shown by White [21], the estimator ψ ^ a solves the score equations S ( ψ a ; y , δ ) = i = 1 n S i ( ψ a ; y i , δ i ) = 0 corresponding to (4) and satisfies
n ( ψ ^ a ψ a * ) M V N ( 0 , A 1 ( ψ a * ) B ( ψ a * ) A 1 ( ψ a * ) )
where
  • ψ a * solves E T { S i ( ψ a ; Y i , Δ i ) } = S i ( ψ a ; y i , δ i ) d F T ( y i ; ψ ) = 0 ,
  • A ( ψ a * ) = E T ψ a S i ( ψ a ; Y i , Δ i ) | ψ a = ψ a * , and
  • B ( ψ a * ) = E T S i ( ψ a ; Y i , Δ i ) S i ( ψ a ; Y i , Δ i ) | ψ a = ψ a *
and E T { · } denotes the expectation under the true data-generating model; see also Gourieroux, Monfort, and Trognon [22] and Qin and Lawless [23]. Appendix A shows how to compute the limiting values and implicit estimands ψ a * when the true distribution is logistic or extreme value (case (i) or (ii) above, respectively). Recall that ψ 1 = θ 1 / θ 2 and ψ 2 = 1 / θ 2 , so to recover the limiting values of estimates for θ 1 , log ( θ 2 ) , ψ 1 + ψ 2 τ , and F T ( τ ; ψ ) , we, respectively, use the following:
  • ψ 1 a * / ψ 2 a * ;
  • log ( ψ 2 a * ) ;
  • ψ 1 a * + ψ 2 a * τ ;
  • F a , 0 ( ψ 1 a * + ψ 2 a * τ ) .
The bias of an estimator is defined as its estimand based on the assumed model minus its intended estimand (e.g., bias ( θ ^ 1 a ) = ψ 1 a * / ψ 2 a * θ 1 ). Plots of bias as a function of the left-censoring rate q for the first two estimands listed above are displayed in Figure 1; here, we use scale θ 2 = 1 (i.e., log ( θ 2 ) = 0 ), and location θ 1 , chosen to align the 95th percentile of the data-generating distribution with that of the standard normal, resulting in θ 1 = 1.30 for the logistic distribution and θ 1 = 0.55 for the extreme value distribution. Examining these plots at q = 0 shows that, in the absence of an LLD, model misspecification may lead to substantial bias. Note that larger LLDs may or may not be associated with less absolute bias, indicating that a loss of information in the lower tail of the distribution may, in some cases, improve performance in terms of bias but the LLD cannot be relied upon for this purpose.
The bias curves for the second two estimands, displayed in Figure 2, further demonstrate the meaningful bias that may result from model misspecification in the absence of an LLD, even at high reliability levels. As before, increasing the left-censoring rate may or may not reduce bias here, although the reliability estimator under the extreme value error distribution shows marked reduction in asymptotic bias with a larger LLD.
In order to investigate the relationship between the LLD and rejection rates for tests of the hypotheses in Section 2, we next examine the asymptotic rejection rate as a function of the left-censoring rate. We begin by reformulating our hypothesis test in terms of the quantity
c ψ = τ θ 1 θ 2
where c = ( 1 , τ ) and ψ = ( ψ 1 , ψ 2 ) = ( θ 1 / θ 2 , 1 / θ 2 ) ; in particular, since F a ( τ ; ψ ) = F a , 0 ( c ψ ) , where F a , 0 is again the standardized c.d.f. under the assumed model, and we wish to test
H 0 : F a ( τ ; ψ ) = p 0 vs . H A : F a ( τ ; ψ ) > p 0 ,
we apply F a , 0 1 to both sides of each hypothesis to get
H 0 : c ψ = F a , 0 1 ( p 0 ) vs . H A : c ψ > F a , 0 1 ( p 0 ) .
Now, let ψ ^ a be the estimate of ψ under the assumed model with limiting value ψ a * . Given the robust (“true”) variance Var T ( ψ ^ a ) of White [21] and the corresponding variance Var a ( ψ ^ a ) under the assumed model, we can compute the asymptotic rejection rate (at the 0.025 one-sided significance level) for our reformulated hypothesis test using
P c ψ ^ a F a , 0 1 ( p 0 ) Var a ( c ψ ^ a ) > 1.96 = 1 P F a , 0 1 ( p 0 ) c ψ a * + 1.96 c Var a ( ψ ^ a ) c c Var T ( ψ ^ a ) c .
Explicit expressions for Var T ( ψ ^ a ) and Var a ( ψ ^ a ) are derived under an assumed normal model in Appendix B and Appendix C.
In a statistical hypothesis test, the “type I error rate” refers to the probability of rejecting H 0 when it is true (i.e., incorrectly concluding that p 1 > p 0 when the truth is p 1 = p 0 in this case), while “power” refers to the probability of rejecting H 0 when it is false (i.e., correctly concluding that p 1 > p 0 ) across multiple samples from the same population. An illustration of the rejection rates resulting from (5) is shown in Figure 3 (type I error; p 1 = 0.95 ) and Figure 4 (power; p 1 = 0.97 , 0.98 , 0.99 ), respectively; here, we use p 0 = 0.95 , n = 80 , scale θ 2 = 1 , and location θ 1 , chosen to align the p 1 quantile of the data-generating distribution with that of the standard normal.
As expected, when the error distribution is normal (and thus matches the assumed model), the type I error rate is at the nominal level of 0.025 (red dashed line in Figure 3) and power is a strictly decreasing function of the left-censoring rate q (Figure 4). When the error distribution is logistic, we see an elevated false positive rate relative to the normal data case (Figure 3) as well as increased power (Figure 4); by contrast, when it is extreme value, the false positive rate is very low and power is also greatly diminished. Inference is invalid in both of these cases due to poor control of the type I error rate. Interestingly, the impact of q on asymptotic power does not depend much on the true reliability level p 1 in the normal and logistic error settings but increases substantially as a function of p 1 with extreme value errors (Figure 4).

4. A Piecewise-Constant Hazard-Based Model

In Appendix D, we investigate the performance of the Shapiro–Francia test [24] of normality under an LLD and find that some location-scale distributions that differ greatly in their tails (and, therefore, differ greatly in their reliability estimates) may be indistinguishable to whole-distribution goodness of fit tests. Based on this finding, we now introduce a flexible class of models based on piecewise-constant hazard models from the field of survival analysis [18]. This approach requires fewer parametric assumptions and intentionally discards observations in the lower tail of the measurement distribution to improve the robustness of inference in the upper tail.
We first consider non-negative values V i = exp ( Y i ) , i = 1 , , n , and corresponding threshold value τ V = exp ( τ ) . To limit the influence of observations in the lower tail on inference in the upper tail, we next restrict attention to values of v > τ V ϵ for some tolerance ϵ 0 . In particular, we take v 0 = τ V ϵ and simply assign a mass to the event V < v 0 given by p = P ( V < v 0 ) . Letting η = logit ( p ) yields
F V ( v 0 ; η ) = P ( V < v 0 ; η ) = exp ( η ) 1 + exp ( η ) = expit ( η ) .
Then, to model F V ( v ) = 1 F V ( v ) for v > v 0 , we partition ( v 0 , ) into K intervals using cut-points v 0 = b 0 < b 1 < < b K = and let B k = [ b k 1 , b k ) for k = 1 , , K . We then use the model
F V ( v ; p , ρ ) = P ( V > v ; p , ρ ) = ( 1 p ) e Λ ( v 0 , v ; ρ )
where the following apply:
  • Λ ( v 0 , v ; ρ ) = k = 1 K W k ( v 0 , v ) ρ k is the “cumulative hazard”;
  • W k ( v 0 , v ) = v 0 v I ( u B k ) d u is the “time at risk” in B k over the interval [ v 0 , v ) .
Letting β k = log ρ k for k = 1 , , K and β = ( β 1 , , β K ) , we can write
F V ( v ; η , β ) = ( 1 expit ( η ) ) exp k = 1 K W k ( v 0 , v ) e β k .

4.1. Maximum Likelihood Estimation

In what follows, we suppress the subscript V on density f V , c.d.f. F V , and survivor function F V of V, for ease of notation. Since d W k ( v 0 , v ) / d y = I ( v B k ) , then the corresponding density (assuming v v 0 so that I ( v B k ) = 1 for some k { 1 , , K } ) is given by
f ( v ; η , β ) = ( 1 expit ( η ) ) exp k = 1 K W k ( v 0 , v ) e β k k = 1 K I ( v B k ) e β k .
Now, for a sample of n observations from independent units, let δ i = I ( v i v 0 ) with δ · = i = 1 n δ i for i = 1 , , n . Further, let S i k = W k ( v 0 , v i ) with S · k = i = 1 n S i k and δ i k = I ( v i B k ) with δ · k = i = 1 n δ i k for k = 1 , , K . Note that since v can only fall in interval B k if v v 0 , then δ i δ i k = δ i k , and since S i k = 0 when δ i = 0 , then δ i S i k = S i k . Lastly, we replace k = 1 K I ( v B k ) e β k with k = 1 K e I ( v B k ) β k in (6). Then, accounting for both v ( 0 , v 0 ) and v ( v 0 , ) , the likelihood is
L ( η , β ; v ) = i = 1 n F ( v 0 ; η ) 1 δ i f ( v i ; η , β ) δ i = i = 1 n [ expit ( η ) ] 1 δ i ( 1 expit ( η ) ) exp k = 1 K W k ( v 0 , v i ) e β k k = 1 K e I ( v i B k ) β k δ i = [ expit ( η ) ] n δ · ( 1 expit ( η ) ) δ · exp i = 1 n k = 1 K δ i S i k e β k i = 1 n k = 1 K e δ i δ i k β k = [ expit ( η ) ] n 1 expit ( η ) expit ( η ) δ · exp k = 1 K β k δ · k S · k e β k .
Now, let θ = ( η , β ) . We have
( θ ) = log L ( θ ) = n log ( expit ( η ) ) η δ · + k = 1 K β k δ · k S · k e β k .
Taking the first partial derivatives of ( θ ) and solving the resulting estimating equations readily yields maximum likelihood estimates of
η ^ = logit 1 δ · n and β ^ k = log δ · k S · k , k = 1 , , K .

4.2. Profile Likelihood Ratio Test

To test process reliability using this model, we use a profile likelihood ratio test [16], which is conducted as follows:
  • Write L ( γ , β ) where γ = γ ( θ ) = F ( τ V ; θ ) : letting
    F 1 ( τ ; β ) = exp k = 1 K W k ( v 0 , τ ) e β k ,
    we have γ = ( 1 expit ( η ) ) F 1 ( τ V ; β ) , so
    expit ( η ) = F 1 ( τ V ; β ) γ F 1 ( τ V ; β ) and 1 expit ( η ) expit ( η ) = γ F 1 ( τ V ; β ) γ .
    Hence, continuing from the likelihood function (7) above, we have
    L ( γ , β ) = γ δ · [ F 1 ( τ V ; β ) γ ] n δ · F 1 ( τ V ; β ) n exp k = 1 K β k δ · k S · k e β k .
  • Compute the profile likelihood estimate β ^ ( γ ) : We can optimize the log-likelihood
    ( γ , β ) = δ · log γ + ( n δ · ) log ( F 1 ( τ V ; β ) γ ) n log F 1 ( τ V ; β ) + k = 1 K β k δ · k S · k e β k = δ · log γ + ( n δ · ) log exp k = 1 K W k ( v 0 , τ V ) e β k γ + n k = 1 K W k ( v 0 , τ V ) e β k + k = 1 K β k δ · k S · k e β k .
    numerically for a given value of γ to get β ^ ( γ ) .
  • Under H 0 : γ = γ 0 , we have
    2 ( γ 0 , β ^ ( γ 0 ) ) ( γ ˜ , β ˜ ) χ ( 1 ) 2 ,
    where η ˜ and β ˜ are the (non-profile) MLEs derived in Section 4.1, and γ ˜ = γ ˜ ( η ˜ , β ˜ ) . Thus, a 100 ( 1 α ) % profile likelihood interval for γ is given by
    ( γ , β ^ ( γ ) ) ( γ ˜ , β ˜ ) χ ( 1 , 1 α ) 2 / 2
    where χ ( 1 , 1 α ) 2 is the 1 α quantile of χ ( 1 ) 2 .
  • Reject H 0 if γ 0 = 1 p 0 falls outside the profile likelihood interval computed in step (3). This constitutes a hypothesis test at significance level α .

5. Empirical Studies

5.1. Bias, Power, Robustness, and Efficiency

To assess the bias and efficiency of estimation and the power of reliability tests for these methods, and to assess their robustness to model misspecification, we consider several simulation studies. In each of the simulations detailed below, two thousand simulation iterations are run and we conduct one-sided tests for p 0 = 0.95 reliability at both the 0.025 and 0.05 significance levels.
For each of the methods, we run a simulation for each combination of (a) sample size n; (b) true reliability p 1 ; (c) true data-generating distribution; (d) fitted model; and (e) LLD value d. Possible sample sizes n include 20, 40, 60, and 80, while possible values for p 1 include 0.95 (representing the null hypothesis), 0.97, and 0.99. All true data-generating distributions are set to have scale parameter θ 2 = 1 ; under this constraint, the possible true distributions include the standard normal, as well as logistic and extreme value distributions, with location parameter θ 1 chosen to align the p 1 quantile of the true distribution with that of the standard normal. See Equations (2) and (3) for the specification of the logistic and extreme value distributions. Possible fitted models vary depending on the modelling method. In the case of the fully parametric models of Section 2, we fit normal models, so the fitted model is correctly specified when the true data-generating distribution is standard normal but misspecified otherwise. In the case of the piecewise-constant hazard-based models, we let Q p denote the pth percentile of the true data-generating distribution and consider models with interval cut-points at b 0 = Q 90 ; b 0 = Q 85 , b 1 = Q 90 , and b 2 = Q 95 ; and b 0 = Q 85 , b 1 = Q 89 , and b 2 = Q 93 ; as well as models with cut-points at the corresponding empirical percentiles. We focus here on cut-points Q p with p 85 to ensure robustness in approximating the upper tail of the measurement distribution near the critical reliability threshold τ . Varying the LLD only affects the fully parametric method since the other models deliberately avoid distinguishing between observations above and below the LLD for sufficiently small values in y 1 , , y n . For this method, one possible value of the LLD is d = Q 10 N where Q p N denotes the pth percentile of the standard normal distribution. Under this LLD, the asymptotic percentage of left-censored observations is 10% when fitting a normal model but greater than 10% when fitting logistic or extreme value models. Thus, to disentangle the effects of these differing left-censoring rates from the effects of model misspecification itself, we also consider an LLD of d = Q 10 , based on the true data-generating distribution.
In order to more extensively explore the spectrum of models from fully parametric to fully non-parametric, we compare the results of the normal and piecewise-constant hazard-based models to those obtained using a non-parametric method, where p-values and smoothed “mid-p” values are computed according to a one-sided exact binomial test [25].
A selection of simulation results is shown in Table 1 and Table 2 (full results can be found in the Supplementary Materials online). Table 1 shows the estimates of F ( τ ) obtained using each of the models described above, while Table 2 shows the corresponding rejection rates for testing H 0 : F ( τ ) = p 0 vs. H A : F ( τ ) > p 0 where p 0 = 0.95 . In each table, the first section of twelve lines shows the results of simulating data under the null hypothesis that reliability is at the nominal level of p 0 = 0.95 , while in the second and third sections, the reliability exceeds p 0 . Note that each column heading for the piecewise-constant hazard-based models indicates the p 1 , , p K empirical percentiles Q ^ p 1 , , p K defining the interval cut-points of the fitted model; the model with cut-points at Q ^ 85 , Q ^ 89 , and Q ^ 93 is omitted when n = 20 since no observations would fall in the interval [ Q ^ 85 , Q ^ 89 ) , and such a model would, therefore, not be used in this case.
The leftmost column of Table 1 shows that while fitting a normal model leads to accurate estimates of F ( τ ) when the data are truly normal, bias becomes appreciable under the logistic and, particularly, the extreme value error distribution; this matches the asymptotic results shown in Figure 2. Similarly, Table 2 shows that type I error is well controlled and power reasonably high for larger sample sizes under correct model specification, but (as expected based on Figure 3 and Figure 4) these rejection rates are inflated and severely depressed, respectively, for the logistic and extreme value error distributions. The last column of Table 1 and the last two columns of Table 2 show that the exact tests address the bias issue and perform roughly the same across all settings since they do not rely on distributional assumptions; both p-values and mid-p-values here result in very low rejection rates, however, even when the true reliability of F ( τ ) = 0.99 far exceeds the nominal reliability of 0.95 that we aim to demonstrate. Prohibitively large sample sizes would be required to achieve acceptable levels of power using exact tests.
The piecewise-constant hazard-based models generally outperform the normal model in terms of both bias in the reliability estimate (Table 1) and power (Table 2) when the data are extreme value, while at the same time, they perform comparably to the normal model when data are generated under a normal or logistic distribution. Bias is least for larger values of F ( τ ) where down-weighted values in the left tail become less relevant for estimation (Table 1). Type I error is relatively well controlled for all simulation settings, particularly at larger sample sizes, which is a notable improvement in the cases of logistic and extreme value data generation (Table 2). Overall, the piecewise-constant hazard-based method provides a balance between the power obtained by correctly specified parametric models and the robustness of non-parametric methods. While the small sample sizes we have focused on here are typical in reliability studies, the performance of these piecewise-constant models is highly scalable as larger samples would allow one to increase the number of cut-points (thereby increasing flexibility and robustness) without sacrificing power.

5.2. Model Selection

Although Table 1 and Table 2 show that no one piecewise-constant hazard-based model uniformly outperforms the others, model cut-points may be chosen by optimizing the worst possible outcome across all simulation settings. Figure 5 shows the minimum sample size needed to achieve the worst-case empirical power of 90% across normal, logistic, and extreme value data-generating distributions, using a piecewise-constant hazard-based model with two cut-points b 0 and b 1 , as well as the largest possible empirical type I error rate across those distributions when the sample size is fixed at this optimal value. As before, each of these distributions has θ 2 = 1 and a 95th percentile equal to that of a standard normal.
Here, we see that when testing for p 0 = 0.95 and the true reliability is p 1 = 0.99 , the two-cut-point model that best balances power and the type I error rate appears to have its cut-points at the 90th and 98th percentiles of the observed data (Figure 5). It is important to note that even in this optimal case, we require a sample size of 150 to achieve the desired power of 90%, and the corresponding type I error rate is slightly inflated at 0.039. If controlling the type I error rate is of greater importance, one may prefer a model with cut-points at the 92nd and 98th percentiles, where we see a required sample size of 180 and a worst-case type I error rate of 0.022.
Figure 5 illustrates the trade-off between type I error rates and the sample sizes required to achieve the specified power level using two-cut-point models (for particular values of p 0 , p 1 , threshold τ , significance level α , and worst-case empirical power), and similar analyses could be conducted for different choices of tolerance ϵ and number of intervals K (as well as other specifications of the reliability test). Piecewise-constant hazard-based models do not typically require many cut-points to accurately capture the shape of the target distribution [26], and in the context of reliability studies, we recommend using two to four cut-points ( 1 K 3 ). Moreover, the smaller the tolerance ϵ , the closer the first cut-point b 0 will be to the exponentiated threshold τ V and the closer the piecewise-constant hazard-based reliability test will be to the exact binomial test from Section 5.1; smaller ϵ values, therefore, result in greater robustness but worse efficiency, as seen by the larger sample sizes and better-controlled type I error rates at the lower ends of Figure 5a and Figure 5b, respectively. The effects of changing the number and spacing of the other cut-points are more difficult to characterize, and fitting multiple models is recommended in practice to assess the sensitivity of results to a given choice.

6. Application

In this section, we apply the piecewise-constant hazard-based models introduced in Section 4 to a study of drink can weights, used as an example in the documentation for the PROC CAPABILITY statement in SAS® version 9.4 [27], and we compare the results to those obtained using the fully parametric models of Section 2. The corresponding dataset consists of 100 can weights that we suppose were measured using an industrial scale with an LLD of 0.9 oz; eight of these measurements fall below the LLD, while 98 fall below the threshold exp ( τ ) = 1.12 . Here, interest may lie in demonstrating that at least 95% of cans weigh at most 1.12 oz (say, in order to meet freight shipment weight limits). We therefore aim to test H 0 : F ( τ ) = p 0 vs. H A : F ( τ ) > p 0 where τ = log ( 1.12 ) , p 0 = 0.95 , and F is the c.d.f. of the distribution of log-can weights. We perform analysis on the log scale, as indicated in Section 2.
Figure 6 shows the empirical c.d.f. of the data for observations above the LLD. Non-parametric methods yield F ^ ( τ ) = 98 / 100 = 0.98 , with exact binomial tests resulting in
p - value = u = 98 100 100 u · 0 . 95 u · ( 1 0.95 ) 100 u = 0.118
and
mid - p - value = p - value 1 2 100 98 · 0 . 95 98 · ( 1 0.95 ) 100 98 = 0.078 .
While the estimate F ^ ( τ ) is above p 0 = 0.95 , neither the p-value nor the mid-p-value provides evidence of process reliability at the 0.025 significance level.
The upper tails of various fitted fully parametric (Section 2) and piecewise-constant hazard-based (Section 4) models are shown in Figure 7, with corresponding reliability estimates and p-values shown in Table 3. Here, we fit logistic and extreme value models in addition to the normal models examined in the derivations of Appendix AAppendix C and the simulation study of Section 5.1. Each piecewise-constant hazard-based model is specified by either the vector of empirical pth percentiles Q ^ p used as its cut-points (Table 3) or, for ease of plotting, the corresponding vector of probabilities p (Figure 7); these models are chosen as follows:
  • ( Q ^ 90 ) is one of the models considered in the simulation study of Section 5.1;
  • ( Q ^ 85 , Q ^ 93 , Q ^ 95 ) is an adaptation of another model ( Q ^ 85 , Q ^ 90 , Q ^ 95 ) from the simulation study, with the middle cut-point adjusted upward to ensure that each interval of the resulting model strictly contains at least one observation;
  • ( Q ^ 92 , Q ^ 98 ) is suggested by the simulation from Section 5.2;
  • ( Q ^ 91 , Q ^ 99 ) and ( Q ^ 94 , Q ^ 97 ) are the models that are, respectively, the closest to and farthest from demonstrating reliability from a grid search of two-cut-point models whose first cut-point is in { Q ^ 91 , Q ^ 92 , Q ^ 93 , Q ^ 94 } and whose second cut-point is in { Q ^ 95 , Q ^ 96 , , Q ^ 99 } .
Note that the third model ( Q ^ 85 , Q ^ 89 , Q ^ 93 ) from the simulation study is omitted because Q ^ 85 = Q ^ 89 for this dataset, meaning that it is effectively a two-cut-point model.
Table 3 shows that, as with the non-parametric method, each of these eight model-based reliability estimates falls above the desired level of p 0 = 0.95 . Unlike the exact binomial test, however, some of the corresponding confidence and likelihood intervals have lower limits above 0.95 and, therefore, suggest that the drink can manufacturing process is reliable. We also see that likelihood intervals based on the piecewise-constant hazard-based models are much less variable than the confidence intervals based on their fully parametric counterparts.
Practitioners often use so-called probability plots to visualize the outcomes of fully parametric model fitting, and these are shown in Figure 8. Here, we plot the log-can weights y on the horizontal axis and ψ ^ 1 + ψ ^ 2 y on the vertical axis but label these axes with can weights exp ( y ) and reliability estimates F a , 0 ( ψ ^ 1 + ψ ^ 2 y ) , respectively, where F a , 0 is the c.d.f. of the assumed standard error distribution; see Section 2 for the definition of ψ .

7. Conclusions

The impact of an LLD on inference in the centre of a distribution has been well studied in the environmental science literature. In this paper, we show that when interest lies in the right tail of the distribution and parametric models are misspecified, an LLD can attenuate bias and even slightly increase the power of tests for reliability, but these modest benefits are not guaranteed and, critically, come at the cost of invalid inference due to poor control of type I error rates. Moreover, the power of goodness of fit tests can be limited even with larger sample sizes, and the presence of an LLD exacerbates this issue; selection of an appropriate fully parametric model, therefore, presents a substantial challenge. In order to circumvent this challenge, we introduce a weakly parametric method of testing process reliability based on piecewise-constant hazard models that de-emphasizes observations in the lower tail of the measurement distribution and successfully balances efficiency with robustness to model misspecification. We use simulations to identify two two-cut-point piecewise-constant hazard-based models that may provide good asymptotic results for particular values of the true and required reliability levels, and apply the methods discussed in this paper to a study of drinking can weights.
While the methods proposed here provide a basic framework for right-tail inference under an LLD, one limitation is that (as seen by Figure 5) the results from piecewise-constant hazard-based models are somewhat sensitive to one’s choice of cut-points. To address this issue, one could consider using splines (say, using survPen in R version 4.3.2 [28]), which may offer more robustness at a modest price in terms of efficiency and power. Practical considerations can also complicate analysis. One such issue is that LLDs may vary between subsets of observations; for example, systematic variation might occur if technological advancements decrease an LLD over time by improving measurement device precision [5,9], while random variation might occur if environmental factors such as wind speed or ambient temperature affect device performance [29]. In other settings involving multiple laboratories, different devices may be used to measure concentrations of analytes, so device-to-device variation in detection limits can also raise challenges. Models for the LLD may be helpful in these situations. Other practical issues that merit attention occur when multiple different LLDs are present (e.g., due to use of multiple measurement devices) and when measurements between the LLD and a so-called “limit of quantification” are said by the device manufacturer to be observed with some uncertainty. More work is required to develop methods suited to these more complicated settings.
Finally, one reviewer raised the notion of Bayesian reliability testing. As noted in Section 1, when one uses fully parametric models of the form presented in Section 2, Bayesian methods have been found to estimate the 95th percentile comparably (at best) to substitution methods [17]. A Bayesian approach could also be compatible, however, with the piecewise-constant hazard-based models that we have proposed here. In particular, one would need to specify a prior for each parameter in the model (e.g., a gamma prior for the exponential hazard rate in each piece [30] and a beta prior for the probability p), calculate the posterior distribution for the true reliability p 1 , and stipulate a decision rule for using this posterior to conduct the test of reliability given by (1). Such a test could be done by constructing a 100 ( 1 α ) % credible interval for p 1 and checking whether it covers p 0 [31]. This is an interesting possible direction for future research.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/math13081274/s1.

Author Contributions

Conceptualization, L.S.B. and R.J.C.; methodology, L.S.B. and R.J.C.; software, L.S.B.; validation, L.S.B.; formal analysis, L.S.B.; resources, R.J.C.; writing—original draft preparation, L.S.B.; writing—review and editing, L.S.B. and R.J.C.; visualization, L.S.B.; supervision, R.J.C.; funding acquisition, R.J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Sciences and Engineering Research Council of Canada through Discovery Grants to Richard J. Cook (RGPIN-2017-04207).

Data Availability Statement

The data that support the findings of this study are openly available at https://documentation.sas.com/doc/en/pgmsascdc/9.4_3.4/qcug/qcug_code_capspec2.htm (accessed on 26 November 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Parameter Estimands in Misspecified Location-Scale Models with Left-Censored Data

Given likelihood (4), the observed log-likelihood is
( ψ a ; y , δ ) = i = 1 n δ i log ψ 2 a 1 2 ( ψ 1 a + ψ 2 a y i ) 2 + ( 1 δ i ) log F 0 ( z d ) .
Now, consider the resulting random vector S i ( ψ a ; Y i , Δ i ) of score functions for a single observation, obtained by taking partial derivatives of a contribution to (A1) with respect to ψ a ; we drop the subscripts i for notational convenience to denote this by S ( ψ a ; Y , Δ ) . Note z d / ψ 1 a = 1 and z d / ψ 2 a = d , so
S 1 ( ψ a ; Y , Δ ) = ψ 1 a = Δ ( ψ 1 a + ψ 2 a Y ) + ( 1 Δ ) f 0 ( z d ) F 0 ( z d )
and
S 2 ( ψ a ; Y , Δ ) = ψ 2 a = Δ 1 ψ 2 a Y ( ψ 1 a + ψ 2 a Y ) + ( 1 Δ ) d f 0 ( z d ) F 0 ( z d )
where f 0 denotes the pdf of N ( 0 , 1 ) . Hence,
E S 1 ( ψ a ; Y , Δ ) = ( 1 q ) ψ 1 a + ψ 2 a E T { Y | Y > d } + q f 0 ( z d ) F 0 ( z d )
and
E S 2 ( ψ a ; Y , Δ ) = ( 1 q ) 1 ψ 2 a ψ 1 a E T { Y | Y > d } ψ 2 a E T { Y 2 | Y > d } + q d f 0 ( z d ) F 0 ( z d )
where E T { · } denotes the expectation under the true data-generating model.
Solving the system of equations given by E S 1 ( ψ a ; Y , Δ ) = 0 and E S 2 ( ψ a ; Y , Δ ) = 0 determines the values ψ a * of the estimands in this analysis. Algebraic manipulations of this system show that
ψ 1 a = 1 ψ 2 a + ψ 2 a d E T { Y | Y > d } E T { Y 2 | Y > d } E T { Y | Y > d } d ,
which can be combined with the two estimating equations to solve for ψ a * ; the conditional expectations can be evaluated numerically.

Appendix B. Robust Asymptotic Variance of Parameter Estimators in Misspecified Location-Scale Models with Left-Censored Data

Let ψ ^ a be the estimate of ψ under the assumed location-scale model with limiting value ψ a * . In this appendix, we derive the robust (“true”) variance
Var T ( ψ ^ a ) = 1 n A 1 ( ψ a * ) B ( ψ a * ) A 1 ( ψ a * )
of ψ ^ a [21] under the assumption of a normal model, where
A ( ψ a * ) = E T ψ a S ( ψ a ; Y , Δ ) | ψ a = ψ a *
and
B ( ψ a * ) = E T S ( ψ a ; Y , Δ ) S ( ψ a ; Y , Δ ) | ψ a = ψ a * .
Firstly, it is straightforward to show that f 0 ( u ) = u f 0 ( u ) and f 0 ( u ) = f 0 ( u ) ( u 2 1 ) where f 0 is the density of a standard normal distribution. Thus, for A ( ψ a ) , we have
ψ 1 a S 1 ( ψ a ; Y , Δ ) = Δ ( 1 Δ ) f 0 ( z d ) F 0 ( z d ) f 0 2 ( z d ) F 0 2 ( z d ) , ψ 2 a S 2 ( ψ a ; Y , Δ ) = Δ 1 ψ 2 a 2 + Y 2 ( 1 Δ ) d 2 f 0 ( z d ) F 0 ( z d ) f 0 2 ( z d ) F 0 2 ( z d ) ,
and
ψ 2 a S 1 ( ψ a ; Y , Δ ) = Δ Y ( 1 Δ ) d f 0 ( z d ) F 0 ( z d ) f 0 2 ( z d ) F 0 2 ( z d ) ,
so
[ A ( ψ a ) ] 11 = 1 q q f 0 ( z d ) F 0 ( z d ) f 0 2 ( z d ) F 0 2 ( z d ) , [ A ( ψ a ) ] 22 = ( 1 q ) 1 ψ 2 a 2 + E T { Y 2 | Y > d } q d 2 f 0 ( z d ) F 0 ( z d ) f 0 2 ( z d ) F 0 2 ( z d ) ,
and
[ A ( ψ a ) ] 12 = [ A ( ψ a ) ] 21 = ( 1 q ) E T { Y | Y > d } q d f 0 ( z d ) F 0 ( z d ) f 0 2 ( z d ) F 0 2 ( z d ) .
Now, for B ( ψ a ) , note that Δ = Δ 2 , ( 1 Δ ) = ( 1 Δ ) 2 , and Δ ( 1 Δ ) = 0 , so
S 1 2 ( ψ a ; Y , Δ ) = Δ ( ψ 1 a + ψ 2 a Y ) 2 + ( 1 Δ ) f 0 ( z d ) F 0 ( z d ) 2 , S 2 2 ( ψ a ; Y , Δ ) = Δ 1 ψ 2 a Y ( ψ 1 a + ψ 2 a Y ) 2 + ( 1 Δ ) d 2 f 0 ( z d ) F 0 ( z d ) 2 ,
and
S 1 ( ψ a ; Y , Δ ) S 2 ( ψ a ; Y , Δ ) = Δ ( ψ 1 a + ψ 2 a Y ) 1 ψ 2 a Y ( ψ 1 a + ψ 2 a Y ) + ( 1 Δ ) d f 0 ( z d ) F 0 ( z d ) 2 = Δ ψ 1 a ψ 2 a + ( 1 ψ 1 a 2 ) Y 2 ψ 1 a ψ 2 a Y 2 ψ 2 a 2 Y 3 + ( 1 Δ ) d f 0 ( z d ) F 0 ( z d ) 2 .
Taking the expectations of these expressions and simplifying algebraically results in
[ B ( ψ a ) ] 11 = ( 1 q ) ψ 1 a 2 + 2 ψ 1 a ψ 2 a E 1 + ψ 2 a 2 E 2 + q f 0 ( z d ) F 0 ( z d ) 2 , [ B ( ψ a ) ] 22 = ( 1 q ) 1 ψ 2 a 2 2 ψ 1 a ψ 2 a E 1 + ( ψ 1 a 2 2 ) E 2 + 2 ψ 1 a ψ 2 a E 3 + ψ 2 a 2 E 4 + q d 2 f 0 ( z d ) F 0 ( z d ) 2 ,
and
[ B ( ψ a ) ] 12 = [ B ( ψ a ) ] 21 = q d f 0 ( z d ) F 0 ( z d ) 2 ( 1 q ) ψ 1 a ψ 2 a + ( 1 ψ 1 a 2 ) E 1 2 ψ 1 a ψ 2 a E 2 ψ 2 a 2 E 3 ,
where we now let E k denote the conditional expectation E T { Y k | Y > d } for notational convenience. Plugging ψ a = ψ a * from Appendix A into the equations above yields A ( ψ a * ) and B ( ψ a * ) .

Appendix C. Assumed Asymptotic Variance of Parameter Estimators in Misspecified Location-Scale Models with Left-Censored Data

As in Appendix B, let ψ ^ a be the estimate of ψ under the assumed location-scale model with limiting value ψ a * . In this appendix, we derive the variance Var a ( ψ ^ a ) of ψ ^ a under the assumed model. Using the delta method [20], we can show
Var a ( ψ ^ a ) G ( ϕ a * ) Var a ( ϕ ^ a ) G ( ϕ a * )
where ϕ a * = ( ϕ 1 a * , ϕ 2 a * ) = ( θ 1 a * , log θ 2 a * ) , ϕ ^ a is the estimate of ϕ = ( θ 1 , log θ 2 ) under the assumed model,
G ( ϕ ) = e ϕ 2 ϕ 1 e ϕ 2 0 e ϕ 2 ,
and Var a 1 ( ϕ ^ a ) is the expected information with respect to ϕ a evaluated at ϕ a = ϕ a * . Based on (A1), the log-likelihood parameterized by ϕ a can be written as
( ϕ a ; y , δ ) = i = 1 n δ i ϕ 2 a + 1 2 y i ϕ 1 a exp ( ϕ 2 a ) 2 + ( 1 δ i ) log F 0 ( z d )
where z d = ( d ϕ 1 a ) / exp ( ϕ 2 a ) . It is straightforward to show z d / ϕ 1 a = 1 / exp ( ϕ 2 a ) and z d / ϕ 2 a = z d . As before, we omit subscripts i to denote the random vector S i ( ϕ a ; Y i , Δ i ) of score functions for a single observation by S ( ϕ a ; Y , Δ ) . Then,
S 1 ( ϕ a ; Y , Δ ) = ϕ 1 a = Δ Y ϕ 1 a [ exp ( ϕ 2 a ) ] 2 1 Δ exp ( ϕ 2 a ) f 0 ( z d ) F 0 ( z d )
and
S 2 ( ϕ a ; Y , Δ ) = ϕ 2 a = Δ Y ϕ 1 a exp ( ϕ 2 a ) 2 1 + ( 1 Δ ) f 0 ( z d ) F 0 ( z d ) ,
so
ϕ 1 a S 1 ( ϕ a ; Y , Δ ) = 1 [ exp ( ϕ 2 a ) ] 2 Δ + ( 1 Δ ) f 0 ( z d ) z d F 0 ( z d ) + f 0 ( z d ) F 0 2 ( z d ) , ϕ 2 a S 2 ( ϕ a ; Y , Δ ) = 2 Δ Y ϕ 1 a exp ( ϕ 2 a ) 2 + ( 1 Δ ) z d f 0 ( z d ) ( z d 2 1 ) F 0 ( z d ) + z d f 0 ( z d ) F 0 2 ( z d ) ,
and
ϕ 2 a S 1 ( ϕ a ; Y , Δ ) = 1 exp ( ϕ 2 a ) 2 Δ Y ϕ 1 a exp ( ϕ 2 a ) + ( 1 Δ ) f 0 ( z d ) ( z d 2 1 ) F 0 ( z d ) + z d f 0 ( z d ) F 0 2 ( z d ) .
Thus
Var a 1 ( ϕ ^ a ) 11 = 1 [ exp ( ϕ 2 a ) ] 2 1 q + q f 0 ( z d ) z d F 0 ( z d ) + f 0 ( z d ) F 0 2 ( z d ) , Var a 1 ( ϕ ^ a ) 22 = 2 ( 1 q ) [ exp ( ϕ 2 a ) ] 2 E 2 2 ϕ 1 a E 1 + ϕ 1 a 2 + q z d f 0 ( z d ) ( z d 2 1 ) F 0 ( z d ) + z d f 0 ( z d ) F 0 2 ( z d ) ,
and
[ Var a 1 ( ϕ ^ a ) ] 12 = Var a 1 ( ϕ ^ a ) 21 = 1 exp ( ϕ 2 a ) 2 ( 1 q ) E 1 ϕ 1 a exp ( ϕ 2 a ) + q f 0 ( z d ) ( z d 2 1 ) F 0 ( z d ) + z d f 0 ( z d ) F 0 2 ( z d ) .
As in Appendix B, E k denotes the conditional expectation E T { Y k | Y > d } .

Appendix D. Testing Goodness of Fit Under LLD

Here, we consider a small simulation study to investigate the finite sample performance of the Shapiro–Francia test for normality under left-censoring. We conduct this test using the gofTestCensored function in the EnvStats package of R version 4.3.2 [32]. Specifically, suppose X = ( X 1 , , X n ) are the order statistics of a distribution with c.d.f. F and interest lies in testing the null hypothesis that F is the c.d.f. of an arbitrary normal distribution against the alternative that it is the c.d.f. of some other distribution. In the absence of censoring, this can be done using the Shapiro–Francia statistic [24], which is defined as
W = i = 1 n a i X i 2 i = 1 n ( X i X ¯ ) 2
where a i is a standardized version of m i , the expectation of the ith order statistic for a random sample of size n from the standard normal distribution; that is, if m = ( m 1 , , m n ) with a = ( a 1 , , a n ) , then a = m / m m . Note that a a = 1 , and since the normal distribution is symmetric, then the averages m ¯ and a ¯ satisfy m ¯ = 0 and a ¯ = 0 . Using these facts, it is straightforward to show that W = r 2 ( a , X ) where
r ( u , v ) = i = 1 n ( u i u ¯ ) ( v i v ¯ ) i = 1 n ( u i u ¯ ) 2 i = 1 n ( v i v ¯ ) 2
denotes the sample Pearson correlation coefficient of two length-n vectors u and v. This W statistic can be approximated by replacing m i in a with the Blom score
m ˜ i = Φ 1 i 3 / 8 N + 1 / 4
for i = 1 , , n [33], where Φ denotes the c.d.f. of the standard normal distribution, resulting in W ˜ = r 2 ( b , X ) where b = m ˜ / m ˜ m ˜ [34]. To adapt W ˜ to left-censored data, we suppose k observations in a realization x of X have been censored, and let x = ( x k + 1 , , x n ) be the subset of x that has been observed exactly. We further let b = ( b k + 1 , , b n ) . Now, if F is the c.d.f. of a normal distribution, then the standardized expected order statistics a will be highly linearly correlated with x, and, thus, so will the approximation b of a. In this case, r ( b , x ) 1 r ( b , x ) , and we can simply use
w ˜ = r 2 ( b , x ) ;
this seems to be the argument invoked by Royston [35], whom the authors of gofTestCensored cite to justify this step [32]. To compute p-values based on (A2), Royston [35] explores transformations Z = g ( W ˜ ) that are approximately normal. The parameters μ Z and σ Z of this normal distribution are estimated by regressing z α = g ( w ˜ α ) on Φ 1 ( α ) for α = 0.9 , 0.95 , 0.99 where the quantiles w ˜ α of W ˜ have been approximated via simulation [36], and we compute
p - value = 1 Φ z μ ^ Z σ ^ Z
where z is the observed value of Z [35].
To generate data for this simulation study, we use three location-scale distributions with location parameter θ 1 and scale parameter θ 2 : normal; logistic with the c.d.f. given by (2); and extreme value with the c.d.f. given by (3). For each of these models, we take θ 2 = 1 , and, given our goal of inference in the upper tail, we choose θ 1 such that the 95th percentile equals that of a standard normal. We specify the LLD via a fixed left-censoring rate in each scenario, leading to an LLD of 1.28 for the normal data, 3.50 for the logistic data, and 1.70 for the extreme value data when the left-censoring rate is 10%; and an LLD of 0.67 for the normal data, 2.40 for the logistic data, and 0.70 for the extreme value data when the left-censoring rate is 25%. The results of these simulations are shown in Table A1.
Table A1. Empirical rejection rates at the 0.05 significance level for the Shapiro–Francia test of normality under various left-censoring rates, based on 2000 simulated samples of size n. All Monte Carlo standard errors for the rejection rate estimates are < 0.001 .
Table A1. Empirical rejection rates at the 0.05 significance level for the Shapiro–Francia test of normality under various left-censoring rates, based on 2000 simulated samples of size n. All Monte Carlo standard errors for the rejection rate estimates are < 0.001 .
Left-Censoring RatenData-Generation Model
NormalLogisticExtreme Value
0%200.0480.1160.332
400.0560.1820.582
600.0560.2150.768
800.0580.2380.879
10%200.0470.1170.124
400.0560.1600.224
600.0560.2090.346
800.0500.2290.491
25%200.0470.1120.080
400.0600.1640.124
600.0590.2220.180
800.0560.2520.264
With data generated under the logistic distribution, we see that the Shapiro–Francia test can have very low power with moderately large sample sizes, even in the absence of left-censoring. On the other hand, with extreme value data generation, power is initially reasonably good for the larger sample sizes but can be severely compromised by even modest levels of left-censoring; this may be due to the fact that much of the substantial difference between the normal and extreme value distributions is found in the tails.

References

  1. United States Food and Drug Administration. Considerations for the Development of Dried Plasma Products Intended for Transfusion: Guidance for Industry; Docket Number FDA-2018-D-3759; United States Food and Drug Administration: Silver Spring, MD, USA, 2019. Available online: https://www.fda.gov/regulatory-information/search-fda-guidance-documents/considerations-development-dried-plasma-products-intended-transfusion (accessed on 17 October 2024).
  2. United States Food and Drug Administration. Submission and Review of Sterility Information in Premarket Notification (510(k)) Submissions for Devices Labeled as Sterile: Guidance for Industry and Food and Drug Administration Staff; Docket Number FDA-2008-D-0611; United States Food and Drug Administration: Silver Spring, MD, USA, 2024. Available online: https://www.fda.gov/regulatory-information/search-fda-guidance-documents/submission-and-review-sterility-information-premarket-notification-510k-submissions-devices-labeled (accessed on 17 October 2024).
  3. United States Food and Drug Administration. Approaches to Establish Thresholds for Major Food Allergens and for Gluten in Food; United States Food and Drug Administration: Silver Spring, MD, USA, 2006. Available online: https://www.fda.gov/food/food-labeling-nutrition/approaches-establish-thresholds-major-food-allergens-and-gluten-food (accessed on 17 October 2024).
  4. Government of Ontario. R.R.O. 1990, Reg. 347: General—Waste Management. In Environmental Protection Act; Government of Ontario: Toronto, ON, Canada, 2024. Available online: https://www.ontario.ca/laws/regulation/900347 (accessed on 17 October 2024).
  5. Newman, M.C.; Dixon, P.M.; Looney, B.B.; Pinder, J.E., III. Estimating Mean and Variance for Environmental Samples with Below Detection Limit Observations. JAWRA J. Am. Water Resour. Assoc. 1989, 25, 905–916. [Google Scholar] [CrossRef]
  6. Helsel, D.R. Much Ado About Next to Nothing: Incorporating Nondetects in Science. Ann. Occup. Hyg. 2010, 54, 257–262. [Google Scholar] [CrossRef] [PubMed]
  7. Hwang, M.; Lee, S.C.; Park, J.-H.; Choi, J.; Lee, H.-J. Statistical methods for handling nondetected results in food chemical monitoring data to improve food risk assessments. Food Sci. Nutr. 2023, 11, 5223–5235. [Google Scholar] [CrossRef] [PubMed]
  8. Gilliom, R.J.; Helsel, D.R. Estimation of Distributional Parameters for Censored Trace Level Water Quality Data: 1. Estimation Techniques. Water Resour. Res. 1986, 22, 135–146. [Google Scholar] [CrossRef]
  9. Helsel, D.R.; Cohn, T.A. Estimation of Descriptive Statistics for Multiply Censored Water Quality Data. Water Resour. Res. 1988, 24, 1997–2004. [Google Scholar] [CrossRef]
  10. Farewell, V.T. Some comments on analysis techniques for censored water quality data. Environ. Monit. Assess. 1989, 13, 285–294. [Google Scholar] [CrossRef]
  11. Shumway, R.H.; Azari, R.S.; Kayhanian, M. Statistical Approaches to Estimating Mean Water Quality Concentrations with Detection Limits. Environ. Sci. Technol. 2002, 36, 3345–3353. [Google Scholar] [CrossRef]
  12. Antweiler, R.C.; Taylor, H.E. Evaluation of statistical treatments of left-censored environmental data using coincident uncensored data sets: I. Summary statistics. Environ. Sci. Technol. 2008, 42, 3732–3738. [Google Scholar] [CrossRef]
  13. Shoari, N.; Dubé, J.-S.; Chenouri, S. On the use of the substitution method in left-censored environmental data. Hum. Ecol. Risk Assess. Int. J. 2015, 22, 435–446. [Google Scholar] [CrossRef]
  14. Tekindal, M.A.; Erdoğan, B.D.; Yavuz, Y. Evaluating Left-Censored Data Through Substitution, Parametric, Semi-parametric, and Nonparametric Methods: A Simulation Study. Interdiscip. Sci. Comput. Life Sci. 2017, 9, 153–172. [Google Scholar] [CrossRef]
  15. Kaplan, E.L.; Meier, P. Nonparametric Estimation from Incomplete Observations. J. Am. Stat. Assoc. 1958, 53, 457–481. [Google Scholar] [CrossRef]
  16. Sprott, D.A. Statistical Inference in Science; Springer: New York, NY, USA, 2000. [Google Scholar] [CrossRef]
  17. Huynh, T.; Quick, H.; Ramachandran, G.; Banerjee, S.; Stenzel, M.; Sandler, D.P.; Engel, L.S.; Kwok, R.K.; Blair, A.; Stewart, P.A. A Comparison of the β-Substitution Method and a Bayesian Method for Analyzing Left-Censored Data. Ann. Occup. Hyg. 2016, 60, 56–73. [Google Scholar] [CrossRef]
  18. Lawless, J.F. Statistical Models and Methods for Lifetime Data, 2nd ed.; John Wiley & Sons: Hoboken, NJ, USA, 2003. [Google Scholar] [CrossRef]
  19. Therneau, T.M.; Grambsch, P.M. Modeling Survival Data: Extending the Cox Model; Springer: New York, NY, USA, 2000. [Google Scholar]
  20. Casella, G.; Berger, R.L. Statistical Inference, 2nd ed.; Duxbury: Pacific Grove, CA, USA, 2002. [Google Scholar]
  21. White, H. Maximum likelihood estimation of misspecified models. Econometrica 1982, 50, 1–25. [Google Scholar] [CrossRef]
  22. Gourieroux, C.; Monfort, A.; Trognon, A. Pseudo Maximum Likelihood Methods: Theory. Econometrica 1984, 52, 681–700. [Google Scholar] [CrossRef]
  23. Qin, J.; Lawless, J. Empirical Likelihood and General Estimating Equations. Ann. Stat. 1994, 22, 300–325. [Google Scholar] [CrossRef]
  24. Shapiro, S.S.; Francia, R.S. An Approximate Analysis of Variance Test for Normality. J. Am. Stat. Assoc. 1972, 67, 215–216. [Google Scholar] [CrossRef]
  25. Berry, G.; Armitage, P. Mid-P confidence intervals: A brief review. J. R. Stat. Soc. Ser. D Stat. 1995, 44, 417–423. [Google Scholar] [CrossRef]
  26. Lawless, J.F.; Zhan, M. Analysis of interval-grouped recurrent event data using piecewise-constant rate functions. Can. J. Stat. 1998, 26, 549–565. [Google Scholar] [CrossRef]
  27. SAS Institute Inc. Reading Spec Limits from an Input Data Set; SAS Institute Inc.: Cary, NC, USA, n.d.; Available online: https://documentation.sas.com/doc/en/pgmsascdc/9.4_3.4/qcug/qcug_code_capspec2.htm (accessed on 11 January 2025).
  28. Fauvernier, M.; Remontet, L.; Uhry, Z.; Bossard, N.; Roche, L. survPen: An R package for hazard and excess hazard modelling with multidimensional penalized splines. J. Open Source Softw. 2019, 4, 1434. [Google Scholar] [CrossRef]
  29. Wigle, A.; Béliveau, A.; Blackmore, D.; Lapeyre, P.; Osadetz, K.; Lemieux, C.; Daun, K.J. Estimation and Applications of Uncertainty in Methane Emissions Quantification Technologies: A Bayesian Approach. ACS ES T Air 2024, 1, 1000–1014. [Google Scholar] [CrossRef]
  30. Arjas, E.; Gasbarra, D. Nonparametric Bayesian Inference from Right Censored Survival Data, Using the Gibbs Sampler. Stat. Sin. 1994, 4, 505–524. [Google Scholar]
  31. Ghosh, J.K.; Delampady, M.; Samanta, T. An Introduction to Bayesian Analysis: Theory and Methods, 1st ed.; Springer: New York, NY, USA, 2006. [Google Scholar]
  32. Millard, S.P. EnvStats: An R Package for Environmental Statistics; Springer: New York, NY, USA, 2013. [Google Scholar]
  33. Blom, G. Statistical Estimates and Transformed Beta-Variables; John Wiley & Sons: Hoboken, NJ, USA, 1958. [Google Scholar]
  34. Weisberg, S.; Bingham, C. An Approximate Analysis of Variance Test for Non-Normality Suitable for Machine Calculation. Technometrics 1975, 17, 133–134. [Google Scholar] [CrossRef]
  35. Royston, P. A Toolkit for Testing for Non-Normality in Complete and Censored Samples. J. R. Stat. Soc. Ser. D Stat. 1993, 42, 37–43. [Google Scholar] [CrossRef]
  36. Verrill, S.; Johnson, R.A. Tables and Large-Sample Distribution Theory for Censored-Data Correlation Statistics for Testing Normality. J. Am. Stat. Assoc. 1988, 83, 1192–1197. [Google Scholar] [CrossRef]
Figure 1. Asymptotic bias of estimates for location parameter θ 1 and log-scale parameter log ( θ 2 ) as a function of the left-censoring rate q under logistic (left panels) and extreme value (right panels) error distributions. The red dashed line represents the reference value of zero asymptotic bias.
Figure 1. Asymptotic bias of estimates for location parameter θ 1 and log-scale parameter log ( θ 2 ) as a function of the left-censoring rate q under logistic (left panels) and extreme value (right panels) error distributions. The red dashed line represents the reference value of zero asymptotic bias.
Mathematics 13 01274 g001
Figure 2. Asymptotic bias of estimates for standardized threshold value ( τ θ 1 ) / θ 2 and reliability F ( τ ) as a function of the left-censoring rate q under logistic (left panels) and extreme value (right panels) error distributions. The red dashed line represents the reference value of zero asymptotic bias.
Figure 2. Asymptotic bias of estimates for standardized threshold value ( τ θ 1 ) / θ 2 and reliability F ( τ ) as a function of the left-censoring rate q under logistic (left panels) and extreme value (right panels) error distributions. The red dashed line represents the reference value of zero asymptotic bias.
Mathematics 13 01274 g002
Figure 3. Asymptotic type I error rate for a hypothesis test of H 0 : p 1 = 0.95 vs. H A : p 1 > 0.95 as a function of the left-censoring rate q under logistic and extreme value (EVD) data generation. The red dashed lines indicate the nominal type I error rate of α = 0.025 . Sample size is n = 80 .
Figure 3. Asymptotic type I error rate for a hypothesis test of H 0 : p 1 = 0.95 vs. H A : p 1 > 0.95 as a function of the left-censoring rate q under logistic and extreme value (EVD) data generation. The red dashed lines indicate the nominal type I error rate of α = 0.025 . Sample size is n = 80 .
Mathematics 13 01274 g003
Figure 4. Asymptotic power for hypothesis tests of H 0 : p 1 = 0.95 vs. H A : p 1 > 0.95 as a function of the left-censoring rate q under normal, logistic, and extreme value (EVD) data generation. Sample size is n = 80 .
Figure 4. Asymptotic power for hypothesis tests of H 0 : p 1 = 0.95 vs. H A : p 1 > 0.95 as a function of the left-censoring rate q under normal, logistic, and extreme value (EVD) data generation. Sample size is n = 80 .
Mathematics 13 01274 g004
Figure 5. (a) Minimum sample size n needed to achieve worst-case empirical power of 90% across normal, logistic, and extreme value data-generating distributions, using a piecewise-constant hazard-based model with two cut-points b 0 and b 1 at the specified empirical quantiles; and (b) worst-case empirical type I error rate across those distributions when the sample size in each case is the value n shown in (a). Tests are conducted at the 0.025 significance level with p 0 = 0.95 and p 1 = 0.99 , and each empirical rejection rate is obtained using 2000 simulations. Green cells indicate more desirable configurations of sample size and control of type I error while red cells indicate configurations with impractically large sample sizes or unacceptably elevated type I error rates.
Figure 5. (a) Minimum sample size n needed to achieve worst-case empirical power of 90% across normal, logistic, and extreme value data-generating distributions, using a piecewise-constant hazard-based model with two cut-points b 0 and b 1 at the specified empirical quantiles; and (b) worst-case empirical type I error rate across those distributions when the sample size in each case is the value n shown in (a). Tests are conducted at the 0.025 significance level with p 0 = 0.95 and p 1 = 0.99 , and each empirical rejection rate is obtained using 2000 simulations. Green cells indicate more desirable configurations of sample size and control of type I error while red cells indicate configurations with impractically large sample sizes or unacceptably elevated type I error rates.
Mathematics 13 01274 g005
Figure 6. Empirical c.d.f. of the log-can weight distribution. Open and closed circles, respectively, denote observations below and above the LLD. The dashed blue line shows the target reliability level of p 0 = 0.95 , while the dashed red lines show the non-parametric reliability estimate p ^ 1 .
Figure 6. Empirical c.d.f. of the log-can weight distribution. Open and closed circles, respectively, denote observations below and above the LLD. The dashed blue line shows the target reliability level of p 0 = 0.95 , while the dashed red lines show the non-parametric reliability estimate p ^ 1 .
Mathematics 13 01274 g006
Figure 7. Fitted c.d.f.s of the log-can weight distribution based on fully parametric and piecewise-constant hazard-based models. Dotted black lines show 95% CIs for the fully parametric models. Each piecewise-constant hazard-based model is specified by the vector of probabilities p defining the empirical percentiles Q ^ p used as its cut-points; the vertical dashed grey lines show the values of these cut-points. The horizontal dashed blue lines show the target reliability level of p 0 = 0.95 , while the dashed red lines show the reliability estimates p ^ 1 based on the fitted models.
Figure 7. Fitted c.d.f.s of the log-can weight distribution based on fully parametric and piecewise-constant hazard-based models. Dotted black lines show 95% CIs for the fully parametric models. Each piecewise-constant hazard-based model is specified by the vector of probabilities p defining the empirical percentiles Q ^ p used as its cut-points; the vertical dashed grey lines show the values of these cut-points. The horizontal dashed blue lines show the target reliability level of p 0 = 0.95 , while the dashed red lines show the reliability estimates p ^ 1 based on the fitted models.
Mathematics 13 01274 g007
Figure 8. Probability plots of the log-can weight distribution based on fully parametric models. Dashed grey lines show the estimated 5th through 95th percentiles of the distribution (horizontal) and the corresponding can weights (vertical). Dotted black lines show 95% CIs. The dashed blue line shows the target reliability level of p 0 = 0.95 , while the dashed red lines show the reliability estimates p ^ 1 .
Figure 8. Probability plots of the log-can weight distribution based on fully parametric models. Dashed grey lines show the estimated 5th through 95th percentiles of the distribution (horizontal) and the corresponding can weights (vertical). Dotted black lines show 95% CIs. The dashed blue line shows the target reliability level of p 0 = 0.95 , while the dashed red lines show the reliability estimates p ^ 1 .
Mathematics 13 01274 g008
Table 1. Estimates of F ( τ ) obtained under 10% left-censoring based on 2000 simulated samples of size n.
Table 1. Estimates of F ( τ ) obtained under 10% left-censoring based on 2000 simulated samples of size n.
F ( τ ) Data-Generating ModelnNormalPiecewise-Constant 1Exact Test
Q ^ 90 Q ^ 85 , 90 , 95 Q ^ 85 , 89 , 93
0.95Normal200.9490.9570.955 0.949
400.9500.9510.9510.9500.949
600.9500.9490.9500.9460.949
800.9490.9470.9500.9470.949
Logistic200.9510.9530.955 0.950
400.9500.9430.9490.9450.949
600.9510.9380.9480.9400.950
800.9510.9350.9470.9410.950
Extreme Value200.9250.9590.957 0.951
400.9230.9540.9520.9520.950
600.9230.9520.9510.9480.949
800.9230.9520.9510.9490.949
0.97Normal200.9680.9720.972 0.969
400.9690.9700.9700.9700.969
600.9690.9690.9690.9680.969
800.9690.9680.9690.9680.969
Logistic200.9730.9690.973 0.970
400.9730.9620.9680.9660.969
600.9740.9580.9660.9610.969
800.9740.9560.9650.9600.970
Extreme Value200.9420.9740.974 0.970
400.9410.9720.9710.9720.969
600.9410.9710.9700.9710.969
800.9410.9710.9700.9710.969
0.99Normal200.9880.9900.990 0.990
400.9890.9900.9890.9900.990
600.9890.9900.9890.9910.990
800.9890.9900.9890.9900.989
Logistic200.9930.9890.991 0.990
400.9940.9870.9890.9880.990
600.9940.9860.9880.9870.990
800.9950.9860.9870.9860.990
Extreme Value200.9640.9890.990 0.991
400.9640.9890.9900.9900.990
600.9640.9890.9900.9910.990
800.9640.9900.9900.9900.990
1 Column heading indicates the p 1 , , p K empirical percentiles Q ^ p 1 , , p K defining the interval cut-points of the fitted piecewise-constant hazard-based model.
Table 2. Empirical null hypothesis rejection rates at the α = 0.025 significance level, testing for 95% reliability under 10% left-censoring, based on 2000 simulated samples of size n.
Table 2. Empirical null hypothesis rejection rates at the α = 0.025 significance level, testing for 95% reliability under 10% left-censoring, based on 2000 simulated samples of size n.
F ( τ ) Data-Generating ModelnNormalPiecewise-Constant 1Exact Test
Q ^ 90 Q ^ 85 , 90 , 95 Q ^ 85 , 89 , 93 p-Valuemid-p
0.95Normal200.0390.0350.015 0.0000.000
400.0290.0310.0420.0480.0000.000
600.0310.0260.0350.0530.0000.045
800.0260.0230.0320.0370.0200.020
Logistic200.0480.0590.017 0.0000.000
400.0490.0470.0440.0560.0000.000
600.0500.0420.0420.0580.0000.050
800.0580.0280.0300.0330.0210.021
Extreme Value200.0020.0270.018 0.0000.000
400.0000.0220.0300.0390.0000.000
600.0010.0230.0290.0440.0000.048
800.0000.0190.0290.0340.0170.017
0.97Normal200.1000.0720.032 0.0000.000
400.1240.0870.0770.0900.0000.000
600.1770.1130.0850.1370.0000.153
800.2250.1250.1110.1280.0900.090
Logistic200.1670.1320.037 0.0000.000
400.2410.1690.1100.1590.0000.000
600.3400.1550.1190.1660.0000.164
800.4050.1540.1200.1440.0930.093
Extreme Value200.0050.0590.023 0.0000.000
400.0030.0640.0620.0800.0000.000
600.0020.0890.0760.1290.0000.156
800.0000.1080.0980.1160.0780.078
0.99Normal200.3550.1920.055 0.0000.000
400.6390.3740.2530.3640.0000.000
600.8480.5240.4060.5570.0000.538
800.9240.6200.5240.6040.4290.429
Logistic200.5830.3390.112 0.0000.000
400.8460.5390.3830.4860.0000.000
600.9580.6030.4740.5800.0000.538
800.9890.6480.5410.6080.4480.448
Extreme Value200.0290.1320.044 0.0000.000
400.0350.2730.2000.2770.0000.000
600.0480.4350.3640.5150.0000.557
800.0720.5900.5420.6030.4590.459
1 Column heading indicates the p 1 , , p K empirical percentiles Q ^ p 1 , , p K defining the interval cut-points of the fitted piecewise-constant hazard-based model.
Table 3. Reliability estimates at threshold τ = log ( 1.12 ) and corresponding confidence/likelihood intervals obtained for the distribution of log-can weights using fully parametric and piecewise-constant hazard-based models.
Table 3. Reliability estimates at threshold τ = log ( 1.12 ) and corresponding confidence/likelihood intervals obtained for the distribution of log-can weights using fully parametric and piecewise-constant hazard-based models.
Model 1 F ^ ( τ ) Confidence/Likelihood Interval 2
Normal0.961 ( 0.920 , 0.983 )
Logistic0.951 ( 0.839 , 0.986 )
Extreme Value0.982 ( 0.965 , 0.992 )
( Q ^ 90 ) 0.976 ( 0.952 , 1.000 )
( Q ^ 91 , Q ^ 99 ) 0.974 ( 0.949 , 1.000 )
( Q ^ 92 , Q ^ 98 ) 0.972 ( 0.922 , 0.994 )
( Q ^ 94 , Q ^ 97 ) 0.973 ( 0.881 , 0.993 )
( Q ^ 85 , Q ^ 93 , Q ^ 95 ) 0.975 ( 0.929 , 1.000 )
1 Each piecewise-constant hazard-based model (last five rows) is specified by the vector of empirical pth percentiles Q ^ p used as its cut-points. 2 This is a two-sided 95% confidence interval for the fully parametric models (first three rows) and a one-sided 97.5% likelihood interval for the piecewise-constant hazard-based models (last five rows).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bumbulis, L.S.; Cook, R.J. Robustness and Efficiency Considerations When Testing Process Reliability with a Limit of Detection. Mathematics 2025, 13, 1274. https://doi.org/10.3390/math13081274

AMA Style

Bumbulis LS, Cook RJ. Robustness and Efficiency Considerations When Testing Process Reliability with a Limit of Detection. Mathematics. 2025; 13(8):1274. https://doi.org/10.3390/math13081274

Chicago/Turabian Style

Bumbulis, Laura S., and Richard J. Cook. 2025. "Robustness and Efficiency Considerations When Testing Process Reliability with a Limit of Detection" Mathematics 13, no. 8: 1274. https://doi.org/10.3390/math13081274

APA Style

Bumbulis, L. S., & Cook, R. J. (2025). Robustness and Efficiency Considerations When Testing Process Reliability with a Limit of Detection. Mathematics, 13(8), 1274. https://doi.org/10.3390/math13081274

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop