Next Article in Journal
On Convergence Rate of MRetrace
Next Article in Special Issue
Mitigating Multicollinearity in Regression: A Study on Improved Ridge Estimators
Previous Article in Journal
Rough and T-Rough Sets Arising from Intuitionistic Fuzzy Ideals in BCK-Algebras
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Negative Binomial Regression via the Kibria–Lukman Strategy: Methodology and Application

by
Adewale F. Lukman
1,*,
Olayan Albalawi
2,
Mohammad Arashi
3,4,
Jeza Allohibi
5,
Abdulmajeed Atiah Alharbi
5 and
Rasha A. Farghali
6
1
Department of Mathematics and Statistics, University of North Dakota, Grand Forks, ND 58202, USA
2
Department of Statistics, Faculty of Science, University of Tabuk, Tabuk 47512, Saudi Arabia
3
Department of Statistics, Faculty of Mathematical Sciences, Ferdowsi University of Mashhad, Mashhad 9177948974, Razavi Khorasan, Iran
4
Department of Statistics, Faculty of Natural and Agricultural Sciences, University of Pretoria, Pretoria 0002, South Africa
5
Department of Mathematics, Faculty of Science, Taibah University, Al-Madinah Al-Munawara 42353, Saudi Arabia
6
Department of Mathematics, Insurance and Applied Statistics, Helwan University, Cairo 11795, Egypt
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(18), 2929; https://doi.org/10.3390/math12182929
Submission received: 13 August 2024 / Revised: 13 September 2024 / Accepted: 15 September 2024 / Published: 20 September 2024
(This article belongs to the Special Issue Application of Regression Models, Analysis and Bayesian Statistics)

Abstract

:
Count regression models, particularly negative binomial regression (NBR), are widely used in various fields, including biometrics, ecology, and insurance. Over-dispersion is likely when dealing with count data, and NBR has gained attention as an effective tool to address this challenge. However, multicollinearity among covariates and the presence of outliers can lead to inflated confidence intervals and inaccurate predictions in the model. This study proposes a comprehensive approach integrating robust and regularization techniques to handle the simultaneous impact of multicollinearity and outliers in the negative binomial regression model (NBRM). We investigate the estimators’ performance through extensive simulation studies and provide analytical comparisons. The simulation results and the theoretical comparisons demonstrate the superiority of the proposed robust hybrid KL estimator (M-NBKLE) with predictive accuracy and stability when multicollinearity and outliers exist. We illustrate the application of our methodology by analyzing a forestry dataset. Our findings complement and reinforce the simulation and theoretical results.

1. Introduction

Count regression models play an important role in modern applied statistics, with applications across fields such as biometrics, ecology, and insurance. These models are advantageous when the response variable, y, is a count data, as in studies that model the number of falls in medical research [1]. The relationship between the response variable y and a set of covariates x = 1 , x 1 , , x p T is often expressed as E y | x = g 1 x T β , where β = β 0 , β 1 , , β p T are the regression coefficients and g is the canonical link function within the framework of generalized linear model (GLM) [2]. This formulation provides a flexible approach to modeling count data while accounting for various covariates, making it widely applicable in many disciplines.
In recent years, the NBR has gained significant attention as an effective tool for modeling count data in the presence of overdispersion. The primary goal is to estimate β accurately. However, multicollinearity among covariates (predictors) can lead to inflated confidence intervals and inaccurate predictions for the response variable [2,3]. Outlying observations can also significantly impact the estimates, making it essential to develop robust strategies to eliminate their misleading effects [3]. Unfortunately, in the GLM, multicollinearity and outliers are often considered separately, leading to incomplete analyses that fail to address the simultaneous impact of multicollinearity and outliers on the model’s performance. Although there have been various studies addressing the issue of multicollinearity in the NBRM, such as the development of ridge and Liu estimators [4,5,6], and the Jackknifed Liu-type estimator [7], the robust inference in the presence of outliers is neglected.
Outliers are common occurrences in real-life applications and can often result in misleading results and conclusions in data science. Causes of outliers include human error, measurement error, and recording error, or can be simply inherent in the data. It is crucial to identify and remove such outliers using appropriate data-cleaning methods and employing methods that are not susceptible to outliers to ensure robustness in data science [8,9]. While considerable research has been conducted on robust regression modeling, relatively few studies have specifically addressed the issue of outliers in NBR [1,8,9,10]. For a comprehensive review of the robust inference for the GLMs, Medina and Ronchetti’s [10] work is a valuable resource.
Lukman et al. [11,12] pointed out the simultaneous impact of multicollinearity and outliers on the Poisson and linear regression models. Roozbeh et al. [13,14] developed numerical algorithms to solve nonlinear equations resulting from the estimation of regression coefficients in the presence of multicollinearity and outliers. While both issues have received detailed attention in linear and Poisson regression models, more focus is needed in the NB framework. To the best of our knowledge, there is currently no available research handling multicollinearity and outliers in the context of NBRMs at the time of writing.
The primary objective of this article is to examine the difficulties associated with multicollinearity and outliers in NBR modeling and propose solutions to overcome these obstacles. To achieve this, we will integrate robust and regularization techniques for NBR modeling to effectively address the issues of multicollinearity and outliers. The main contribution of this work is the introduction of a new estimator for the NBRM that combines the shrinkage estimator with the robust regression to account for both multicollinearity and outliers in NBR.
The rest of the paper is as follows. Section 2 reviews some existing methods and develops new estimators. Additionally, this section includes analytical comparisons, where we theoretically compare the proposed methods with existing alternatives using performance measures. In Section 3, an extensive simulation study is considered, where we intensively study the effect of multicollinearity and outliers on the NBRM. The study is supported by a real-life application in Section 4, followed by the conclusion in Section 5. We have included the code used for the real-life application to facilitate replication of our results in Supplementary Material.

2. Methodology

2.1. Negative Binomial Regression Model

NBR is a generalization of Poisson regression designed to handle overdispersion, where the variance of the response variable exceeds the mean. NBR provides a more flexible alternative to Poisson regression for count data that exhibit greater dispersion by introducing an additional parameter to model this extra variability. Given a sample of n observations y 1 , y 2 , , y n , the probability mass function (p.m.f.) of the NB distribution is expressed as the following equation:
f y i | x i = Γ γ 1 + y i Γ γ 1   Γ 1 + y i   γ 1 γ 1 + μ i γ 1 μ i γ 1 + μ i y i     y i = 0 , 1 , 2 , .   i = 1 , 2 , , n .
The overdispersion parameter γ is defined as ( γ = 1 τ   ). μ i denotes the mean of the distribution for the i-th observation and Γ (.) is the gamma function. The conditional mean and variance are calculated as follows:
E y i | μ i , τ i = μ i .
Var y i | μ i , τ i = μ i 1 + γ μ i .
The maximum likelihood (ML) method is commonly used to estimate the parameter vector β by maximizing the log-likelihood function. The mean parameter μ i is often modelled as a function of the explanatory variables xi through a log-link: l o g μ i = x i T β . Thus, the log-likelihood function is given as follows:
l β , γ = J = 1 n l o g Γ γ 1 + y i l o g Γ y i + 1 l o g Γ γ 1 + γ 1 l o g γ 1 γ 1 + μ i + y i l o g μ i γ 1 + μ i .
The score function, which results from taking the first derivative of the log-likelihood function to β is given as follows:
  l β , γ     β j   = i = 1 n y i μ i 1 + γ μ i x i j = 0    j = 1 , 2 , , p .
This system of equations is nonlinear and, as mentioned, does not have a closed-form solution. We employ numerical methods to find the parameter estimates. Specifically, we use iterative reweighted least squares (IRLS) to solve the nonlinear system. At each iteration r, the estimate β r + 1 is as follows:
β ^ r + 1 = β ^ r + I β ^ r 1   S β ^ r ,
where S β ^ r is the score function evaluated at the r-th iteration. I β r = E 2 l β , γ β     β = X W r X is a p × p Fisher information matrix and it is evaluated at β r and W r is a n × n diagonal matrix; the iterations will end when the difference | | β ^ r β ^ r + 1 | | < ϵ , ϵ is a certain small value that usually equals 10−6. Thus, the maximum likelihood estimates (NBMLE) are obtained iteratively as follows:
β ^ r + 1 = X W ^ r X 1 X W ^ r z r ,
where zr is n × 1 vector of adjusted response variable such that z i r = log μ ^ i r + y i μ ^ i r μ ^ i r ,   and   the   weight   W ^ r = d i a g μ ^ i r 1 + θ μ ^ i r . The final equation at convergence is
β ^ N B M L E = X W ^ X 1 X W ^ z ,
where z is the final adjusted response vector such that z i r = log μ ^ i + y i μ ^ i μ ^ i and W ^ = d i a g μ ^ i 1 + θ μ ^ i .   The var-covariance matrix can be obtained as follows:
Cov β ^ N B M L E = X W ^ X 1 .

2.2. Shrinkage Estimators

Biased estimators, such as the ridge regression estimator developed by Hoerl and Kennard [15], the Liu and Liu-type estimators [16,17], and the Kibria–Lukman estimator [18], are among those introduced to address the multicollinearity problem in linear regression. These estimators are also used in generalized linear models to account for multicollinearity in binary logistic regression [19,20], zero-inflated Poisson regression [21,22,23], Poisson regression [24,25,26,27], NBR [4,5,8,28,29] and Conway–Maxwell–Poisson [30], among others.
Månsson [4] introduced the ridge regression estimator (NBRRE) for the NBR as follows:
β ^ N B R R E = X W ^ X + k I 1 X W ^ X   β ^ N B M L E , k > 0 .
where k is the ridge parameter, defined in this study as: k = 1 max β ^ N B M L E 2 .
The asymptotic covariance matrix is calculated as follows:
Cov β ^ N B R R E = X W ^ X + k I 1 X W ^ X X W ^ X + k I 1 .  
Following the works of [15,16], Kibria and Lukman [18] developed the Kibria–Lukman estimator (KLE), a single-parameter biased estimator designed to address multicollinearity in linear regression models. They demonstrated that this estimator outperforms both the Ridge and Liu estimators in terms of estimation accuracy and stability. Thus, the KLE for the NBRM is given by:
β ^ N B K L E = X W ^ X + k I 1 X W ^ X k I   β ^ N B M L E , k > 0 .  
The asymptotic covariance matrix is calculated as follows:
Cov β ^ N B K L E = X W ^ X + k I 1 X W ^ X k I X W ^ X 1 X W ^ X k I X W ^ X + k I 1 .
Under certain conditions of their simulation study, they proved that the KL estimator outperforms both the ridge regression and Liu estimators.

2.3. Shrinkage-Robust Estimators

The occurrence of outliers is a prevalent issue in regression analysis. Outliers are observations that stand out from the rest of the data [31]. Outliers were divided into vertical outliers and good and bad leverage points [32,33]. Vertical outliers are those observations that are not outlying in the space of explanatory variables (the x-dimension) but have outlying values for the related error term (on the y-direction). Outlying observations in the explanatory variables’ space located near the regression line are good leverage points. Bad leverage points are observations that are both off-center in the explanatory variable space and distant from the actual regression line. Outliers exist in both ways (x-direction of the explanatory variables and y-direction of the response variable). Outliers can produce biased regression estimates and misleading regression inferences as well as overestimate the standard errors. As a result, while doing power analyses, overestimate the required sample size [1,34].
The most popular robust regression technique is M-estimation [35], which is almost as effective as the least squares estimator. The M-estimator minimizes the function of the standardized residual to minimize the sum of squared residuals as the goal. This approach performs better than the conventional least squares estimator because it assigns fewer weights to atypical observations. The letter M denotes that the estimation is of the maximum likelihood kind. This method allows for removing certain data, which is sometimes not the best course of action, especially if the material being removed is crucial. The M-estimator is obtained by minimizing the residual function as follows:
β ^ M = M i n β   i = 1 n ρ   y i β j x i j s ,  
where s is an estimate of scale from a linear combination of the residuals and the function ρ assigns the contribution of each residual to the objective function. A system of normal equations is obtained by taking the first partial derivatives with respect to β j   ,   j = 1 , 2 , , p and setting them equal to zero.
Aeberhard et al. [1] introduced robust estimators for the parameters of the NB distribution using the M-estimator. They achieved robustness in the y-direction by bounding the Person residuals r i = y i μ i V a r 1 / 2 y i | μ i , τ i in the response that appears in the score function and on the design by introducing weights w x i . Thus, the Mallows quasi-likelihood estimator is obtained as follows:
i = 1 n S β r i V a r 1 / 2 y i | μ i , τ i   μ i φ w x i x i E i β = 0 ,
where S . r i is a continuous and bounded function that depends on few tuning constants, w x i is a weight controlling the impact of possible leverage points, and E i β = E S β r i V a r 1 / 2 y i | μ i , τ i   μ i φ w x i x i   is an adjustment term guaranteeing Fisher consistency at the model. The choice of the S . r i   function that bounds the Pearson residual is crucial because the resulting estimator has a bounded effect function thanks to the S . r i function’s boundedness criterion. In this study, we consider Tukey’s bi-weight function that has the advantage of being tuned by a single constant. The Tukey weight is a weight function that can be used in robust regression methods, including M-estimators based on Pearson residuals, to down-weight influential observations. The Tukey weight function is given as follows:
S T u k e y r i , c = r i c 2 1 2   r i   0   r i > c   r i c ,   i = 1 , 2 , , n ,  
where r i is the Pearson residual for the i t h observation, i = 1 , 2 , , n , c is a tuning parameter that determines the cut-off point for the weight function, and S T u k e y r i , c is the weight assigned to i t h observation, i = 1 , 2 , , n . The Huber weight function and the Tukey weight function are comparable; however, the Tukey weight function provides a more seamless change from high weights to low weights for observations with significant Pearson residuals. The Tukey weight function gives observations with Pearson residuals less than or equal to a c weight of 1, and it gradually reduces that weight as the Pearson residuals rise over c . A weight of 0 is applied to observations with Pearson residuals higher than c .
Researchers have proposed robust-biased estimators as viable alternatives when dealing with a regression model that encounters both multicollinearity and outliers in linear and Poisson regression models. These estimators combined biased estimators with robust estimators such as M-estimators. For more in-depth information, refer to the following works [11,12,36,37,38,39,40,41].
Silvapulle [36] proposed a robust ridge regression estimator for a linear regression model as follows:
β ^ M R R E = X X + k I 1 X X   β ^ M O L S ,     k > 0 .  
Majid et al. [41] proposed a robust Kibria–Lukman estimator for a linear regression model as follows:
β ^ M K L E = X X + k I 1 X X k I   β ^ M O L S ,     k > 0 .  
Following the work of Silvapulle [36], Majid et al. [41], Lukman et al. [11], and Arum et al. [40], we extended the robust ridge regression estimator and the robust KL estimator to the NBRM; the new estimators were found by combining the ridge regression estimator and the KL estimator with the M-estimator instead of the maximum likelihood estimators as follows:
β ^ M N B R R E = X W ^ X + k I 1 X W ^ X   β ^ M M L E     k > 0 .  
β ^ M N B K L E = X W ^ X + k I 1 X W ^ X k I   β ^ M M L E     k > 0 .  
The asymptotic covariance matrices are calculated as follows:
C o v β ^ M N B R R E = X W ^ X + k I 1 ( X W ^ X ) V a r β ^ M M L E ( X W ^ X )   X W ^ X + k I 1 .  
C o v β ^ M N B K L E = X W ^ X + k I 1 X W ^ X k I   V a r β ^ M M L E X W ^ X k I X W ^ X + k I 1

2.4. Theoretical Comparisons between Estimators

For convenience, we use the spectral decomposition of the estimated weighted information matrix ( X W ^ X ) ; assume that there exists a matrix G such that G X W ^ X G = Λ = d i a g λ j , j = 1 , 2 , ,   p ,where λ 1 λ 2 λ p are the ordered eigenvalues of X W ^ X and G orthogonal Matrix of order ( p × p ), its columns are the corresponding eigenvectors for λ 1 , λ 2 , , λ p . Assume that H = X G ,   G X W ^ X G = H H = Λ   , then α ^ N B M L E = G β ^ N B M L E . Thus, Equations (8), (10), (12), (19) and (20) can be written in canonical form as follows:
α ^ N B M L E = Λ 1 H   W ^ z .  
α ^ N B R R E = Λ + k I 1 Λ   α ^ N B M L E .  
α ^ N B K L E = Λ + k I 1 Λ k I   α ^ N B M L E .  
α ^ M N B R R E = Λ + k I 1 Λ   α ^ M M L E .  
α ^ M N B K L E = Λ + k I 1 Λ k I   α ^ M M L E .
Specifically, we replace α ^ N B M L E , which is not robust, with the robust version proposed by Aeberhard et al. [1]. We have defined this robust version in our study as α ^ M M L E .
The performance of the ξ ˜   of the regression parameter ξ is evaluated using the scalar mean squared error (SMSE), which is defined as follows:
SMSE ξ ˜ = t r C o v ξ ˜ + b i a s   ξ ˜ b i a s ξ ˜ .
where C o v ξ ˜ represents the covariance matrix of ξ ˜   and b i a s ξ ˜ = E ξ ˜ ξ .
The SMSE of the mentioned estimators above are obtained as follows:
SMSE α ^ N B M L E   = j = 1 p 1 λ j
SMSE α ^ N B R R E = j = 1 p λ j λ j + k 2 + j = 1 p α j 2 k 2 λ j + k 2   .
SMSE α ^ N B K L E = j = 1 p λ j k 2 λ j + k 2 λ j + 4 j = 1 p α j 2 k 2 λ j + k 2   .  
SMSE α ^ M N B R R E = j = 1 p λ j 2 λ j + k 2 ψ j j + j = 1 p α j 2 k 2 λ j + k 2 .  
SMSE α ^ M N B K L E = j = 1 p λ j k 2 λ j + k 2 ψ j j + 4 j = 1 p α j 2 k 2 λ j + k 2 ,  
where α ^ M M L E = Ψ , and S M S E α ^ M M L E   = j = 1 p ψ j j . We assume the following conditions are satisfied:
(I)
Ψ is finite.
(II)
ψjj is skew-symmetric and nondecreasing.
(III)
The errors have a zero mean and finite variance.
Theorem 1. 
The estimator  α ^ M N B K L E  is superior to the estimator  α ^ M L E  in the sense of SMSE criterion, i.e.,  S M S E α ^ M N B K L E S M S E α ^ M L E < 0  if  ψ j j < λ j + k 2 4 α j 2 k 2 / λ j k 2 .
Proof. 
The difference between S M S E α ^ M N B K L E   and   S M S E α ^ MLE is as follows:
SMSE α ^ M N B K L E   SMSE α ^ MLE = j = 1 p λ j k 2 λ j + k 2 ψ j j + 4 j = 1 p α j 2 k 2 λ j + k 2 j = 1 p 1 λ j .  
j = 1 p λ j k 2 λ j ψ j j + 4 α j 2 k 2 λ j λ j λ j + k 2 λ j λ j + k 2
The difference between SMSE α ^ M N B K L E   and   SMSE α ^ MLE in Equation (34) will be less than zero if ψ j j < λ j + k 2 4 α j 2 k 2 / λ j k 2 ; thus, it means that α ^ M N B K L E is better than α ^ MLE since it has smaller scalar mean squared errors. □
Theorem 2. 
The estimator  α ^ M N B K L E  is superior to the estimator  α ^ N B R R E  in the sense of SMSE criterion, i.e.,  S M S E α ^ M N B K L E S M S E α ^ N B R R E < 0  if  ψ j j < λ j 3 α j 2 k 2 / λ j k 2 .
Proof. 
The difference between SMSE α ^ M N B K L E   and   SMSE α ^ N B R R E is as follows:
SMSE α ^ M N B K L E   SMSE α ^ N B R R E = j = 1 p λ j k 2 λ j + k 2 ψ j j + 4 j = 1 p α j 2 k 2 λ j + k 2 j = 1 p λ j λ j + k 2 j = 1 p α j 2 k 2 λ j + k 2
j = 1 p λ j k 2 ψ j j + 4 α j 2 k 2 α j 2 k 2 λ j λ j + k 2  
The difference between SMSE α ^ M N B K L E   and   SMSE α ^ N B R R E in Equation (35) will be less than zero if ψ j j < λ j 3 α j 2 k 2 / λ j k 2 ; thus, it means that α ^ M N B K L E is better than α ^ N B R R E since it has smaller scalar mean squared errors. □
Theorem 3. 
The estimator  α ^ M N B K L E  is superior to the estimator  α ^ M N B R R E  in the sense of SMSE criterion, i.e.,  S M S E α ^ M N B K L E S M S E α ^ M N B R R E < 0  if  ψ j j < 3 α j 2 k 2 / 2 λ j k k 2 .
Proof. 
The difference between SMSE α ^ M N B K L E   and   SMSE α ^ M N B R R E is as follows:
SMSE α ^ M N B K L E   SMSE α ^ M N B R R E = j = 1 p λ j k 2 λ j + k 2 ψ j j + 4 j = 1 p α j 2 k 2 λ j + k 2 j = 1 p λ j 2 λ j + k 2 ψ j j j = 1 p α j 2 k 2 λ j + k 2
j = 1 p λ j k 2 ψ j j + 4 α j 2 k 2 λ j 2 ψ j j α j 2 k 2 λ j + k 2  
The difference between SMSE α ^ M N B K L E   and   SMSE α ^ M N B R R E in Equation (36) will be less than zero if ψ j j < 3 α j 2 k 2 / 2 λ j k k 2 ; thus, it means that α ^ M N B K L E is better than α ^ M N B R R E since it has smaller scalar mean squared errors. □
Theorem 4. 
The estimator  α ^ M N B K L E  is superior to the estimator  α ^ N B K L E  in the sense of SMSE criterion, i.e.,  S M S E α ^ M N B K L E S M S E α ^ N B K L E < 0  if  ψ j j < 1 / λ j .
Proof. 
The difference between SMSE α ^ M N B K L E   and   SMSE α ^ N B K L E is as follows:
SMSE α ^ M N B K L E   SMSE α ^ N B K L E = j = 1 p λ j k 2 λ j + k 2 ψ j j + 4 j = 1 p α j 2 k 2 λ j + k 2 j = 1 p λ j k 2 λ j λ j + k 2 4 j = 1 p α j 2 k 2 λ j + k 2
j = 1 p λ j k 2 ψ j j λ j λ j k 2 λ j λ j + k 2
The difference between SMSE α ^ M N B K L E   and   SMSE α ^ N B K L E in Equation (37) will be less than zero if ψ j j < 1 / λ j ; thus, it means that α ^ M N B K L E is better than α ^ N B K L E since it has smaller scalar mean squared errors. □

3. Simulation Study

To ensure a robust evaluation, we meticulously designed the simulation study, carefully specifying the variables to influence the characteristics of the proposed estimator. We selected a suitable metric, drawing inspiration from well-established references to evaluate the outcomes [11,12,41,42,43,44]. To generate the regressors, we followed the equation:
x i j = 1 ρ 2 1 2 m i j + ρ m i , p + 1 ,   i = 1 , 2 , , n ,   j = 1 , 2 , 3 ,   ,   p
Here, m i j represents independent standard normal pseudo-random numbers. We considered various scenarios by varying the number of regressors ( p = 3 and 5) and the level of multicollinearity ( ρ = 0.8 ,   0.9 ,   0.99 ,   0.999 ) . The response variable y was generated using the “rnbinom” function, which simulates NB random variables. The mean of y was computed based on the exponential of the linear combination of predictors ( μ i = exp x i T , β ) where the over-dispersion parameter (σ) is set to 5. The length of y was varied among n = 30, 50, 100, and 200, ensuring that the regression parameters β were chosen such that β β = 1 [42,43,44].
The line y [sample (n, n*0.08)] = 50 was used to introduce outliers in the response variable. The sample function randomly selects a subset of n*b indices, where n*b represents b% of the total number of observations in the response variable. The percentage of outliers (b) was chosen arbitrarily as 1 and 8, respectively. The selected indices corresponded to the positions of the outliers, and their values were then assigned to a predefined outlier value of 50. Consequently, 1% or 8% of the observations in the response variable were replaced with the outlier value of 50.
The model is estimated using the various methods discussed in this study. The “robNB” package is useful when dealing with count data with outliers, where the number of occurrences of an event is non-negative and discrete. It produced robust standard errors and other robust inferential procedures.
To ensure the reliability of our results, we conducted 2000 replications of the experiment. To evaluate the performance of the estimators, we calculated the estimated mean squared error (MSE) using the following formula:
MSE = 1 2000   i = 1 2000 j = 1 p β ^ i j β j 2
Here, β ^ i j represents the estimated value of the jth parameter in the ith replication, and β j denotes the corresponding true parameter value. The estimated MSE values for the proposed estimator, with other estimators, can be found in Table 1 and Table 2.
Table 1 and Table 2 present the estimated mean squared error (MSE) values for the proposed estimator and other estimators considered in this study. Based on the results obtained from the simulation, several important conclusions have been drawn:
i. Impact of Multicollinearity: To investigate the effect of multicollinearity on the estimates of regression parameters, we examined the correlation coefficients ρ = 0.8, 0.9, 0.99, and 0.999. It was observed that as the correlation between explanatory variables increases, the MSE of both classical estimators and robust hybrid estimators also increases (Figure 1). The MSE values for robust hybrid estimators, denoted with α ^ M N B R R E and α ^ M N B K L E , show a reduction in magnitude compared to their non-hybrid counterparts.
ii. Influence of Sample Size: Comparing the performance of estimators for different sample sizes (n = 30, 50, 100, and 200), it becomes evident that the MSE decreases as the sample size increases. Detailed graphical representations can be found in Figure 2. For instance, with ρ = 0.8, the MSE decreases from n = 30 to n = 200 for all estimators. Similarly, for other correlation coefficients, there is a consistent pattern of decreasing MSE values as the sample size grows.
iii. Impact of Explanatory Variables: The total number of explanatory variables has a considerable influence on the MSE values of estimators. The MSE values tend to be higher for all estimators when there are more explanatory variables.
iv. Impact of Outliers: The percentage of outliers in the model increased the estimated MSE values. In Table 1 and Table 2 ( p = 3 and p = 5), we observe that as the percentage of outliers increases from 1% to 8%, the MSE values for all estimators also increase significantly. The presence of outliers introduces more variability and influences the accuracy of the estimators, leading to higher MSE values.
Overall, the MLE and NBRRE estimators typically exhibit higher MSE values compared to the robust hybrid estimators (M-NBRRE and M-NBKLE) across different sample sizes and correlation coefficients. Among the robust hybrid estimators, M-NBKLE generally outperforms M-NBRRE in terms of lower MSE values.

4. Application

In this application, we conducted a comprehensive analysis of the nuts dataset, as documented in Hilbe [45] and recently adopted by Algamal [46]. The dataset is available in the library COUNT. The dataset comprises fifty-two (52) observations and seven (7) explanatory variables, focusing on the behavior of squirrels and various forest characteristics across different plots within Scotland’s Abernathy Forest. Specifically, the response variable represents the number of cones stripped by squirrels. The explanatory variables include the number of trees per plot (x1), number of DBH (diameter at breast height) per plot (x2), mean tree height per plot (x3), canopy closure (as a percentage) (x4), standardized number of trees per plot (x5), standardized mean tree height per plot (x6), and standardized canopy closure (as a percentage) (x7).
Due to challenges encountered during the modeling process, we excluded variable x7 from the analysis. We proceeded to fit Poisson and NBR models to identify the most appropriate model for the dataset. See Table 3 for the outcome. Notably, the Akaike Information Criterion (AIC) and residual variance from Table 3 unequivocally indicate that the NBRM provides an excellent fit to the data. Moreover, the dispersion test revealed the presence of overdispersion in the model, rendering the Poisson regression model unsuitable for this dataset.
Additionally, the estimated variance inflation factors are 4.474, 16.332, 39.61, and 40.89, indicating a clear presence of correlated regressors in the model. Consequently, the model suffers from both multicollinearity and overdispersion. In our further examination, we conducted outlier diagnostics. The residual plots revealed the existence of outliers in the model. Specifically, the standardized Pearson residual plot against the leverage identified cases 2, 11, and 25 as outliers (See Figure 3). To tackle these challenges, we proceeded with model fitting using both non-robust estimators (MLE, NBRRE, and NBKLE) and robust estimators (M-NBRRE and M-NBKLE). See Table 4 for the results, which will aid in understanding the impact of these estimators on our model and its interpretations.
Table 4 presents the estimated regression coefficients obtained from different estimators considered in this study. The estimators used comprise non-robust estimators (MLE, NBRRE, and NBKLE) and robust estimators (M-NBRRE and M-NBKLE). The values in parentheses next to each coefficient represent the standard errors of the estimates, providing insight into the precision of the coefficient estimates. Smaller standard errors indicate greater confidence in the estimate’s accuracy, while larger standard errors suggest more variability in the regression estimate.
Upon examination, we observe intriguing differences in the coefficient signs for x1 and x4 between the non-robust and robust estimators. This contrast highlights the sensitivity of non-robust estimators to outliers, potentially leading to biased coefficient estimates. In contrast, robust estimators, designed to withstand the impact of influential observations, demonstrate more consistent and reliable coefficient estimates.
To gauge the performance of these estimators, we rely on the SMSE. The results in Table 4 revealed that the SMSE values for the robust estimators, M-NBRRE and M-NBKLE, are substantially lower than those for the non-robust estimators (MLE, NBRRE, and NBKLE). This substantial difference suggests that the robust estimators exhibit superior predictive accuracy and resilience to the influence of outliers, making them more suitable for this dataset.
Particularly noteworthy is the robust hybrid KL estimator (M-NBKLE), which emerges as the most promising method among all the estimators. Not only does it provide accurate coefficient estimates, but it also achieves the lowest SMSE and has relatively small standard errors, indicating that its estimates are both reliable and precise. With its integration of robustness and superior predictive power, the M-NBKLE estimator dominates other approaches, providing researchers with a reliable tool to obtain accurate and stable coefficient estimates even in the presence of extreme observations.

5. Conclusions

NBR is widely used for modeling non-negative integer data, particularly when overdispersion is present. However, interpreting the model becomes more challenging when assumptions, such as the absence of multicollinearity among predictors, are violated. In practice, meeting these assumptions can be difficult, and multicollinearity can significantly reduce the efficiency of maximum likelihood estimators. Ridge and Kibria–Lukman estimators have been applied to address multicollinearity, but their performance often deteriorates in the presence of outliers. Robust estimation techniques have been proposed to handle the impact of outliers in NBR. Recent studies have shown that the combined effects of outliers and multicollinearity can significantly degrade model performance, a problem well-documented in linear and Poisson regression models. This insight has motivated us to propose a new method to address these issues, specifically, in NBR.
This study addresses the challenges of multicollinearity and outliers in NBR by integrating Ridge and Kibria–Lukman estimators with robust estimators, resulting in new hybrid estimators termed M-NBRRE and M-NBKLE. We evaluated their performance theoretically and through simulations, with the results confirming the robustness and effectiveness of the hybrid estimators. Notably, M-NBKLE demonstrated superior performance, particularly by achieving a lower estimated mean squared error. A real-world application was also conducted, focusing on squirrel behavior and forest characteristics across plots in Scotland’s Abernathy Forest, further supporting the dominance of M-NBKLE.
Our findings contribute to the growing body of research on count regression models and offer valuable tools for researchers needing robust data analysis techniques across various fields. This work represents a significant step toward enhancing the accuracy and reliability of NBR, ensuring robust inferences, and enabling sound decision-making in practical data analysis. In future research, we plan to incorporate additional evaluation metrics, such as the Pearson chi-square statistic, which could provide further insights into model fit and performance, especially for count data. We also aim to refine these estimators by incorporating techniques like principal component regression or partial least squares to enhance estimator performance.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/math12182929/s1. We have included the code used for the real-life application to facilitate replication of our results.

Author Contributions

Conceptualization, A.F.L., M.A. and R.A.F.; methodology, A.F.L. and R.A.F.; software, A.F.L.; formal analysis, A.F.L.; resources, All authors; writing—original draft preparation, All authors; writing—review and editing, All authors; visualization, R.A.F.; supervision, O.A.; funding acquisition, O.A., J.A. and A.A.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data will be made available upon request from the corresponding author.

Acknowledgments

We want to express our sincere gratitude to the Editorial team and reviewers for their meticulous review and invaluable feedback, which have greatly enhanced the quality of our manuscript. Mohammad Arashi’s work was supported in part by the Iran National Science Foundation (INSF) under grant No. 4015320.

Conflicts of Interest

There are no conflicts of interest to declare in this study.

References

  1. Aeberhard, W.H.; Cantoni, E.; Heritier, S. Robust inference in the negative binomial regression model with an application to falls data. Biometrics 2014, 70, 920–931. [Google Scholar] [CrossRef]
  2. McCullagh, P.; Nelder, J.A. Generalized Linear Models, 2nd ed.; Chapman & Hall: London, UK, 1989. [Google Scholar]
  3. Algamal, Z.Y.; Abonazel, M.R.; Awwad, F.A.; Eldin, E.T. Modified jackknife ridge estimator for the Conway-Maxwell-Poisson model. Sci. Afr. 2023, 19, e01543. [Google Scholar] [CrossRef]
  4. Månsson, K. On ridge estimators for the negative binomial regression model. Econ. Model. 2012, 29, 178–184. [Google Scholar] [CrossRef]
  5. Mansson, K. Developing a Liu estimator for the negative binomial regression model: Method and application. J. Stat. Comput. Simul. 2013, 83, 1773–1780. [Google Scholar] [CrossRef]
  6. Alobaidi, N.N.; Shamany, R.E.; Algamal, Z.Y. A new ridge estimator for the negative binomial regression model. Thail. Stat. 2021, 19, 116–125. [Google Scholar]
  7. Jabur, D.; Rashad, N.; Algamal, Z. Jackknifed Liu-type estimator in the negative binomial regression model. Int. J. Nonlinear Anal. Appl. 2022, 13, 2675–2684. [Google Scholar]
  8. Abonazel, M.R.; El-sayed, S.M.; Saber, O.M. Performance of robust count regression estimators in the case of overdispersion, zero inflated, and outliers: Simulation study and application to German health data. Commun. Math. Biol. Neurosci. 2021, 2021, 55. [Google Scholar]
  9. Tüzen, F.; Erbaş, S.; Olmuş, H. A simulation study for count data models under varying degrees of outliers and zeros. Commun. Stat.-Simul. Comput. 2020, 49, 1078–1088. [Google Scholar] [CrossRef]
  10. Medina, M.A.; Ronchetti, E. Robust statistics: A selective overview and new directions. WIREs Comput. Stat. 2015, 7, 372–393. [Google Scholar] [CrossRef]
  11. Lukman, A.F.; Arashi, M.; Prokaj, V. Robust biased estimators for Poisson regression model: Simulation and applications. Concurr. Comput. Pract. Exp. 2023, 35, e7594. [Google Scholar] [CrossRef]
  12. Lukman, A.F.; Farghali, R.A.; Kibria, B.M.G.; Oluyemi, O.A. Robust-stein estimator for overcoming outliers and multicollinearity. Sci. Rep. 2023, 13, 9066. [Google Scholar] [CrossRef] [PubMed]
  13. Roozbeh, M.; Babaie-Kafaki, S.; Aminifard, Z. Two penalized mixed–integer nonlinear programming approaches to tackle multicollinearity and outliers effects in linear regression models. J. Ind. Manag. Optim. 2021, 17, 3475–3491. [Google Scholar] [CrossRef]
  14. Roozbeh, M.; Babaie-Kafaki, S.; Maanavi, M. A heuristic algorithm to combat outliers and multicollinearity in regression model analysis. Iran. J. Numer. Anal. Optim. 2022, 12, 173–186. [Google Scholar]
  15. Hoerl, A.E.; Kennard, R.W. Ridge regression: Biased estimation for nonorthogonal problems. Technometrics 1970, 12, 55–67. [Google Scholar] [CrossRef]
  16. Liu, K. A new class of biased estimate in linear regression. Commun. Stat. 1993, 22, 393–402. [Google Scholar]
  17. Liu, K. Using Liu-type estimator to combat collinearity. Commun. Stat.-Theory Methods 2003, 32, 1009–2003. [Google Scholar] [CrossRef]
  18. Kibria, B.M.G.; Lukman, A.F. A New Ridge-Type Estimator for the Linear Regression Model: Simulations and Applications. Scientifica 2020, 2020, 9758378. [Google Scholar] [CrossRef]
  19. Asar, Y.; Genc, A. New shrinkage parameters for the liu-type logistic estimators. Commun. Stat.-Simul. Comput. 2016, 45, 1094–1103. [Google Scholar] [CrossRef]
  20. Asar, Y.; Genç, A. Two-parameter ridge estimator in the binary logistic regression. Commun. Stat.-Simul. Comput. 2017, 46, 7088–7099. [Google Scholar] [CrossRef]
  21. Kibria, B.G.; Månsson, K.; Shukur, G. Some ridge regression estimators for the zero-inflated Poisson model. J. Appl. Stat. 2013, 40, 721–735. [Google Scholar] [CrossRef]
  22. Kibria, B.G.; Månsson, K.; Shukur, G. A Ridge Regression Estimator for the Zero-Inflated Poisson Model; Royal Institute of Technology, CESIS-Centre of Excellence for Science and Innovation Studies: Stockholm, Sweden, 2011. [Google Scholar]
  23. Al-Taweel, Y.; Algamal, Z. Almost unbiased ridge estimator in the zero-inflated Poisson regression model. TWMS J. Appl. Eng. Math. 2022, 12, 235–246. [Google Scholar]
  24. Månsson, K.; Shukur, G. A Poisson ridge regression estimator. Econ. Model. 2011, 28, 1475–1481. [Google Scholar] [CrossRef]
  25. Asar, Y.; Genç, A. A new two-parameter estimator for the Poisson regression model. Iran. J. Sci. Technol. Trans. Sci. 2018, 42, 793–803. [Google Scholar] [CrossRef]
  26. Lukman, A.F.; Aladeitan, B.; Ayinde, K.; Abonazel, M.R. Modified ridge-type for the Poisson Regression Model: Simulation and Application. J. Appl. Stat. 2021, 49, 2124–2136. [Google Scholar] [CrossRef]
  27. Lukman, A.F.; Adewuyi, E.; Månsson, K.; Kibria, B.M.G. A new estimator for the multicollinear poisson regression model: Simulation and application. Sci. Rep. 2021, 11, 3732. [Google Scholar] [CrossRef]
  28. Huang, J.; Yang, H. A two-parameter estimator in the negative binomial regression model. J. Stat. Comput. Simul. 2014, 84, 124–134. [Google Scholar] [CrossRef]
  29. Çetinkaya, M.K.; Kaçıranlar, S. Improved two-parameter estimators for the negative binomial and Poisson regression models. J. Stat. Comput. Simul. 2019, 89, 2645–2660. [Google Scholar] [CrossRef]
  30. Abonazel, M.R.; Saber, A.A.; Awwad, F.A. Kibria–Lukman estimator for the Conway–Maxwell Poisson regression model: Simulation and applications. Sci. Afr. 2023, 19, e01553. [Google Scholar] [CrossRef]
  31. Barnett, V.; Lewis, T. Outliers in Statistical Data; Wiley: New York, NY, USA, 1994. [Google Scholar]
  32. Chatterjee, S.; Hadi, A.S. Influential observations, high leverage points, and outliers in linear regression. Stat. Sci. 1986, 1, 379–416. [Google Scholar]
  33. Rousseeuw, P.J.; Leroy, A.M. Robust Regression and Outlier Detection; Series in Applied Probability and Statistics; Wiley Interscience: New York, NY, USA, 1987; 329p. [Google Scholar] [CrossRef]
  34. Wasim, D.; Suhail, M.; Albalawi, O.; Shabbir, M. Weighted penalized m-estimators in robust ridge regression: An application to gasoline consumption data. J. Stat. Comput. Simul. 2024, 1–30. [Google Scholar] [CrossRef]
  35. Huber, P.J. Robust Regression: Asymptotics, Conjectures and Monte Carlo. Ann. Stat. 1973, 1, 799–821. [Google Scholar] [CrossRef]
  36. Silvapulle, M. Robust ridge regression based on an M-estimator. Aust. J. Stat. 1991, 33, 319–333. [Google Scholar] [CrossRef]
  37. Ertas, H. A modified ridge M-estimator for linear regression model with multicollinearity and outliers. Commun. Stat.-Simul. Comput. 2018, 47, 1240–1250. [Google Scholar] [CrossRef]
  38. Dawoud, I.; Abonazel, M. Robust Dawoud–Kibria estimator for handling multicollinearity and outliers in the linear regression model. J. Stat. Comput. Simul. 2021, 91, 3678–3692. [Google Scholar] [CrossRef]
  39. Abonazel, M.; Dawoud, I. Developing robust ridge estimators for Poisson regression model. Concurr. Comput. Pract. Exp. 2022, 34, e6979. [Google Scholar] [CrossRef]
  40. Arum, K.; Ugwuowo, F.; Oranye, H.; Alakija, T.; Ugah, T.; Asogwa, O. Combating outliers and multicollinearity in linear regression model using robust Kibria-Lukman mixed with principal component estimator, simulation and computation. Sci. Afr. 2023, 19, e01566. [Google Scholar] [CrossRef]
  41. Majid, A.; Ahmed, S.; Aslam, M.; Kashif, M. A robust Kibria–Lukman estimator for linear regression model to combat multicollinearity and outliers. Concurr. Comput. Pract. Exp. 2023, 35, e7533. [Google Scholar] [CrossRef]
  42. Kibria, B.M.G. Performance of some new ridge regression estimators. Commun. Stat.-Simul. Comput. 2003, 32, 419–435. [Google Scholar] [CrossRef]
  43. Dawoud, I.; Lukman, A.F.; Haadi, A. A new biased regression estimator: Theory, simulation and application. Sci. Afr. 2022, 15, e01100. [Google Scholar] [CrossRef]
  44. Kibria, B.M.G. More than hundred (100) estimators for estimating the shrinkage parameter in a linear and generalized linear ridge regression models. J. Econom. Stat. 2022, 2, 233–252. [Google Scholar]
  45. Hilbe, J.M. Negative Binomial Regression, 2nd ed.; Cambridge University Press: Cambridge, MA, USA, 2011. [Google Scholar]
  46. Algamal, Z. Variable Selection in Count Data Regression Model based on Firefly Algorithm. Stat. Optim. Inf. Comput. 2019, 7, 520–529. [Google Scholar] [CrossRef]
Figure 1. Graph of MSE against the level of correlation for each estimator when the outlier is 1%.
Figure 1. Graph of MSE against the level of correlation for each estimator when the outlier is 1%.
Mathematics 12 02929 g001
Figure 2. Graph of MSE against the sample size for the estimators corresponding to ρ = 0.8 when the outlier is 1%.
Figure 2. Graph of MSE against the sample size for the estimators corresponding to ρ = 0.8 when the outlier is 1%.
Mathematics 12 02929 g002
Figure 3. Diagnostic plot of squirrel dataset NB modelling.
Figure 3. Diagnostic plot of squirrel dataset NB modelling.
Mathematics 12 02929 g003
Table 1. Estimated MSE values for p* = 3 with 1% and 8% outliers.
Table 1. Estimated MSE values for p* = 3 with 1% and 8% outliers.
n1% Outliers8%
30501002003050100200
ρ = 0.8 ρ = 0.8
α ^ M L E 0.24330.17590.14070.13711.73021.71171.32971.2726
α ^ N B R R E 0.22260.14890.13220.13121.52841.51061.28431.2561
α ^ M N B R R E 0.15430.14800.13200.02740.40170.32180.18680.0971
α ^ N B K L E 0.20540.13380.13010.12991.34381.28561.24011.2397
α ^ M N B K L E 0.13230.12580.12270.02550.33640.31780.16230.0933
ρ = 0.9 ρ = 0.9
α ^ M L E 2.29990.51320.36060.16714.34352.46491.42631.3409
α ^ N B R R E 2.19460.41540.28720.14943.64602.26171.33721.3218
α ^ M N B R R E 1.32130.33460.21280.04392.08741.45980.42950.2087
α ^ N B K L E 2.14080.34810.22850.13413.02002.09201.25391.2630
α ^ M N B K L E 1.25240.25080.18950.04041.50431.36880.35160.1169
ρ = 0.99 ρ = 0.99
α ^ M L E 4.28882.09522.03251.262213.62509.16682.85362.0822
α ^ N B R R E 3.47811.25611.30100.82759.66635.44771.91291.7719
α ^ M N B R R E 2.71751.12270.56590.30055.99705.01011.40410.3160
α ^ N B K L E 3.44300.63160.58570.53426.99713.12661.60771.5263
α ^ M N B K L E 2.37950.50890.43020.11493.71443.07110.81170.1274
ρ = 0.999 ρ = 0.999
α ^ M L E 13.654414.36007.74666.7521149.293770.448733.474211.6880
α ^ N B R R E 5.65217.48972.45141.5667104.428646.418823.58427.8009
α ^ M N B R R E 3.80325.16301.74711.312667.550824.14802.13841.6055
α ^ N B K L E 5.01853.96741.20861.158071.748432.059617.34645.9552
α ^ M N B K L E 3.13952.61551.13610.498532.85147.92321.47560.5311
Table 2. Estimated MSE values for p* = 5 with 1% and 8% outliers.
Table 2. Estimated MSE values for p* = 5 with 1% and 8% outliers.
n1% Outliers8%
30501002003050100200
ρ = 0.8 ρ = 0.8
α ^ M L E 0.37450.34130.33360.32592.31252.01201.61641.5833
α ^ N B R R E 0.28780.27940.24850.18492.26101.73291.35881.2773
α ^ M N B R R E 0.27910.26330.20460.04581.00640.82540.39550.1100
α ^ N B K L E 0.25060.24600.23870.14682.21641.63701.29321.2714
α ^ M N B K L E 0.20880.20010.19580.04001.00280.70870.33950.1082
ρ = 0.9 ρ = 0.9
α ^ M L E 3.75420.67880.63530.59625.51002.81251.65463.3937
α ^ N B R R E 3.51190.48560.46470.42544.36962.45321.52533.3180
α ^ M N B R R E 2.54760.43230.42170.05653.01102.06340.54180.2955
α ^ N B K L E 3.34060.53660.52460.46184.26232.38501.40563.2442
α ^ M N B K L E 2.37750.50320.24810.04623.01062.00420.44040.2774
ρ = 0.99 ρ = 0.99
α ^ M L E 6.88415.19154.67333.855014.388511.55787.97389.2243
α ^ N B R R E 4.18803.23443.13332.737412.84317.70366.25747.5123
α ^ M N B R R E 3.89323.43181.61211.43136.02145.55072.14960.3618
α ^ N B K L E 3.54522.00111.85651.807010.66537.06994.85874.1347
α ^ M N B K L E 3.11882.13641.46390.23915.04504.44542.04900.3477
ρ = 0.999 ρ = 0.999
α ^ M L E 71.140445.717938.593732.2868245.5499100.221387.348126.0989
α ^ N B R R E 43.611529.316212.761810.2636128.813162.344550.537515.0965
α ^ M N B R R E 39.337028.90902.81732.683171.053145.04814.42003.1227
α ^ N B K L E 26.543719.322610.42177.788786.093659.112947.782412.0998
α ^ M N B K L E 23.487518.52392.31131.085870.745943.23663.65192.7898
Table 3. Model adequacy check.
Table 3. Model adequacy check.
Metrics P o i s s o n N e g a t i v e   B i n o m i a l
Residual Variance661.2661.2
AIC873.1397.9
Table 4. Estimated regression coefficients for the squirrel dataset.
Table 4. Estimated regression coefficients for the squirrel dataset.
Coef.Estimators
α ^ M L E α ^ N B R R E α ^ M N B R R E α ^ N B K L E α ^ M N B K L E
x10.0267
(0.0128)
0.0235
(0.0120)
−2.7247
(0.0122)
0.0203
(0.0115)
−2.7247
(0.0119)
x21.9640
(1.8877)
1.0210
(1.3625)
0.0043
(1.5516)
0.0780
(0.9834)
0.0065
(1.2754)
x30.0228
(0.0456)
0.0339
(0.0429)
0.0267
(0.0438)
0.0451
(0.0413)
0.0267
(0.0425)
x40.0139
(0.0101)
0.0152
(0.0100)
−0.0350
(0.0100)
0.0165
(0.0100)
−0.0350
(0.0100)
x50.4493
(0.1659)
0.4350
(0.1646)
0.0490
(0.1651)
0.4207
(0.1637)
0.0488
(0.1645)
SMSE3.59341.88551.65603.59270.4692
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lukman, A.F.; Albalawi, O.; Arashi, M.; Allohibi, J.; Alharbi, A.A.; Farghali, R.A. Robust Negative Binomial Regression via the Kibria–Lukman Strategy: Methodology and Application. Mathematics 2024, 12, 2929. https://doi.org/10.3390/math12182929

AMA Style

Lukman AF, Albalawi O, Arashi M, Allohibi J, Alharbi AA, Farghali RA. Robust Negative Binomial Regression via the Kibria–Lukman Strategy: Methodology and Application. Mathematics. 2024; 12(18):2929. https://doi.org/10.3390/math12182929

Chicago/Turabian Style

Lukman, Adewale F., Olayan Albalawi, Mohammad Arashi, Jeza Allohibi, Abdulmajeed Atiah Alharbi, and Rasha A. Farghali. 2024. "Robust Negative Binomial Regression via the Kibria–Lukman Strategy: Methodology and Application" Mathematics 12, no. 18: 2929. https://doi.org/10.3390/math12182929

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop