Next Article in Journal
A Convolutional Sequence-to-Sequence Attention Fusion Framework for Commonsense Causal Reasoning
Previous Article in Journal
Modelling Infectious Disease Dynamics: A Robust Computational Approach for Stochastic SIRS with Partial Immunity and an Incidence Rate
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Kibria–Lukman-Type Estimator for Regularization and Variable Selection with Application to Cancer Data

by
Adewale Folaranmi Lukman
1,
Jeza Allohibi
2,*,
Segun Light Jegede
3,
Emmanuel Taiwo Adewuyi
4,
Segun Oke
5 and
Abdulmajeed Atiah Alharbi
2
1
Department of Mathematics, University of North Dakota, Grand Forks, ND 58202, USA
2
Department of Mathematics, Faculty of Science, Taibah University, Al-Madinah Al-Munawara 42353, Saudi Arabia
3
Department of Mathematical Sciences, Kent State University, Kent, OH 44242, USA
4
Department of Statistics, Ladoke Akintola University of Technology, Ogbomoso 212102, Nigeria
5
Department of Physics, Chemistry and Mathematics, Alabama A&M University, Huntsville, AL 35762, USA
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(23), 4795; https://doi.org/10.3390/math11234795
Submission received: 16 October 2023 / Revised: 10 November 2023 / Accepted: 21 November 2023 / Published: 28 November 2023

Abstract

:
Following the idea presented with regard to the elastic-net and Liu-LASSO estimators, we proposed a new penalized estimator based on the Kibria–Lukman estimator with L1-norms to perform both regularization and variable selection. We defined the coordinate descent algorithm for the new estimator and compared its performance with those of some existing machine learning techniques, such as the least absolute shrinkage and selection operator (LASSO), the elastic-net, Liu-LASSO, the GO estimator and the ridge estimator, through simulation studies and real-life applications in terms of test mean squared error (TMSE), coefficient mean squared error (βMSE), false-positive (FP) coefficients and false-negative (FN) coefficients. Our results revealed that the new penalized estimator performs well for both the simulated low- and high-dimensional data in simulations. Also, the two real-life results show that the new method predicts the target variable better than the existing ones using the test RMSE metric.

1. Introduction

The linear regression model describes the response variable as a linear combination of one or more predictors, and it is widely applicable in fields such as medicine, biological science, economics, etc. The model is defined as follows:
y i = x i T β + ε i
where y = y 1 , , y n T is a n × 1 vector of the dependent (response) variable; X = x 1 , , x n T , x i R n × p and i = 1 , , n are the known n × p matrix of predictors; β = β 1 , , β p T is a p × 1 vector of regression coefficients; and ε = ε 1 , , ε n T is the random error vector such that E ε i = 0 ,. V a r ε i = σ 2 R + , i = 1 , , n .
The model parameters are usually estimated using the least squares (LS) estimator. The LS estimator minimized the residual sum of squares (RSS) of model (1) and produced the following expression:
β ^ n L S = arg min β R p 1 2 n y X β 2 2
where 2 denotes the Euclidean norm. The high-dimensional quantitative structure–activity relationship (QSAR) modeling method is a popular example of the linear regression models. A high-dimensional problem arises when there are more features (variables) than the sample size (n) in a dataset. The popular OLSE does not give a unique solution via high-dimensional QSAR modeling since the number of molecular descriptors exceeds the compound number [1,2]. Likewise, the performance of the OLSE suffers a setback when there is linear dependency among the features (predictors), a condition called multicollinearity [3,4,5,6]. The setback includes a high standard error, insignificant t-tests, wrong conclusions, etc. [6,7,8].
The shrinkage estimators were mainly adopted to account for the linear dependency among the predictors. The shrinkage estimators include the ridge estimator [9], the Liu estimator [10], the Liu-type estimator [11], the two-parameter estimator [12], the modified ridge-type estimator [6], the modified Liu estimator [13], the Kibria–Lukman (KL) estimator [14], the modified KL estimator [15], and the Dawoud–Kibria (DK) estimator [16]. These shrinkage estimators account for more multicollinearity than the OLSE. The ridge estimator minimizes the residual sum of the squares (RSS) subject to an L2-norm of the coefficients, and it is defined as follows:
β ^ n λ = arg min β R p 1 2 n y X β 2 2 + λ 2 β 2 2 ,
where λ > 0 represents the penalized parameter. The estimator produces a more stable estimate and a better prediction than the LS estimator. The ridge estimator will only perform regularization in high-dimensional settings and will not perform variable selection. It fails to reduce any of its coefficients to zero. Hence, it becomes difficult to interpret the regression models in high-dimensional settings.
Variable selection forms a crucial aspect of modeling high-dimensional problems. It is worthwhile to state that removing the important features (predictors) produces more biased regression estimates and predictions. Likewise, including irrelevant predictors degrades the estimation efficiency and leads to imprecise predictions [17]. Thus, variable selection seeks to select the most relevant variables from a large dataset and remove irrelevant and redundant variables. Moreover, it improves the performance of the model prediction and reduces the computational cost. The problem of variable selection has been studied extensively in the literature [17,18,19,20,21,22].
The least absolute penalty and selection operator (LASSO) [18] is a well-known variable selection technique obtained by subjecting the RSS to an L1-norm penalty, and it is defined as follows:
β ^ n L = arg min β R p 1 2 n y X β 2 2 + λ β 1 ,
where λ > 0 is the penalized parameter. LASSO retains the regularization property of the ridge and performs variable selection due to the introduction of the L1-norm penalty. It also improves the model prediction by eliminating irrelevant variables. However, the estimator often chooses only one variable from the group of pair-wise correlated predictors, and it is dominated by the ridge estimator when the multicollinearity level is high [1,8,18,22].
The elastic-net (E-net) approach [1] is another popular variable selection technique obtained by subjecting the RSS to the L1 and L2 norms penalty simultaneously. This technique retains the features of the ridge and the LASSO estimators. It is defined as follows:
β ^ n E = arg min β R p 1 2 n y X β 2 2 + λ 2 2 β 2 2 + λ 1 β 1 ,
where λ 1 , λ 2 > 0 , 2 is the L2-norm, and 1 is L1-norm. The drawback of this method is that the L2-penalty term introduced extra bias that increases its variance [23].
Recently, the GO estimator [23] was developed as an improvement on the elastic net model. The estimator combines the features of the two-parameter estimator [24] and the LASSO. It is defined as follows:
β ^ n G O = arg min β R p 1 2 n y X β 2 2 + λ 2 2 β d β ^ n L S 2 2 + λ 1 β 1 ,
where λ 1 , λ 2 , d > 0 .
The Liu-LASSO [21] is another type of E-net that is defined as follows:
β ^ n L L = arg min β R p 1 2 n y X β 2 2 + 1 2 β d β ^ n L S 2 2 + λ 1 β 1 ,
where λ 1 , d > 0 . Both the GO estimator and the Liu-LASSO adopted two other shrinkage estimators (the two-parameter and the Liu estimators) other than the ridge estimator. Research shown that these shrinkage estimators compete well with the ridge estimator. Recently, the Kibria–Lukman (KL) estimator [14] shows competitive advantage over the ridge and the Liu estimator. Thus, the KL estimator will be adopted in a manner similar to the Liu-LASSO to form a new variable selection technique. The goal is to produce a new estimator in the class of the previously mentioned ones and assess its ability to deal with multicollinearity and high-dimensional problems.
The outline of the article is as follows: We define the new estimator in Section 2 and derive its properties. We analyze low-dimensional data and high-dimensional QSAR data to compare the proposed estimator to the existing ones, as shown in Section 4. The simulation study and conclusion are given in Section 3 and Section 5, respectively.

2. Proposed Estimator (KL1)

The Kibria–Lukman estimator is an optimization problem given as follows:
β ^ n K L = arg min β R p 1 2 n y X β 2 2 + λ 2 2 β + β ^ n L S 2 2 ,
where λ 2 > 0 . The KL estimator possesses smaller bias compared to the ridge and the Liu estimator. Therefore, in accordance with the Liu-LASSO estimator, we proposed the KL1 estimator, and it is defined as follows:
β ^ n K L 1 = arg min β R p 1 2 n y X β 2 2 + 1 2 β + β ^ n L S 2 2 + λ 1 β 1 ,
where λ 1 , λ 2 > 0 . The estimator automatically retains the regularization and the variable selection property due to the L1-norm penalty.

Coordinate Descent for KL1

We assume that the predictor variables are standardized and the response variable is centered. We attempt to find β ^ j , which minimizes the following objective function:
β ^ j = 1 2 n i = 1 n y i x i β 2 + 1 2 j = 1 p β j + β ^ j L S 2 + λ 1 j p β j = 1 2 n i = 1 n [ y i x i j β j k j x i k β k ] 2 + 1 2 j = 1 p β j + β ^ j L S 2 + λ 1 β j + λ 1 k j β k = 1 2 n i = 1 n [ y i y ¯ i ( j ) x i j β j ] 2 + 1 2 j = 1 p β j + β ^ j L S 2 + λ 1 β j + λ 1 k j β k
where y ¯ i ( j ) = k j x i k β k . The objective function (10) is differentiable everywhere except for | β j | . The first derivative for f β ^ j when β j 0 is given as follows:
f β ^ j β j = 1 n i = 1 n x i j y i y ¯ i ( j ) + β ^ j L S + 1 n i = 1 n x i j 2 β j + β j + λ 1 s g n ( β j )
where s g n β j = 1   β j > 0 0   β j = 0 1   β j < 0 . The soft threshold estimate of β j was obtained by setting Equation (11) to zero. Hence,
β j S 1 n i = 1 n x i j y i y ¯ i j β ^ j L S , λ 1 1 n i = 1 n x i j 2 + 1
where y i y ¯ i j is the partial residual for fitting β j , and S(z, λ 1 ) is the soft-thresholding operator defined by [22] as follows:
S ( z ,   λ 1 ) = z + λ 1 ,       z < λ 1 0 ,                       | z | λ 1 z λ 1 ,     z > λ 1

3. Simulation Studies

In this section, we designed three different cases to explore the performances of the proposed method (KL1) and other estimators. The simulation model is defined as follows:
y = X β + σ ϵ
where the random error is ϵ ~ N 0 , I . The predictor X is generated such that X ~ N p ( 0 ,   ) where k j = ρ | k j | . ρ is the correlation between the predictors. Each independent simulated dataset is divided into three parts: a training set, a validation set and a test set. The division is performed at a ratio of 60%:20%:20%. The training data were used to train the model, while the validation data were used to predict the response while tuning the model’s hyperparameters. The tuning parameters used in the coordinate descent algorithm (Algorithm 1) specify a grid in the interval [0, 1]. The test data are used to provide an unbiased evaluation of a final model fit on the training dataset. The simulation is conducted in two senses of dimensionality—high-dimensional data and low-dimensional data.
Algorithm 1 Pseudo-code for the coordinate descent algorithm
Input: maximum iteration (max_iter),
   tolerance (tol),hyperparameters ( λ 1 ),
   predictors (matrix X), response (vector Y)
Initialize:   β ~ j ,   j = 1 , , p
While iter < max_iter do:
    β j   β ~ j
   While  j < p do:
      u j 1 n i = 1 n x i j y i y ¯ i j β ^ j L S
      v j 1 n i = 1 n x i j 2 + 1
      β j S u j v j , λ 1 v j
   End While
   If β ~ β β ~ β < t o l :
    Break
   Else:
      β ~ j β j
End While
Output: β j ,   j = 1 , , p
The low-dimensional data simulation was carried out for sample sizes of n = 100 ,   200 ,   and   400 . The simulation was also computed to determine the various correlations between the predictors and variances such that ρ = 0.5 ,   0.8 ,   and   0.99 and σ = 5 ,   10 . We simulated 8 true β s with 50 datasets for different cases as follows:
Case 1: β = 3 , 1.5 , 0 , 0 , 2 , 0 , 0 , 0
Case 2: β = 3 , 0 , 0 , 0 , 0 , 0 , 0 , 0
Case 3: β = 3 , 1.5 , 2 , 1.6 , 2 , 1.5 , 1.5 , 1.5
The high-dimensional data simulation was carried out for sample sizes of n = 150   a n d   200 . The simulation was also computed to determine the various correlations between the predictors and variances such that ρ = 0.5 ,   0.8 ,   a n d   0.99 and σ = 5 ,   10 . We simulated 100 true β s with 50 datasets for different cases as follows:
Case 1: β = 0.5 , , 0.5 15 , 0 , , 0 25 , 0.7 , , 0.7 15 , 0 , , 0 30 , 0.7 , , 0.7 15
Case 2: β = 0.5 , , 0.5 15 , 0 , , 0 25 , 0 , , 0 15 , 0 , , 0 30 , 0 , , 0 15
Case 3: β = 0.5 , , 0.5 15 , 0.2 , , 0.2 25 , 0.7 , , 0.7 15 , 0.4 , , 0.4 30 , 0.9 , , 0.9 15
The model performance was measured based on the mean square error. The mean square errors of the predictions from the test data were estimated as given in (15). The mean square error for estimating true coefficients β-MSE through the training model was also observed in (16). The mean and standard errors of these MSEs were recorded and reported. We also reported the median of the false positives (FP) and false negatives (FN) involved in estimating the true coefficients of the model. By false positive, we mean the coefficients that were initially zero from the simulation model specification but returned a positive real number as the estimated coefficient. False-negative coefficients return negative estimated coefficients, though the true coefficients are positive. This allows us to see how well the estimator reports the significant coefficients when it performs dimension reduction.
T M S E = 1 n t e s t y ^ t e s t y t e s t y ^ t e s t y t e s t
β M S E = 1 R β β ^ β β ^
The simulation results of the low-dimensional and high-dimensional simulations are reported in Supplementary Materials (Sections S1 and S2). We also report the spread of the TMSEs for the test prediction and βMSE for true coefficients for the high-dimensional cases in Supplementary Materials (Section S3).
A general observation of the results of the simulated data (Supplementary Materials, Sections S1 and S2) indicates that the values of the MSEs of the estimators increase as the standard deviation ( σ ) and the degree of multicollinearity ( ρ ) increase. There are also decreases in the values of the MSEs of the estimators as the sample size increases. Although our main focus is feature reduction in high-dimensional data, we observe a prediction for our proposed estimator better than those of other estimators.
For the low-dimensional data result discussed in case 1 (Supplementary Materials, Section S1, Tables S1–S3), we observed that the KL1 and GO estimators performed better than the other estimators. The KL1 estimator performed better at predicting the test sample based on the MSE criterion, while the GO estimator had the minimum MSE for estimating the beta coefficients. In case 2 (Supplementary Materials, Section S1, Tables S4–S6), the KL1 estimator performed better both at model prediction and beta coefficient estimation. The performance of the KL1 estimator in case 3 (Supplementary Materials, Section S1, Tables S7–S9) was the same as that in case 1. In the high-dimensional simulation result, the KL1 estimator performed better in model prediction, beta coefficient estimation and sensitivity with regard to detecting the right coefficients using FN and FP (Supplementary Materials, Section S2, Tables S10–S15).
Furthermore, though the GO estimator may compete with the KL1 estimator in the MSEs in terms of estimating the beta coefficients for the low-dimensional data, it is more appropriate to look at the FP and the FN. The FP and the FN proves that the KL1 estimator is preferable to the GO estimator where feature reduction is needed, that is, in cases 1 and 2.

4. Real-Life Application

In this section, we analyzed two datasets to evaluate the performances of the following estimators: the ridge, LASSO (L), E-net (E), GO, Liu-LASSO (LL) and the proposed (KL1) estimators. We used two datasets: the prostate cancer dataset and the asphalt binder dataset. Each dataset was analyzed by splitting them into two—training (80%) and test (20%) data. Model fitting and parameter tunning were carried out via k-fold cross classification, with k = 10 . The tuning parameters were selected from a sequence 0 to 1. We computed the average errors over the folds for each of the tuning parameters. We chose the tuning parameters that minimized the mean square error. The coefficients from the better model were used to predict the response in the test sample, and the model performance was judged via the test mean squared error (TMSE), which is defined as follows:
T M S E = 1 n t e s t i = 1 n t e s t y t e s t x i T β ^ 2
Figure 1 below gives the general framework of the analytical process employed in the analysis of the real-life dataset. This process is followed through for implementing all the estimation techniques outlined in Section 2.

4.1. Dataset I (Prostate Cancer Data)

The prostate data are employed to model the relationship between the level of prostate-specific antigen (lpsa) and a number of clinical measures derived from men who were about to receive a radial prostatectomy [1,21,22]. The model is defined as follows:
y i = β 0 + β 1 x i 1 + β 2 x i 2 + + β 8 x i 8 + ϵ i     i = 1 , 97 ,
where response y i is the level of prostate-specific antigen, x i 1 is the logarithm of cancer volume, x i 2 is the logarithm of prostate weight, x i 3 is age in years, x i 4 is the natural logarithm of the amount of benign prostatic hyperplasia, x i 5 is seminal vesicle invasion, x i 6 is the logarithm of capsule penetration, x i 7 is the Gleason score, and x i 8 is a percentage of Gleason score 4 or 5. Table 1 gives the statistical summary of the dataset and the correlation plot in Figure 2. The response variable was centered, while the predictors were standardized. The following estimators were compared: the ridge estimator, the LASSO estimator, the Liu-LASSO estimator, the GO estimator, the elastic-net (E-net), and the KL1 estimator. The estimated TMSE and the regression coefficients are provided in Table 2.
The results in Table 1 show that the LASSO estimator, Liu-LASSO estimator and E-Net selected six active sets. The GO estimator failed to shrink any of the variables to zero for this particular data, and the ridge estimator, as expected, only performed regularization. The proposed estimator, namely KL1, selected five active sets and possessed the lowest test mean squared error. This makes the KL1 estimator a useful fit for this dataset.

4.2. Dataset 2 (Asphalt Binder Data)

The data were employed to model the effects of some chemical compositions on surface free energy [8]. The model was defined as follows:
y i = β 0 + β 1 x i 1 + β 2 x i 2 + + β 12 x i 12 + ϵ i     i = 1 , 23 ,
where y i denotes surface free energy, x i 1 denotes saturates, x i 2 denotes aromatics, x i 3 denotes resins, x i 4 denotes asphaltenes, x i 5 denotes wax, x i 6 denotes carbon, x i 7 denotes hydrogen, x i 8 denotes oxygen, x i 9 denotes nitrogen, x i 10 denotes sulfur, x i 11 denotes nickel and x i 12 denotes vanadium. The summary statistics are presented in Table 3 and the correlation plot in Figure 3. The response variable is centered, and the predictors are standardized to have zero mean and unit standard deviation before model fitting. The following estimators were compared: the ridge estimator, the LASSO estimator, the Liu-LASSO estimator, the GO estimator, the elastic-net (E-net), and the KL1 estimator. The estimated TMSE and the regression coefficients are provided in Table 4. The ridge estimator did not shrink any of its coefficients, as expected. LASSO estimator, the Liu-LASSO estimator, the E-net estimator and the KL1 estimator selected ten active sets for the regression coefficients. Of all the estimators, the KL1 estimator possessed the smallest test mean squared error.

5. Conclusions

In this study, we proposed a new penalized operator (KL1). It is a shrinkage and feature selection method that adopts the coordinate descent algorithm. The KL1 is an equivalent form of the Elastic net and the Liu-LASSO, which produces a sparse model with high prediction accuracy. The new estimator inherits the selection features of estimators such as Lasso and Elastic-net and equally possesses the stability of the ridge estimator. We applied the new estimator and the existing ones to two dataset (health data and chemical data). The application and simulation results supported the performance of KL1 and its dominance over the ridge estimator, elastic-net, GO, Liu-LASSO and LASSO. Although, the existing methods compete favorably.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/math11234795/s1, Section S1. Quality of Estimators for Low Dimensional Simulated Data (Tables S1–S9); Section S2. Quality of Estimators for High Dimensional Simulated Data (Tables S10–S15); Section S3. Mean Square Error of the Estimated Coefficients for High Dimensional Simulated Data (Figures S1–S6).

Author Contributions

Conceptualization, A.F.L.; methodology, S.L.J. and E.T.A.; software, S.L.J. and E.T.A.; validation, J.A. and A.A.A.; investigation, J.A. and A.A.A.; writing—review & editing, J.A. and A.A.A.; supervision, S.O.; funding acquisition, J.A. and A.A.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Taibah University, grant number 445-9-469.

Data Availability Statement

All data generated or analyzed during this study are included in Supplementary Materials attached to this article: DATA 1_Prostate Cancer data.csv and DATA 2_asphalt Data.csv. And the code for both the simulation and the real-life applications is accessible via this link: https://github.com/Teniola17/Teniola17/tree/main, accessed on 25 September 2023.

Acknowledgments

The authors extend their appreciation to the Deputyship for Research & Innovation, Ministry of Education in Saudi Arabia for funding this research work through project number 445-9-469. Also, the authors would like to extend their appreciation to Taibah University for its supervision support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zou, H.; Hastie, T. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. B Stat. Methodol. 2005, 67, 301–320. [Google Scholar] [CrossRef]
  2. Fu, W.J. Penalized Regressions: The Bridge versus the Lasso. J. Comput. Graph. Stat. 1998, 7, 397–416. [Google Scholar]
  3. Dawoud, I.; Abonazel, M.R.; Awwad, F.A.; Tag Eldin, E. A New Tobit Ridge-Type Estimator of the Censored Regression Model with Multicollinearity Problem. Front. Appl. Math. Stat. 2022, 8, 952142. [Google Scholar] [CrossRef]
  4. Ugwuowo, F.I.; Oranye, H.E.; Arum, K.C. On the jackknife Kibria-Lukman estimator for the linear regression model. Commun. Stat. Simul. Comput. 2021, 1, 1–13. [Google Scholar] [CrossRef]
  5. Idowu, J.I.; Oladapo, O.J.; Owolabi, A.T.; Ayinde, K.; Akinmoju, O. Combating multicollinearity: A new two-parameter approach. Nicel Bilim. Derg. 2023, 5, 90–116. [Google Scholar] [CrossRef]
  6. Lukman, A.F.; Ayinde, K.; Binuomote, S.; Clement, O.A. Modified Ridge-Type Estimator to Combat Multicollinearity: Application to Chemical Data. J. Chemom. 2019, 33, e3125. [Google Scholar] [CrossRef]
  7. Gujarati, D.N. Basic Econometrics, 4th ed.; McGraw-Hill: New York, NY, USA, 2004. [Google Scholar]
  8. Arashi, M.; Asar, Y.; Yüzbaşı, B. SLASSO: A scaled LASSO for multicollinear situations. J. Stat. Comput. Simul. 2021, 91, 3170–3183. [Google Scholar] [CrossRef]
  9. Hoerl, A.E.; Kennard, R.W. Ridge Regression: Biased Estimation for Nonorthogonal Problems. Technometrics 1970, 12, 55–67. [Google Scholar] [CrossRef]
  10. Liu, K. A new class of biased estimate in linear regression. Commun. Stat. 1993, 22, 393–402. [Google Scholar]
  11. Liu, K. Using Liu-type estimator to combat collinearity. Commun. Stat. Theory Methods 2003, 32, 1009–1020. [Google Scholar] [CrossRef]
  12. Şiray, G.Ü.; Toker, S.; Özbay, N. Defining a two-parameter estimator: A mathematical programming evidence. J. Stat. Comput. Simul. 2021, 91, 2133–2152. [Google Scholar] [CrossRef]
  13. Dawoud, İ.; Abonazel, M.R.; Awwad, F.A. Modified Liu estimator to address the multicollinearity problem in regression models: A new biased estimation class. Sci. Afr. 2022, 17, e01372. [Google Scholar] [CrossRef]
  14. Kibria, B.M.G.; Lukman, A.F. A New Ridge-Type Estimator for the Linear Regression Model: Simulations and Applications. Scientifica 2020, 2020, 9758378. [Google Scholar] [CrossRef] [PubMed]
  15. Aladeitan, B.B.; Adebimpe, O.; Lukman, A.F.; Oludoun, O.; Abiodun, O. Modified Kibria-Lukman (MKL) estimator for the Poisson Regression Model: Application and simulation. F1000Research 2021, 10, 548. [Google Scholar] [CrossRef] [PubMed]
  16. Dawoud, I.; Kibria, B.M.G. A new biased estimator to combat the multicollinearity of the gaussian linear regression model. Stat. J. 2020, 3, 526–541. [Google Scholar] [CrossRef]
  17. Wang, H.; Li, G.; Jiang, G. Robust regression shrinkage and consistent variable selection through the LAD-lasso. J. Bus. Econ. Stat. 2007, 25, 347–355. [Google Scholar] [CrossRef]
  18. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B Methodol. 1996, 58, 267–288. [Google Scholar] [CrossRef]
  19. Friedman, J.; Hastie, T.; Höfling, H.; Tibshirani, R. Pathwise coordinate optimization. Ann. Appl. Stat. 2007, 1, 302–332. [Google Scholar] [CrossRef]
  20. Breheny, P.; Huang, J. Coordinate descent algorithms for nonconvex penalized regression, with applications to biological feature selection. Ann. Appl. Stat. 2011, 5, 232. [Google Scholar] [CrossRef]
  21. Genç, M. A new double-regularized regression using Liu and lasso Regularization. Comput. Stat. 2021, 37, 159–227. [Google Scholar] [CrossRef]
  22. Ozkale, M.R.; Kaciranlar, S. The restricted and unrestricted two-parameter estimators. Commun. Stat. Theory Methods 2007, 36, 2707–2725. [Google Scholar] [CrossRef]
  23. Genc, M.; Ozkale, M.R. Usage of the GO estimator in high dimensional linear models. Comput. Stat. 2021, 36, 217–239. [Google Scholar] [CrossRef]
  24. Knight, K.; Fu, W. Asymptotics for lasso-type estimators. Ann. Stat. 2000, 28, 1356–1378. [Google Scholar]
Figure 1. Roadmap for real-life data analysis.
Figure 1. Roadmap for real-life data analysis.
Mathematics 11 04795 g001
Figure 2. Correlation heatmap of prostate data.
Figure 2. Correlation heatmap of prostate data.
Mathematics 11 04795 g002
Figure 3. Correlation heatmap for asphalt dataset.
Figure 3. Correlation heatmap for asphalt dataset.
Mathematics 11 04795 g003
Table 1. Summary statistics of the prostate cancer dataset.
Table 1. Summary statistics of the prostate cancer dataset.
FeatureMeanSDKurtosisSkewness
y2.4783871.1543290.588369−0.00043
lcavol1.350011.178625−0.51681−0.2503
lweight3.6526860.4966315.5288521.216012
age63.865987.4451171.162491−0.82848
lbph0.1003561.450807−1.751440.133813
svi0.2164950.413995−0.045741.398441
lcp−0.179371.39825−0.959090.728634
gleason6.7525770.7221342.6703191.26048
pgg4524.3814428.20403−0.269620.968105
Table 2. Regression coefficients and T M S E .
Table 2. Regression coefficients and T M S E .
Coef.LASSOLiu-LASSOGoestE-NetKL1Ridge
lcavol0.01390.01390.01560.01410.01390.0157
lweight0.97210.96850.97370.96870.96410.9699
age−0.0024−0.0011−0.0049−0.00120.0000−0.0037
lbph0.00060.00060.00070.00060.00060.0007
svi0.00000.0000−0.00150.00000.0000−0.0021
lcp0.00000.0000−0.00210.00000.0000−0.0036
gleason0.00460.00660.00960.00710.00910.0127
pgg45−0.0035−0.0035−0.0036−0.0034−0.0034−0.0035
TMSE2.61672.61092.63292.61252.60382.6284
Table 3. Summary statistics for asphalt dataset.
Table 3. Summary statistics for asphalt dataset.
FeatureMeanSDKurtosisSkewness
y18.42134.104044−0.115391.157349
x18.8173912.8635090.354511−0.56121
x234.621745.639951−0.583270.298716
x340.404357.2728310.4249320.025278
x414.673915.876549−0.59483−0.44598
x52.7660871.417686−1.546060.032829
x684.482611.745293−1.21593−0.1494
x710.364780.3321810.4635330.679561
x80.9665220.2979114.3685681.163691
x90.6995650.252991−0.240030.514575
x104.3717392.15229−1.17692−0.07012
x1171.3478341.79562−1.144990.434363
x12243.1739354.63558.1526152.915543
Table 4. Regression coefficients and M S E y .
Table 4. Regression coefficients and M S E y .
Coef.LASSOLiu-LASSOGoestE-NetKL1Ridge
saturates0.00100.00100.00120.00100.00100.0012
aromatics0.93230.93170.93260.93170.93100.9320
resins−0.0524−0.0523−0.0526−0.0523−0.0521−0.0525
asphaltenes−0.0600−0.0600−0.0600−0.0600−0.0600−0.0600
wax−0.0337−0.0337−0.0338−0.0337−0.0336−0.0338
carbon−0.0070−0.0070−0.0109−0.0066−0.0071−0.0113
hydrogen0.06210.06210.06230.06210.06210.0623
oxygen0.01340.01330.01380.01340.01330.0138
nitrogen0.00000.00000.00000.00000.00000.0041
sulfur0.00000.00000.00980.00000.00000.0123
nickel−0.0027−0.0027−0.0070−0.0025−0.0026−0.0084
vanadium−0.0004−0.0004−0.0003−0.0004−0.0004−0.0003
TMSE9.19059.18499.21279.18629.17819.2165
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lukman, A.F.; Allohibi, J.; Jegede, S.L.; Adewuyi, E.T.; Oke, S.; Alharbi, A.A. Kibria–Lukman-Type Estimator for Regularization and Variable Selection with Application to Cancer Data. Mathematics 2023, 11, 4795. https://doi.org/10.3390/math11234795

AMA Style

Lukman AF, Allohibi J, Jegede SL, Adewuyi ET, Oke S, Alharbi AA. Kibria–Lukman-Type Estimator for Regularization and Variable Selection with Application to Cancer Data. Mathematics. 2023; 11(23):4795. https://doi.org/10.3390/math11234795

Chicago/Turabian Style

Lukman, Adewale Folaranmi, Jeza Allohibi, Segun Light Jegede, Emmanuel Taiwo Adewuyi, Segun Oke, and Abdulmajeed Atiah Alharbi. 2023. "Kibria–Lukman-Type Estimator for Regularization and Variable Selection with Application to Cancer Data" Mathematics 11, no. 23: 4795. https://doi.org/10.3390/math11234795

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop