Next Article in Journal
Identification of Influential Factors in the Adoption of Irrigation Technologies through Neural Network Analysis: A Case Study with Oil Palm Growers
Previous Article in Journal
Biostimulants as a Response to the Negative Impact of Agricultural Chemicals on Vegetation Indices and Yield of Common Buckwheat (Fagopyrum esculentum Moench)
Previous Article in Special Issue
Genomic Prediction of Wheat Grain Yield Using Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Estimation of Error Variance in Genomic Selection for Ultrahigh Dimensional Data

by
Sayanti Guha Majumdar
,
Anil Rai
and
Dwijesh Chandra Mishra
*
Division of Agricultural Bioinformatics, ICAR-Indian Agricultural Statistics Research Institute, New Delhi 110012, India
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Agriculture 2023, 13(4), 826; https://doi.org/10.3390/agriculture13040826
Submission received: 18 March 2022 / Revised: 26 April 2022 / Accepted: 16 May 2022 / Published: 4 April 2023
(This article belongs to the Special Issue Machine Learning and Biological Data in Crop Genetics and Breeding)

Abstract

:
Estimation of error variance in the case of genomic selection is a necessary step to measure the accuracy of the genomic selection model. For genomic selection, whole-genome high-density marker data is used where the number of markers is always larger than the sample size. This makes it difficult to estimate the error variance because the ordinary least square estimation technique cannot be used in the case of datasets where the number of parameters is greater than the number of individuals (i.e., p > n). In this article, two existing methods, viz. Refitted Cross Validation (RCV) and kfold-RCV, were suggested for such cases. Moreover, by considering the limitations of the above methods, two new methods, viz. Bootstrap-RCV and Ensemble method, have been proposed. Furthermore, an R package “varEst” has been developed, which contains four different functions to implement these error variance estimation methods in the case of Least Absolute Shrinkage and Selection Operator (LASSO), Least Squares Regression (LSR) and Sparse Additive Models (SpAM). The performances of the algorithms have been evaluated using simulated and real datasets.

1. Introduction

Genomic Selection (GS) is an advanced form of marker-assisted selection used in the breeding of animals and plants. GS was first described by Meuwissen et al. in 2001 [1]. GS increases genetic gain by reducing the duration of the breeding cycle. Moreover, in GS, there is no need to have information about the exact location of the associated gene. Markers from the whole genome are used in GS to estimate the breeding value, which is known as Genomic Estimated Breeding Value (GEBV), used for the selection of individuals. So, GS deals with millions of markers all over the genome, and this leads to a statistical issue known as p > n problem, where p is the number of variables (markers) and n is the sample size (number of individuals).
There are several statistical models available to estimate the GEBV in the case of GS. However, the performance of these models needs to be evaluated properly. For that, we need to estimate the error variance of each model. However, error variance estimation is a difficult problem in additive models when p > n. Standard least squares estimation techniques cannot be applied in this case. Poor estimation of error variance of a genomic selection model also leads to poor estimation of the GEBV in GS. In order to resolve this problem, several variance estimators have been proposed but mostly all suffer large biases in finite samples. Fan et al. proposed the Refitted Cross Validation (RCV) method in 2012 to overcome the downward bias of the Least Absolute Shrinkage and Selection Operator (LASSO) estimator for the estimation of error variance [2]. They have also mentioned a modification of the RCV method called k-fold RCV where the dataset is divided into k numbers of subgroups instead of two groups as in the case of RCV. Other error variance estimation methods include the oracle estimator, residual sum of squares-based estimators [2], cross-validation based estimators [2], Smoothly Clipped Absolute Deviation Penalty (SCAD) estimator [3], scaled sparse linear regression estimators [4] and method of moments estimators [5]. Chen et al. (2018) have proposed an estimator for error variance in an ultrahigh dimensional sparse additive model by effectively integrating sure independence screening and Refitted Cross Validation techniques [6]. Yu and Bien have proposed the natural LASSO estimator and organic LASSO estimator for the error variance estimation in the high dimensional linear model [7]. A ridge regression and random matrix theory-based novel estimator of error variance has been proposed by Liu et al. [8]. A residual-based efficient estimator for error variance has been developed by Li et al., 2020 [9].
However, hardly any review is available related to the application of these methods in variance estimation for genomic selection. In this article, it is probably the first time we have used the algorithm of RCV method and k-fold RCV method proposed by Fan et al. in 2012 in genomic selection for variance estimation. Moreover, we have extended the work done by Fan et al. and developed a new algorithm for variance estimation called “bootstrap RCV”. It has been observed that sometimes particularly in big data, such as genotypic data, RCV method suffers from a computational problem. In order to resolve this problem, we have developed a new computationally efficient algorithm based on the Ensemble approach. Performances of these algorithms have been tested using simulated data for different genetic architectures (different heritability and epistasis level). All these algorithms are implemented in R to develop an R package “varEst” [10] for estimating the error variance of Genomic Selection models.

2. Materials and Methods

2.1. Simulated Dataset

For this study, 24 genomic datasets, including both genotype and phenotype data, were simulated. There were 10 true features known as Quantitative Trait Loci (QTLs) in the datasets. An R package QTL Bayesian interval mapping (“qtlbim”) [11,12] was used to simulate the genome with 10 chromosomes. The package qtlbim follows Cockerham’s model for simulating quantitative trait loci with epistasis [13]. The Cockerham’s model for quantitative trait controlled by two epistatic genes A and B, from a sample of size n of an F2 population, the trait value of the k th individual with genotype i j can be described as follows:
y i j k = G i j + ε i j k = μ + a 1 x 1 + d 1 z 1 + a 2 x 2 + d 2 z 2 + i a a w a a + i a d w a d + i d a w d a + i d d w d d + ε i j k
where G i j denotes the genotypic value of the genotype   i j , μ is the mean, a 1 is the additive effect of locus A, d 1 is the dominance effect of locus A, a 2 is the additive effect of locus B, d 2 is the dominance effect of locus B, i a a is additive × additive effect of loci A and B, i a d is additive × dominance effect of loci A and B, i d a is dominance × additive effect of loci A and B, i d d is dominance × dominance effect of loci A and B. The coded variables are defined as
x 1 = {   1   i f   A   i s   A A   0   i f   A   i s   A a 1   i f   A   i s   a a   x 2 = {   1   i f   B   i s   B B   0   i f   B   i s   B b 1   i f   B   i s   b b
z 1 = { 1 2   i f   A   i s   A a 1 2   o t h e r w i s e z 2 = { 1 2   i f   B   i s   B b 1 2   o t h e r w i s e
w a a = x 1 × x 2       w a d = x 1 × z 2       w d a = z 1 × x 2       w d d = z 1 × z 2
and ε i j k is residual of the k th individual with genotype i j .
The genome contains 1000 markers, equally spaced over the chromosomes, and each chromosome has 100 markers. The phenotypic values are normally distributed and generated on the basis of 10 QTLs (true features). There are two categories of datasets, one of which is additive in nature and does not contain any epistatic effect, and the other category contains both additive and epistatic effects. So, the additive datasets have 10 QTLs, one each in every chromosome, whereas non-additive datasets have one pair of QTLs in 5 of the 10 chromosomes, the rest of the 5 chromosomes do not contain any QTL. In this way, 24 different datasets with different genetic architecture (i.e., with different levels of narrow-sense heritability (0.1, 0.2, 0.3, 0.5, 0.7 and 0.9) and epistatic effects, i.e., additive × additive effect (0, 5, 10 and 15)) are generated with genotypic and phenotypic information for F2 population. Each dataset contains 200 individuals and 1000 biallelic markers. A total of 500 replicates were simulated for each genetic architecture to predict phenotype (GEBV).

2.2. Real Dataset

We have two real datasets for this study, one is for wheat, and another is maize data [14]. Wheat lines were genotyped by Triticarte Pty. Ltd. (Canberra, Australia) using 1447 Diversity Array Technology. This data set includes 599 lines observed for trait grain yield (GY) for four mega environments. However, for our convenience, we have just considered GY for the first mega environment. The final number of DArT markers after editing was 1279; hence the same has been used in this study. The second dataset, which is the maize dataset, is generated by CIMMYT’s Global Maize Program. It originally included 300 maize lines with 1148 SNP markers. Here also, the trait under study is GY, evaluated under drought and watered conditions. After some editing, 264 maize lines with 1135 SNPs markers were available for the final study. The editing was done to improve the quality of the data by removing missing genotypes and outliers.

2.3. Methods for Error Variance Estimation

In this section, details of the algorithm for Refitted Cross Validation (RCV) and k-fold Refitted Cross Validation (k-RCV) method of error variance estimation in genomic selection are described.

2.3.1. Refitted Cross Validation (RCV)

The Refitted Cross Validation method is originally given by Fan et al. in 2012 for error variance estimation in ultrahigh dimensional regression [2]. A sample of size n is assumed and split randomly into two groups. In the first stage, the most relevant variables were selected with the help of an ultrahigh dimensional variable selection method from those two datasets separately. This step results in two small sets of important variables (markers). Here, in this study, we have applied SpAM [15], LASSO [16] and Linear Least Squares Regression as feature selection methods. The second stage involves the re-estimation of the regression coefficient β and variance σ2 with the help of the ordinary least squares regression method. For this purpose, the variables selected from the first dataset are considered to estimate the variance from the second dataset and vice-versa. Finally, the average of those two variances, which we get from the second stage, is considered the final estimator of σ2 (Figure 1). So, in the second stage, refitting is done to reduce the effect of spurious variables in the datasets, and also it takes care of the p > n problem in the high dimensional genomic datasets. The algorithm of the RCV method has been described below:
  • Consider a dataset ( X n × p , Y n × 1 ) . We denote it as S ;
  • S is divided into two groups randomly, each subsample having size of n/2;
    S = { S ( 1 ) , S ( 2 ) }
    where S ( 1 ) = ( X ( n / 2 ) × p ( 1 ) , Y ( n / 2 ) × 1 ( 1 ) ) and S ( 2 ) = ( X ( n / 2 ) × p ( 2 ) , Y ( n / 2 ) × 1 ( 2 ) ) .
  • Variable selection has been performed on S ( 1 ) using SpAM or LASSO or Least Squares Regression and let M ( 1 ) is a set of selected variables from subsample S ( 1 ) ;
  • Using only the selected variables i.e., M ( 1 ) , error variance is estimated through S ( 2 ) with ordinary least squares estimation;
    σ ^ ( 2 ) 2 = Y ( 2 ) T ( I n 2 P M ( 1 ) ( 2 ) ) Y ( 2 ) n 2 | M ( 1 ) |
    where P M ( 1 ) ( 2 ) = X M ( 1 ) ( 2 ) ( X M ( 1 ) ( 2 ) T X M ( 1 ) ( 2 ) ) 1 X M ( 1 ) ( 2 ) T .
  • In the next stage, variable selection has been performed on S ( 2 ) using Sparse Additive Models (SpAM) or LASSO or Least Squares Regression and M ( 2 ) is a set of selected variables from subsample S ( 2 ) ;
  • Using only the selected variables i.e., M ( 2 ) , variance is estimated in S ( 1 ) with ordinary least squares estimation;
    σ ^ ( 1 ) 2 = Y ( 1 ) T ( I n 2 P M ( 2 ) ( 1 ) ) Y ( 1 ) n 2 | M ( 2 ) |
    where P M ( 2 ) ( 1 ) = X M ( 2 ) ( 1 ) ( X M ( 2 ) ( 1 ) T X M ( 2 ) ( 1 ) ) 1 X M ( 2 ) ( 1 ) T .
  • The final estimator of variance is
    σ ^ R C V 2 = σ ^ ( 1 ) 2 + σ ^ ( 2 ) 2 2
Let error variance estimator of SpAM, LASSO and Least Squares Regression is represented by σ ^ S p 2 , σ ^ L A S S O 2 and σ ^ L S R 2 , respectively. In the above algorithm, σ ^ R C V 2 represents σ ^ S p 2 , σ ^ L A S S O 2 and σ ^ L S R 2 . This means all the three variances from three models ( σ S p 2 , σ ^ L A S S O 2 and σ ^ L S R 2 ) will be obtained by following the above algorithm of RCV.

2.3.2. k-Fold Refitted Cross Validation (k-RCV)

k-fold RCV is an extended version of the original RCV method [2]. In this case, the data is divided into k equal size groups instead of 2 groups. Variables are selected from one group and variance is estimated from rest of the k-1 groups. Likewise, all the groups are covered and, in the end, the average value of all the variances from each group is the final error variance (Figure 2). The algorithm is described below.
  • Consider a dataset ( X n × p , Y n × 1 ) ;
  • The dataset is divided into k groups of size m × p so that m k = n . The new datasets are denoted as S ( i ) = ( X m × p ( i ) , Y m × 1 ( i ) ) , i = 1 , 2 , , k ;
  • Let S = { S ( i ) } ,   i = 1 , 2 , , k , and S ( i ) denotes the ith subsample;
  • S ( i ) = S S ( i ) , i.e., S ( i ) is a subset of S excluding i th subsample;
  • Variable selection has been done on all subsamples of set S ( i ) ;
  • V ( i ) is a set of selected variables on subsample S ( i ) , written as V ( i ) = { M ( r ) } , i , r = 1 , 2 , , k and r i , where M ( r ) is set of selected variables from subsamples S ( i ) , ( i r );
  • Now, the ordinary least squares method is used to estimate variance in S ( i ) using only the selected variables V ( i ) . Thus, we get k 1 variances i.e., one each from S ( i ) , i = 1 , 2 , , k −1.
    σ ^ ( r ) ( i ) 2 = Y ( i ) T ( I n k P M ( r ) ( i ) ) Y ( i ) n k | M ( r ) |
    where P M ( r ) ( i ) = X M ( r ) ( i ) ( X M ( r ) ( i ) T X M ( r ) ( i ) ) 1 X M ( r ) ( i ) T
    σ ^ ( i ) 2 = 1 k 1 r = 1 k 1 σ ^ ( r ) ( i ) 2
  • So, the final estimator of variance is
    σ ^ R C V k 2 = 1 k i = 1 k σ ^ ( i ) 2
Variables are selected in step v with the help of three feature selection methods separately using SpAM, LASSO and Least Squares Regression models. Let error variance estimator of SpAM, LASSO and Least Squares Regression is represented by σ ^ S p 2 , σ ^ L A S S O 2 and σ ^ L S R 2 respectively. In the above algorithm, σ ^ R C V k 2 represents σ S p 2 , σ ^ L A S S O 2 and σ ^ L S R 2 This means all the three variances from three models ( σ S p 2 , σ ^ L A S S O 2 and σ ^ L S R 2 ) will be obtained by following above algorithm of k-RCV.

2.3.3. Proposed Algorithm for Error Variance Estimation

Bootstrap with RCV

In this method, bootstrap samples are taken from the original datasets and then RCV method is applied to each of these bootstrap samples (Figure 3). The algorithm for this method can be written as:
  • Consider a dataset ( X n × p , Y n × 1 ) . We denote it as S ;
  • A bootstrap sample is selected by drawing a sample of size n by simple random sampling with replacement from the actual dataset S . The bootstrap sample is denoted by S ( i ) * = ( X * n × p ( i ) , Y * n × 1 ( i ) ) , i = 1 , 2 , , k . Thus, k numbers of bootstrap samples each of size n have been drawn, S * = { S ( i ) * } ;
  • Now each S ( i ) * is divided into two groups randomly, each subsample having size of n/2;
    S ( i ) * = { S * ( i ) ( 1 ) , S * ( i ) ( 2 ) }
    where S * ( i ) ( 1 ) = ( X * ( n / 2 ) × p ( 1 ) , Y * ( n / 2 ) × 1 ( 1 ) ) and S * ( i ) ( 2 ) = ( X * ( n / 2 ) × p ( 2 ) , Y * ( n / 2 ) × 1 ( 2 ) )
  • Variable selection has been performed on S * ( i ) ( 1 ) and M ( i ) ( 1 ) is a set of selected variables from subsample S * ( i ) ( 1 ) ;
  • Using only the selected variables variance is estimated in S * ( i ) ( 2 ) with ordinary least squares estimation;
    σ ^ * ( i ) ( 1 ) 2 = Y * ( 2 ) T ( I n 2 P M ( i ) ( 1 ) ( 2 ) ) Y * ( 2 ) n 2 | M ( i ) ( 1 ) |
    where P M ( i ) ( 1 ) ( 2 ) = X M ( i ) ( 1 ) * ( 2 ) ( X M ( i ) ( 1 ) * ( 2 ) T X M ( i ) ( 1 ) * ( 2 ) ) 1 X M ( i ) ( 1 ) * ( 2 ) T .
  • In the next stage, variable selection has been performed on S * ( i ) ( 2 ) and M ( i ) ( 2 ) is a set of selected variables from subsample S * ( i ) ( 2 ) ;
  • Using only the selected variables variance is estimated in S * ( i ) ( 1 ) with ordinary least squares estimation;
    σ ^ * ( i ) ( 2 ) 2 = Y * ( 1 ) T ( I n 2 P M ( i ) ( 1 ) ( 1 ) ) Y * ( 1 ) n 2 | M ( i ) ( 2 ) |
    where P M ( i ) ( 2 ) ( 1 ) = X M ( i ) ( 2 ) * ( 1 ) ( X M ( i ) ( 2 ) * ( 1 ) T X M ( i ) ( 2 ) * ( 1 ) ) 1 X M ( i ) ( 2 ) * ( 1 ) T
  • For S ( i ) * ,we get the estimator of variance as
    σ ^ * ( i ) 2 = σ ^ * ( i ) ( 1 ) 2 + σ ^ * ( i ) ( 2 ) 2 2
  • The final estimator of variance is
    σ ^ R C V b 2 = 1 k i = 1 k σ ^ * ( i ) 2
Variable selection is performed with the help of three feature selection methods separately using SpAM, LASSO and Least Squares Regression models. Let error variance estimator of SpAM, LASSO and Least Squares Regression is represented by σ ^ S p 2 , σ ^ L A S S O 2 and σ ^ L S R 2 , respectively. In the above algorithm, σ ^ R C V b 2 represents σ S p 2 , σ ^ L A S S O 2 and σ ^ L S R 2 . This means all the three variances from three models ( σ S p 2 , σ ^ L A S S O 2 and σ ^ L S R 2 ) will be obtained by following the above algorithm of bootstrap-RCV.

Ensemble Method

In this method, both bootstrapping and simple random sampling without replacement are combined to estimate error variance. Variables are selected from the original datasets, and all possible samples of a particular size are taken from the selected variables set with simple random sampling without replacement. With these samples of selected variables, error variance is estimated from bootstrap samples of the original datasets using the Least Squares Regression method. Finally, the average of all the estimated variances is considered the final estimate of the error variance (Figure 4). The algorithm is as follows:
  • Consider a dataset ( X n × p , Y n × 1 ) . We denote it as S ;
  • A bootstrap sample is selected by drawing a sample of size n by simple random sampling with replacement from the actual dataset S . The bootstrap sample is denoted by S ( i ) * = ( X * n × p ( i ) , Y * n × 1 ( i ) ) , i = 1 , 2 , , k . Thus, k numbers of bootstrap samples each of size n have been drawn, S * = { S ( i ) * } ;
  • Variable selection has been performed on S and M ( 1 ) is a set of selected variables from S ;
  • All possible subsamples of size ( | M ( 1 ) | / 2 ) is drawn from M ( 1 ) by simple random sampling without replacement. The subsamples of these selected variables are denoted by M ( r ) , r = 1 , 2 , , L ;
  • Using only the subsample of selected variables, i.e., M ( r ) , variance is estimated in S ( i ) * with ordinary least squares estimation;
    σ ^ ( i ) ( r ) 2 = Y * ( i ) T ( I n P M ( r ) ( i ) ) Y * ( i ) n | M ( r ) |
    where P M ( r ) ( i ) = X M ( r ) * ( i ) ( X M ( r ) * ( i ) T X M ( r ) * ( i ) ) 1 X M ( r ) * ( i ) T
  • The estimated variance for all M ( r ) is
    σ ^ ( i ) 2 = 1 L r = 1 L σ ^ ( i ) ( r ) 2
  • The final estimator of variance is
    σ ^ e n s e m b l e 2 = 1 k i = 1 k σ ^ ( i ) 2
Variables are selected with the help of three feature selection methods separately using SpAM, LASSO and Least Squares Regression models. Let error variance estimator of SpAM, LASSO and Least Squares Regression is represented by σ ^ S p 2 , σ ^ L A S S O 2 and σ ^ L S R 2 , respectively. In the above algorithm, σ ^ e n s e m b l e 2 represents σ S p 2 , σ ^ L A S S O 2 and σ ^ L S R 2 . This means all the three variances from the three models ( σ S p 2 , σ ^ L A S S O 2 and σ ^ L S R 2 ) will be obtained by following the above algorithm of the Ensemble method.

2.4. Development of R Package “varEst” for Implementation of the Proposed Algorithms

The objective of “varEst” package is to estimate the error variance of fitted genomic selection models from ultrahigh dimensional genomic datasets. “varEst” contains four different functions for variance estimation, viz. rcv, krcv, bsrcv and ensemble (Table 1), which are based on four different algorithms, viz. Refitted Cross Validation (RCV), k-fold Refitted Cross Validation (k-RCV), Bootstrap with RCV and Ensemble method, respectively. In each function, there are options to choose the variable selection method among SpAM (“spam”), Least Squares Regression (“lsr”) and LASSO (“lasso”). The package “varEst” is available on CRAN at https://CRAN.R-project.org/package=varEst (accessed on 10 January 2022) for installation and use.

2.5. Application in Simulated and Real Dataset

From these simulated data sets, the GEBV was estimated using statistical models, viz. SpAM, LASSO and Linear Least Squares Regression using R. In the case of SpAM, the samQL function of the SAM package [17] was used to select 10 highly significant markers as we knew the number of true features (QTL) was 10 in each dataset. These significant markers were used for the prediction of the GEBV of the testing dataset with the help of the predict function using the SpAM fitted model on the training dataset. In order to implement the LASSO method, the glmnet function of the glmnet package [18] in R is used with default parameter values. Markers with non-zero coefficients are chosen as selected features, and among the non-zero coefficients, we can choose highly significant 10 markers on the basis of their coefficient value. In order to fit the Linear Least Squares Regression model, the approach of Meuwissen et al. (2001) was followed because, in datasets, we have more markers than the number of individuals. So, according to this approach, the simple linear regression model was fitted for each of the markers and QTL separately, i.e., we have fitted 1010 simple linear regression models. Then 100 of those markers and QTLs with the most significant p-values were selected. Further, a final linear regression model was fitted with these 100 markers simultaneously using the lm function in the stats package in R [19]. Then, from the final fitted model, 10 highly significant markers/QTLs are chosen as the final selected features. Finally, phenotypic values were predicted for testing the data set by using available marker data and estimated regression coefficient of marker effects.
In the next step, error variance was estimated for each model using four estimation methods, viz, RCV, k-RCV, bs-RCV and Ensemble. For that purpose, R functions have been used from the package “varEst” for all four methods. Moreover, the true value of error variance was estimated as we already know the true features from the simulated datasets. So, using only the true features (QTLs), the error variance was estimated with the ordinary least squares regression method. Then, the performances of the methods were compared among themselves to find the better method that could minimize the difference between the error variance estimated by the suggested methods and the true error variance.
Likewise, all four methods (viz. RCV, k-RCV, bootstrap RCV and Ensemble) have been applied in real datasets for estimation of error variance of estimating GEBV with SpAM, LASSO and Least Squares Regression.

3. Results and Discussion

In this article, one R package called “varEst” for error variance estimation methods, viz. RCV, k-RCV, bootstrap RCV and Ensemble method has been developed. These methods have been compared on the basis of their performances in ultra-high dimensional simulated genomic datasets as well as real datasets, and the performance is evaluated for SpAM, LASSO and Least Squares Regression models. The estimated error variances from the simulated datasets are compared with the calculated true error variance of the respective datasets. The true error variance has been estimated by ordinary least squares estimate after incorporating only the true features (which is 10 QTLs in this case) into the model. The performance of SpAM is represented graphically in Figure 5. The comparison between the four error variance estimation method using SpAM as the variable (marker) selection method is shown in Figure 5. It can be observed that in a low heritability level, the Ensemble method performs better than the other three methods, whereas, in the case of higher heritability, the RCV method gives the error variance closer to the true value. It can also be observed that the error variances increase with the increase in epistatic effect in the datasets. This may be due to the fact that when the epistatic effect is present in the datasets, the additive or linear models like SpAM, LASSO, and Least Squares Regression cannot select the most relevant features (markers) from the datasets and leads to increased error variance.
In Figure 6, the performance of the four error variance estimation methods using LASSO as variable (marker) selection method are compared. It can be noticed from Figure 6 that the performances of all the error variance estimation methods have performed similar and the error variances in this case are almost same as the true error variance except in high heritability scenario (i.e., 0.7 and 0.9). Moreover, the error variances tend to increase when epistatic effect increases in the datasets.
Figure 7 shows the comparative performances of RCV, k-RCV, Bootstrap RCV and Ensemble method where Least Squares Regression is used as the variable selection method in the datasets. It is observed that the k-RCV method gives the lowest value of error variance which is closer to the true variance, whereas, in case of SpAM and LASSO, the lowest value of error variance is estimated by RCV method. Like the previous cases, in this case also the error variances increase with the increase in epistatic effects in the datasets. It is also evident from the results that the error variance of LASSO is lowest among the three models, i.e., SpAM, LASSO and LSR. Though the error variance estimated by ensemble method is higher than the value of error variance estimated by RCV-based methods, the trend it follows in different genetic architectures is similar to that of RCV-based methods.
In the case of real datasets, the results are shown in Table 2 and Table 3. In real datasets, error variance could not be estimated with all the methods, such as SpAM and LSR, whereas LASSO gives results for all four methods. Basically, the problem arises in RCV and RCV-based methods due to data splitting, which is the fundamental step in the case of RCV. Here, data splitting refers to dividing the whole dataset into two or k numbers of subsets so that the sample size (i.e., n) is reduced, but the number of the predictor variable (i.e., the number of markers (p) in this case) remains same in each subset (Fan et al., 2012). In the case of genomic selection, the data we are dealing with is marker data which is categorical in nature (i.e., having values 1, 2 and 3 in our study). It has been observed that after data splitting, a singularity issue arises in matrix operations on those new subsets having a small sample size. This is because, in most cases, two or more columns become identical to each other or have the same genotype value as the number of individuals, i.e., the sample size has been reduced by data splitting. However, we can observe that the Ensemble method performs well in the case of SpAM, LASSO and also LSR in real datasets. So, we can infer that the Ensemble method can handle the shortcomings of RCV and RCV-based methods and gives an error variance estimate for all three genomic selection models. All the R scripts for data simulation and four error variance estimation methods and their results are available as Supplementary files (Script S1–S7).
Here, we have proposed four error variance estimation methods for genomic selection. There are several statistical models available for genomic selection. Some models perform well in additive genetic architecture, whereas some of them perform better in the presence of the epistatic effect. However, in the practical situation, we do not know what level of epistatic effect and additive effect is present in the data. The performances of genomic selection models are largely dependent on the genetic architecture of the data. So, it is necessary to know which statistical model to use for a particular data or particular trait. Here comes the applicability of the error variance estimation of a model. For a particular trait, we can estimate the error variance of available genomic selection models, and we can choose the model with the lowest error variance. Estimation of error variance is also helpful to know whether to use linear or non-linear genomic selection models.

4. Conclusions

In this study, four algorithms for error variance estimation in genomic selection for ultrahigh dimensional data have been suggested. These algorithms are implemented in an R package “varEst” which is made available in CRAN for free public access. The performances of the four functions available in “varEst” are evaluated based on both simulated and real datasets. From the performances of these functions, it is evident that the proposed Ensemble method can eliminate some of the shortcomings of RCV and RCV-based methods in the case of real datasets. This approach is suitable when the sample size is small. In a small sample size, using RCV and RCV-based methods are not suitable as these methods involve data splitting. The performance of the error variance estimation methods also depends on the variable selection accuracy of the genomic selection models. The models that can select more accurate variables or markers, the less is the error variance of the model. Thus, the estimation of error variance is also useful for assessing the genomic selection models.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/agriculture13040826/s1, Script S1: Data simulation; Script S2: Data simulation; Script S3: RCV; Script S4: kRCV; Script S5: Bootstrap RCV; Script S6: Ensemble method; Script S7: Results from simulated data.

Author Contributions

Conceptualization, A.R.; Formal analysis, S.G.M.; Investigation, S.G.M. and A.R.; Methodology, S.G.M., A.R. and D.C.M.; Software, S.G.M.; Supervision, A.R. and D.C.M.; Writing—original draft, S.G.M. and A.R.; Writing—review and editing, A.R. and D.C.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Scripts for simulation of the data are available as Supplementary files.

Acknowledgments

We acknowledge the support from the Network Project for Agricultural Bioinformatics and Computational Biology under the CABin Scheme, Indian Council of Agricultural Research (ICAR), New Delhi.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Meuwissen, T.H.E.; Hayes, B.J.; Goddard, M.E. Prediction of total genetic value using genome-wide dense marker maps. Genetics 2001, 157, 1819. [Google Scholar] [CrossRef] [PubMed]
  2. Fan, J.; Guo, S.; Hao, N. Variance estimation using refitted cross-validation in ultrahigh dimensional regression. J. R. Stat. Soc. Ser. B. 2012, 74, 37–65. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Fan, J.; Li, R. Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties. J. Am. Stat. Assoc. 2001, 96, 1348–1360. [Google Scholar] [CrossRef]
  4. Städler, N.; Bühlmann, P.; Geer, S.; Städler, N.; Bühlmann, P.; Geer, S. ℓ1-penalization for mixture regression models. TEST 2010, 19, 209–256. [Google Scholar] [CrossRef] [Green Version]
  5. Dicker, L.H. Variance estimation in high-dimensional linear models. Biometrika 2014, 101, 269–284. [Google Scholar] [CrossRef]
  6. Chen, Z.; Fan, J.; Li, R. Error Variance Estimation in Ultrahigh-Dimensional Additive Models. J. Am. Stat. Assoc. 2018, 113, 315–327. [Google Scholar] [CrossRef] [PubMed]
  7. Yu, G.; Bien, J. Estimating the error variance in a high-dimensional linear model. Biometrika 2019, 106, 533–546. [Google Scholar] [CrossRef]
  8. Liu, X.; Zheng, S.; Feng, X. Estimation of error variance via ridge regression. Biometrika 2020, 107, 481–488. [Google Scholar] [CrossRef]
  9. Li, Z.; Lin, W. Efficient error variance estimation in non-parametric regression. Aust. N. Z. J. Stat. 2020, 62, 467–484. [Google Scholar] [CrossRef]
  10. Guha Majumdar, S.; Rai, A.; Mishra, D.C. varEst: Variance Estimation, Version 0.1.0; R Package; 2019. Available online: https://CRAN.R-project.org/package=varEst (accessed on 10 January 2022).
  11. Yandell, B.S.; Mehta, T.; Banerjee, S.; Shriner, D.; Venkataraman, R.; Moon, J.Y.; Neely, W.W.; Wu, H.; von Smith, R.; Yi, N. R/qtlbim: QTL with Bayesian Interval Mapping in experimental crosses. Bioinformatics 2007, 23, 641–643. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Yandell, B.; Nengjun, Y.; Mehta, T.; Banerjee, S.; Shriner, D.; Venkataraman, R.; Moon, J.; Neely, W.; Wu, H.; Smith, R. qtlbim: QTL Bayesian Interval Mapping, Version 2.0.5; R Package; 2012. Available online: https://pages.stat.wisc.edu/~yandell/qtl/software/qtlbim/ (accessed on 10 January 2022).
  13. Kao, C.H.; Zeng, Z.B. Modeling epistasis of quantitative trait loci using Cockerham’s model. Genetics 2002, 160, 1243. [Google Scholar] [CrossRef] [PubMed]
  14. Crossa, J.; De Los Campos, G.; Pérez, P.; Gianola, D.; Burgueño, J.; Araus, J.L.; Makumbi, D.; Singh, R.P.; Dreisigacker, S.; Yan, J.; et al. Prediction of Genetic Values of Quantitative Traits in Plant Breeding Using Pedigree and Molecular Markers. Genetics 2010, 186, 713. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Ravikumar, P.; Lafferty, J.; Liu, H.; Wasserman, L. Sparse additive models. J. R. Stat. Soc. Ser. B 2009, 71, 1009–1030. [Google Scholar] [CrossRef]
  16. Tibshirani, R. Regression Shrinkage and Selection Via the Lasso. J. R. Stat. Soc. Ser. B 1996, 58, 267–288. [Google Scholar] [CrossRef]
  17. Zhao, T.; Li, X.; Liu, H.; Roeder, K. SAM: Sparse Additive Modelling, Version 1.0.5; R Package; 2014. Available online: https://CRAN.R-project.org/package=SAM (accessed on 10 January 2022).
  18. Friedman, J.; Hastie, T.; Tibshirani, R. Regularization Paths for Generalized Linear Models via Coordinate Descent. J. Stat. Softw. 2010, 33, 1–22. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2019. [Google Scholar]
Figure 1. Refitted Cross Validation (RCV).
Figure 1. Refitted Cross Validation (RCV).
Agriculture 13 00826 g001
Figure 2. k-fold Refitted Cross Validation (k-RCV).
Figure 2. k-fold Refitted Cross Validation (k-RCV).
Agriculture 13 00826 g002
Figure 3. Bootstrap with RCV.
Figure 3. Bootstrap with RCV.
Agriculture 13 00826 g003
Figure 4. Ensemble method.
Figure 4. Ensemble method.
Agriculture 13 00826 g004
Figure 5. Comparison of RCV, k-RCV, bs-RCV and Ensemble method for SpAM. RCV: Refitted Cross Validation, k-RCV: kfold Refitted Cross Validation, bs-RCV: Bootstrap RCV, SpAM: Sparse Additive Models.
Figure 5. Comparison of RCV, k-RCV, bs-RCV and Ensemble method for SpAM. RCV: Refitted Cross Validation, k-RCV: kfold Refitted Cross Validation, bs-RCV: Bootstrap RCV, SpAM: Sparse Additive Models.
Agriculture 13 00826 g005
Figure 6. Comparison of RCV, k-RCV, bs-RCV and Ensemble method for LASSO. RCV: Refitted Cross Validation, k-RCV: kfold Refitted Cross Validation, bs-RCV: Bootstrap RCV, LASSO: Least Absolute Shrinkage and Selection Operator.
Figure 6. Comparison of RCV, k-RCV, bs-RCV and Ensemble method for LASSO. RCV: Refitted Cross Validation, k-RCV: kfold Refitted Cross Validation, bs-RCV: Bootstrap RCV, LASSO: Least Absolute Shrinkage and Selection Operator.
Agriculture 13 00826 g006
Figure 7. Comparison of RCV, k-RCV, bs-RCV and Ensemble method for Least Squared Regression. RCV: Refitted Cross Validation, k-RCV: kfold Refitted Cross Validation, bs-RCV: Bootstrap RCV, LSR: Least Squares Regression.
Figure 7. Comparison of RCV, k-RCV, bs-RCV and Ensemble method for Least Squared Regression. RCV: Refitted Cross Validation, k-RCV: kfold Refitted Cross Validation, bs-RCV: Bootstrap RCV, LSR: Least Squares Regression.
Agriculture 13 00826 g007
Table 1. Available functions in package “varEst”.
Table 1. Available functions in package “varEst”.
Function NameOutput
rcvError variance ( σ ^ R C V 2 )
krcvError variance ( σ ^ R C V k 2 )
bsrcvError variance ( σ ^ R C V b 2 )
ensembleError variance ( σ ^ e n s e m b l e 2 )
Table 2. Error variance estimation in maize dataset.
Table 2. Error variance estimation in maize dataset.
SpAMLASSOLSR
RCV-1.4324-
kRCV-1.4298-
bsRCV-1.3548-
Ensemble1.82881.65902.5848
Note: RCV: Refitted Cross Validation, kRCV: kfold Refitted Cross Validation, bsRCV: Bootstrap RCV, LSR: Least Squares Regression, LASSO: Least Absolute Shrinkage and Selection Operator, SpAM: Sparse Additive Models.
Table 3. Error variance estimation in wheat dataset.
Table 3. Error variance estimation in wheat dataset.
SpAMLASSOLSR
RCV0.78840.79650.7993
kRCV-0.7328-
bsRCV0.84700.83920.8319
Ensemble1.00040.99930.9996
Note: RCV: Refitted Cross Validation, kRCV: kfold Refitted Cross Validation, bsRCV: Bootstrap RCV, LSR: Least Squares Regression, LASSO: Least Absolute Shrinkage and Selection Operator, SpAM: Sparse Additive Models.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guha Majumdar, S.; Rai, A.; Mishra, D.C. Estimation of Error Variance in Genomic Selection for Ultrahigh Dimensional Data. Agriculture 2023, 13, 826. https://doi.org/10.3390/agriculture13040826

AMA Style

Guha Majumdar S, Rai A, Mishra DC. Estimation of Error Variance in Genomic Selection for Ultrahigh Dimensional Data. Agriculture. 2023; 13(4):826. https://doi.org/10.3390/agriculture13040826

Chicago/Turabian Style

Guha Majumdar, Sayanti, Anil Rai, and Dwijesh Chandra Mishra. 2023. "Estimation of Error Variance in Genomic Selection for Ultrahigh Dimensional Data" Agriculture 13, no. 4: 826. https://doi.org/10.3390/agriculture13040826

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop