Next Article in Journal
A Global Identification of Protein Disulfide Isomerases from ‘duli’ Pear (Pyrus betulaefolia) and Their Expression Profiles under Salt Stress
Previous Article in Journal
Population Screening for Hereditary Haemochromatosis—Should It Be Carried Out, and If So, How?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Penalized Regression Method for Genomic Prediction Reduces Mismatch between Training and Testing Sets

by
Osval A. Montesinos-López
1,
Cristian Daniel Pulido-Carrillo
1,
Abelardo Montesinos-López
2,*,
Jesús Antonio Larios Trejo
3,
José Cricelio Montesinos-López
4,
Afolabi Agbona
5,6 and
José Crossa
7,8,9,10,*
1
Facultad de Telemática, Universidad de Colima, Colima 28040, Mexico
2
Centro Universitario de Ciencias Exactas e Ingenierías (CUCEI), Universidad de Guadalajara, Guadalajara 44430, Mexico
3
Facultad de Ciencias de la Educación, Universidad de Colima, Colima 28040, Mexico
4
Department of Public Health Sciences, University of California Davis, Davis, CA 95616, USA
5
International Institute of Tropical Agriculture (IITA), Ibadan 200113, Nigeria
6
Molecular & Environmental Plant Sciences, Texas A&M University, College Station, TX 77843, USA
7
International Maize and Wheat Improvement Center (CIMMYT), Km 45, Carretera Mexico-Veracruz, Texcoco 52640, Mexico
8
Louisiana State University, Baton Rouge, LA 70803, USA
9
Distinguished Scientist Fellowship Program and Department of Statistics and Operations Research, King Saud University, Riyah 11451, Saudi Arabia
10
Colegio de Postgraduados, Montecillos 56230, Mexico
*
Authors to whom correspondence should be addressed.
Genes 2024, 15(8), 969; https://doi.org/10.3390/genes15080969
Submission received: 29 June 2024 / Revised: 17 July 2024 / Accepted: 22 July 2024 / Published: 23 July 2024
(This article belongs to the Section Plant Genetics and Genomics)

Abstract

:
Genomic selection (GS) is changing plant breeding by significantly reducing the resources needed for phenotyping. However, its accuracy can be compromised by mismatches between training and testing sets, which impact efficiency when the predictive model does not adequately reflect the genetic and environmental conditions of the target population. To address this challenge, this study introduces a straightforward method using binary-Lasso regression to estimate β coefficients. In this approach, the response variable assigns 1 to testing set inputs and 0 to training set inputs. Subsequently, Lasso, Ridge, and Elastic Net regression models use the inverse of these β coefficients (in absolute values) as weights during training (WLasso, WRidge, and WElastic Net). This weighting method gives less importance to features that discriminate more between training and testing sets. The effectiveness of this method is evaluated across six datasets, demonstrating consistent improvements in terms of the normalized root mean square error. Importantly, the model’s implementation is facilitated using the glmnet library, which supports straightforward integration for weighting β coefficients.

1. Introduction

Genomic selection (GS) is a transformative methodology in plant breeding, offering substantial reductions in phenotyping costs. GS achieves this by using a statistical machine learning algorithm trained on a reference population that includes both phenotypic and genotypic data. The trained model then predicts phenotypic values for a target population based solely on their genotypic data [1]. Consequently, GS is widely adopted in breeding programs, enabling the early selection of candidate lines without the necessity of phenotyping.
GS has been successfully implemented in various crop breeding programs, leading to significant advancements [2]. Notable examples include maize, where GS has enhanced yield, disease resistance, and drought tolerance. In wheat, GS has improved grain yield, disease resistance, and quality traits [3,4]. For rice, GS has increased yield, pest and disease resistance, and environmental adaptability [5,6]. In soybean, GS has boosted yield, disease resistance, and oil/protein content [7,8]. Barley has seen enhanced yield, malting quality, and disease resistance due to GS [9,10]. In cassava, GS has improved yield, disease resistance, and nutritional content [11,12]. Potato breeding has benefited from increased yield, disease resistance, and tuber quality through GS [13,14]. Lastly, GS has improved yield, disease resistance, and fruit quality in tomato breeding. These examples underscore the effectiveness of GS in enhancing key traits across a diverse range of crops.
According to Montesinos-López et al. [1], the successful implementation of GS in crop breeding depends on several critical factors. High-quality and extensive genotypic and phenotypic data are essential for training accurate predictive models. A large, diverse reference population that accurately represents the genetic diversity of the target population enhances model robustness. The choice of advanced statistical models and machine learning algorithms, along with adequate computational resources, also significantly impacts the predictive power of GS. Traits with higher heritability and those influenced by fewer genes are typically predicted more accurately. Environmental factors and genotype-by-environment interactions must be considered to improve prediction accuracy for traits affected by external conditions. Clear breeding objectives, well-designed programs, and cost considerations are crucial for the effective integration of GS. Additionally, the presence of skilled personnel with expertise in quantitative genetics, statistics, bioinformatics, and computational biology is vital. Collaborative efforts among breeding programs, research institutions, and industry stakeholders can provide the necessary resources, data, and expertise, further enhancing the successful application of GS in crop breeding.
A significant mismatch between the reference (training) population and the target (testing) population can significantly undermine the accuracy of GS predictions [15]. This issue arises when the genetic composition, allele frequencies, and linkage disequilibrium patterns differ between the two populations. Such discrepancies can lead to incorrect marker–trait associations and biased predictions. Additionally, genotype-by-environment interactions may not be adequately captured if the populations experience different environmental conditions. Variability in phenotypic expression and genetic backgrounds further complicates prediction accuracy. To mitigate these problems, it is essential to ensure that the reference population is representative of the target population, incorporating diverse genetic backgrounds and environmental conditions. This approach enhances the reliability of GS predictions by better capturing the genetic and environmental complexities present in the target population.
In this context, a method using penalized Lasso, Ridge, and Elastic Net regression [16] is proposed to reduce the mismatch between training and testing sets. This proposed method is evaluated in eight real datasets, each containing multiple environments. The model is implemented in a uni-environment fashion, with each environment containing several families. Given that the genetic material of each family differs, the training set in each dataset consists of all families minus one, and the remaining family is used as the testing set. We compare the proposed method where the coefficient of regression is weighted by the absolute value of the regression coefficient with conventional penalized methods, assessing prediction performance in terms of normalized root mean square errors.

2. Material and Methods

2.1. Datasets

Table 1 has details of the 8 datasets included in this study. The data used were obtained from Montesinos-Lopez et al. (2023) [17] and comprise two crops, Maize and Soybean, which have different numbers of DNA sequences and different numbers of environments (location–year combination) and numbers of traits.

2.2. Statistical Model

In a general context, we have a covariate vector X = X 1 ,   ,   X p T , such as molecular markers, and we want to use this information to predict or explain how this variable affects a real-value response Y (of size n), for example, grain yield or any other quantitative trait the breeders want to predict. The linear multiple regression model assumes a relationship given by
Y = β 0 + j = 1 p X j β j + ϵ
where ϵ is a random error vector with mean 0, E ϵ = 0 and is independent of X . This error is included in the model to capture measurement errors and the effects of other unregistered explanatory variables that can help explain the mean response. Then, the conditional mean of this model is E Y | X = β 0 + j = 1 p X j β j , and the conditional distribution of Y given X is only affected by the information of X .
To estimate the parameters β 0 and β = ( β 1 ,   ,   β p ) T , we usually have a set of data x i T , y i , i = 1 ,   ,   n , often known as training data, where x i = x i 0 ,   x i 1 ,   ,   x i p T is a vector of feature measurement and y i is the response variable measurement corresponding to the i -th individual drawn. In the context of large p and small n , the ordinary least squares method that does not penalized the parameters is used, and in this context, the estimates are not optimal. One common method to estimate β and β 0 when p >> n is the Elastic regression (ENet) method (1), which consists of taking the ( β 0 , β ) value that minimizes the penalized residual sum of squares [1,16], defined as
P R S S λ β = i = 1 n y i β 0 j = 1 p x i j β j 2 + λ ( 1 α ) j = 1 p β j 2 + α j = 1 p β j  
where λ 0 is the regularization parameter, which determines the level or degree to which the β coefficients are shrunk towards zero, α ( 0,1 ) . When λ = 0 and α   = 0, the ordinal least square (OLS) is the solution to the β coefficients, but when λ is large, P R S S λ β is dominated by the penalization term, and the OLS solution must shrink towards 0 [18]. In general, when the number of parameters (p) to be estimated is larger than the number of observations (n), the estimator can be highly variable. In this situation, the intuition of penalized regression tries to alleviate this by constraining the β coefficients [19]. Also, it is important that when α   = 0, the loss function given in (1) is reduced to the loss function of Ridge regression (Ridge; [1,16]) while when α   = 1, the loss function corresponds to the Lasso regression [16].

2.3. Proposed Approach

The proposed method is described below in 4 steps.
  • Step 1. We assume that we have a uni-environment or single-environment dataset in which there are at least two families, and for each family, there are some lines. Since our goal is to predict a complete family, the information of this family constitutes the testing set { X _ t s t , y _ t s t }, and the information of the remaining families is the training set { X _ t r n , y _ t r n }. Then, in the original dataset { X _ t r n , y _ t r n , X _ t s t , y _ t s t }, we remove the original response variable column { y _ t r n , y _ t s t } and add a fictitious (new) response variable column { y f _ t r n ,   y f _ t s t } that replaces the source of the data with zero (that is, y f _ t r n = 0) for samples (observations) on the training set and with one (i.e., y f _ t s t = 1) for the samples in the testing set [17]. In other words, the fictitious response variables with 0 s correspond to the remaining families, and those with 1 s correspond to the family we want to predict. Note that since the response variable is a fictitious response variable, with a training label ( y f _ t r n = 0) and a testing label ( y f _ t s t = 1), not the original response variable, we should avoid any complication in defining these two populations when performing the random cross-validation and also avoid any type of data leakage confusion.
  • Step 2. We then implement the Lasso regression with inputs { X _ t r n , y f _ t r n , X _ t s t , y f _ t s t } and we extract the β coefficients ( β b i n = ( β 1 , b i n ,   ,   β p , b i n ) T ) for each marker.
  • Step 3. The next step is to compute weights as w j = 1/abs( β j ,   b i n ), where j = 1 ,   ,   p ,   p denotes the number of markers.
  • Step 4. Then, the weights computed in step 3 are used to train a weighted Lasso and Ridge regression. Regarding the Weighted Lasso (WLasso), the loss function used to train the models is given as follows:
P R S S λ β w L = i = 1 n y i β 0 j = 1 p x i j β j 2 + λ j = 1 p w j β j
Meanwhile, for the weighted Ridge regression (WRidge), the weighted loss function is equal to
P R S S λ β R = i = 1 n y i β 0 j = 1 p x i j β j 2 + λ j = 1 p w j β j 2
and for the weighted Elastic net regression (WENet), the weighted loss function is equal to
P R S S λ β R = i = 1 n y i β 0 j = 1 p x i j β j 2 + λ 1 α j = 1 p w j β j 2 + α j = 1 p w j β j
To implement these loss functions, we used the glmnet library in [16], which is implemented in R statistical software (v. 2024) [20].

2.4. How to Avoid Data Leakage

Avoiding data leakage in genomic prediction studies is crucial to ensure the validity and generalizability of your model’s performance. Data leakage occurs when information from outside the training dataset is used to create the model, leading to overly optimistic performance estimates. This can be avoided by properly separating training sets from testing sets; this means that no data points used for training the model should be included in the testing set, and vice versa.
Furthermore, we use cross-validation techniques where the dataset is split into k subsets, and the model is trained on k − 1 subsets and tested on the remaining subset. This helps in validating the model’s performance across different subsets of the data, minimizing the risk of data leakage. In the prediction of longitudinal data, it is important to ensure that the training data precede the testing data chronologically. This simulates real-world scenarios where predictions are made on future, unseen data.
We must ensure that there are no duplicates or near-duplicates between the training and testing sets. In genomic studies, this means carefully managing how genotypes and phenotypes are split, ensuring no overlap. Feature selection should be performed within the cross-validation loop. Performing feature selection before splitting the data can lead to information from the test set leaking into the model, thereby inflating performance metrics. We should use resampling techniques like bootstrapping within the training data only. We must not mix data points from the test set during resampling, as this can cause leakage.

2.5. Evaluation of Prediction Accuracy

Our evaluation procedure preserves data leakage by employing a cross-validation approach wherein each iteration excludes one family at a time from the training set, and the entire proposed approach described above is applied. Therefore, data from the testing set do not influence the weightings derived from the training set.
Specifically, in each iteration, data from a single family served as the testing set, while data from the remaining families constituted the training set [1]. The number of iterations corresponded to the total number of families, ensuring that each family was used as the testing set exactly once. This method was implemented to evaluate the model’s capability to predict information for an entire family based on data from other diverse families. To compare the predictive performance of the proposed method (WRidge, WLasso, and WENet) with the conventional glmnet method (Ridge, Lasso and ENet), the normalized root mean square error (NRMSE) was employed as the evaluation metric.
The NRMSE, a measure of prediction error, was calculated by first computing the square root of the mean square error (MSE) and then dividing this value by the average of the observed values in the testing set. The NRMSE is a metric used to evaluate the accuracy of a predictive model, particularly useful when comparing models with different scales. It is a normalized version of the root mean square error (RMSE), which adjusts for the scale of the data, making it easier to interpret and compare across different datasets. NRMSE = 0: perfect predictions. NRMSE > 0: the further the value from 0, the worse the model’s predictions. NRMSE < 1: indicates a good fit, depending on the context and domain. Some advantages of NRMSE are that it is scale-invariant as it adjusts to the scale of the data, making it easier to compare across different datasets, and has an intuitive interpretation because it provides a relative measure of the prediction error in relation to the range of the observed values. Using NRMSE helps provide a more standardized assessment of model performance, making it particularly useful in genomic prediction studies where traits can have varying scales.

3. Results

The results are provided in five sections that contain the results of the datasets Maize_1, Maize_3, Soybean_1, and Soybean_2 and across all of these datasets, presented in Figure 1, Figure 2, Figure 3, Figure 4 and Figure 5 and Table 2 and Table 3. The results for the rest of the datasets (Maize_2, Maize_4, Soybean_3, Soybean_4) are given in Appendix A containing Table A1 and Figure A1, Figure A2, Figure A3 and Figure A4.

3.1. Maize_1

Figure 1 shows the prediction performance in terms of the normalized root mean square error (NRMSE) for Maize Dataset 1. The first three original prediction models (ENet, Lasso, Ridge) were compared with their weighted versions (WENet, WLasso, WRidge) using two traits for these predictions: GDD_ASI and GDD_DTT.
Figure 1. Normalized root mean square error (NRMSE) for the “Maize_1” dataset. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge) and the 2 traits (GDD_ASI and GDD_DTT).
Figure 1. Normalized root mean square error (NRMSE) for the “Maize_1” dataset. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge) and the 2 traits (GDD_ASI and GDD_DTT).
Genes 15 00969 g001
In the first comparison, we can observe that, in terms of NRMSE for the GDD_ASI trait, the ENet model, with an NRMSE value of 1.427, was inferior to the weighted WENet model, with an NRMSE value of 1.307, being superior in prediction by 9.185%. Similarly, for the GDD_DTT trait, the WENet model, with an NRMSE value of 1.307, also outperformed the Enet, with an NRMSE value of 1.429, with a difference of 9.358%. The individual values can be observed in more detail in Table 2.
In the second comparison, we can observe that, in terms of NRMSE for the GDD_ASI trait, the Lasso model, with an NRMSE value of 1.433, was inferior to the weighted WLasso model, with an NRMSE value of 1.332, surpassing it in prediction by 7.567%. Similarly, for the GDD_DTT trait, the WLasso model, with an NRMSE value of 1.332, also outperformed the Lasso, with an NRMSE value of 1.433, with a difference of 7.564%. The individual values can be observed in more detail in Table 2.
In the third comparison, we can observe that, in terms of NRMSE for the GDD_ASI trait, the Ridge model, with an NRMSE value of 1.371, was inferior to the weighted WRidge model, with an NRMSE value of 1.303, being superior in prediction by 5.171%. Similarly, for the GDD_DTT trait, the WRidge model, with an NRMSE value of 1.303, also outperformed the Ridge, with an NRMSE value of 1.371, with a difference of 5.180%. The individual values can be observed in more detail in Table 2.

3.2. Maize_3

In the first comparison (Figure 2), we can observe that, in terms of NRMSE for the GDD_ASI trait, the ENet model, with an NRMSE value of 1.405, was inferior to the WENet model, with an NRMSE value of 1.369, which outperformed it in prediction by 2.647%. Similarly, for the GDD_DTT trait, the WENet model, with an NRMSE value of 1.369, was also superior to Enet, with an NRMSE value of 1.404, with an advantage of 2.549%. The individual values can be observed in more detail in Table 2.
Figure 2. Normalized root mean square error (NRMSE) for the “Maize 3” dataset. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge) and the 2 traits (GDD_ASI and GDD_DTT).
Figure 2. Normalized root mean square error (NRMSE) for the “Maize 3” dataset. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge) and the 2 traits (GDD_ASI and GDD_DTT).
Genes 15 00969 g002
In the second comparison (Figure 2), we can observe that, in terms of NRMSE for the GDD_ASI trait, the Lasso model, with an NRMSE value of 1.409, was inferior to the WLasso model, with an NRMSE value of 1.370, which outperformed it in prediction by 2.866%. Similarly, for the GDD_DTT trait, the WLasso model, with an NRMSE value of 1.370, was also superior to Lasso, with an NRMSE value of 1.407, with an advantage of 2.721% (see Table 2).
For the third comparison displayed in Figure 2, note that regarding the NRMSE for the GDD_ASI trait, the Ridge model, with an NRMSE value of 1.380, was inferior to the WRidge model, with an NRMSE value of 1.331, which outperformed it in prediction by 3.656%. Similarly, for the GDD_DTT trait, the WRidge model, with an NRMSE value of 1.333, was also superior to Ridge, with an NRMSE value of 1.381, with an advantage of 3.598%. The individual values can be observed in more detail in Table 2.

3.3. Soybean_1

In the first comparison (Figure 3), it is evident that, in terms of NRMSE for the R8 trait, the ENet model, with an NRMSE value of 1.241, outperformed the weighted WENet model, with an NRMSE value of 1.274. The weighted model, in this instance, fell short in prediction by 2.621%. Similarly, for the Height trait, the WENet model, with an NRMSE value of 1.239, was inferior to Enet, with an NRMSE value of 1.193, exhibiting a shortfall of 3.729%. Detailed individual values can be observed in Table 2.
Figure 3. Normalized root mean square error (NRMSE) for the “Soybean 1” dataset. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge) and the 2 traits (Height and R8).
Figure 3. Normalized root mean square error (NRMSE) for the “Soybean 1” dataset. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge) and the 2 traits (Height and R8).
Genes 15 00969 g003
In Figure 3, it is noticeable that, in terms of NRMSE for the R8 trait, the Lasso model, with an NRMSE value of 1.254, surpassed the weighted WLasso model, with an NRMSE value of 1.255, by 0.04%. Likewise, the weighted WLasso model, with an NRMSE value of 1.227, fell short in prediction for the Height trait, being inferior to the Lasso model, with an NRMSE value of 1.195, by 2.675%. Detailed individual values can be observed in Table 2.
Figure 3 displays that in terms of NRMSE for the R8 trait, the Ridge model, with an NRMSE value of 1.270, outperformed the weighted WRidge model, with an NRMSE value of 1.284. The weighted model, in this instance, fell short in prediction by 1.112%. Similarly, for the Height trait, the WRidge model, with an NRMSE value of 1.191, was inferior to Ridge, with an NRMSE value of 1.143, exhibiting a shortfall of 4.018%. Detailed individual values can be observed in Table 2.

3.4. Soybean_2

In the first comparison (Figure 4), it is evident that, in terms of NRMSE for the R8 trait, the ENet model, with an NRMSE value of 1.554, was inferior to the weighted WENet model, with an NRMSE value of 1.470, which outperformed it in prediction by 5.711%. Similarly, for the Height trait, the WENet model, with an NRMSE value of 1.316, also surpassed Enet, with an NRMSE value of 1.361, exhibiting superiority of 3.427%. Detailed individual values can be observed in Table 2.
Figure 4. Normalized root mean square error (NRMSE) for the “Soybean 2” dataset. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge) and the 2 traits (Height and R8).
Figure 4. Normalized root mean square error (NRMSE) for the “Soybean 2” dataset. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge) and the 2 traits (Height and R8).
Genes 15 00969 g004
In Figure 4, it is noticeable that, in terms of NRMSE for the R8 trait, the Lasso model, with an NRMSE value of 1.559, was inferior to the weighted WLasso model, with an NRMSE value of 1.473, which outperformed it in prediction by 5.828%. Similarly, for the Height trait, the WLasso model, with an NRMSE value of 1.313, also surpassed Lasso, with an NRMSE value of 1.357, exhibiting superiority of 3.317%. Detailed individual values can be observed in Table 2.
Figure 4 displays that regarding NRMSE for the R8 trait, the Ridge model, with an NRMSE value of 1.545, was inferior to the weighted WRidge model, with an NRMSE value of 1.515, which outperformed it in prediction by 2.013%. Similarly, for the Height trait, the WRidge model, with an NRMSE value of 1.332, also surpassed Ridge, with an NRMSE value of 1.342, exhibiting superiority of 0.682%. Detailed individual values can be observed in Table 2.

3.5. Across_Data

In the first comparison (Figure 5), it is evident that, in terms of NRMSE, the ENet model, with an NRMSE value of 1.409, was inferior to the weighted WENet model, with an NRMSE value of 1.378, which outperformed it in prediction by 2.196%. Detailed individual values can be observed in Table 3. Figure 5 shows that the Lasso model, with an NRMSE value of 1.413, was worse regarding the weighted WLasso model, with an NRMSE value of 1.377, which outperformed it in prediction by 2.630%. Detailed individual values can be observed in Table 3. On the other hand, the Ridge model, with an NRMSE value of 1.387, was superior to the weighted WRidge model, with an NRMSE value of 1.369, which outperformed it in prediction by 1.309%.
Figure 5. Normalized root mean square error (NRMSE) for all datasets. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge).
Figure 5. Normalized root mean square error (NRMSE) for all datasets. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge).
Genes 15 00969 g005

4. Discussion

There is no doubt that climate change and a growing global population are increasing the pressure to boost productivity worldwide [21]. Therefore, innovative methods like GS are very appealing, as they have the potential to significantly enhance crop productivity on a global scale. For this reason, the adoption of GS is being accelerated in many public and private breeding programs due to its tremendous potential. However, like many new technologies, the practical and successful implementation of GS depends on several factors that impact its accuracy. Consequently, research is being conducted to mitigate these factors affecting its performance. In this study, we propose a method to increase the efficiency of GS in contexts where there is a limited relationship between the training and testing sets.
The core idea of the proposed method is to assign greater weight to features (markers) that less distinctly differentiate between the training and testing sets and, conversely, to assign less weight to those that do. To achieve this, the method first trains a binary classifier (Lasso regression) with a fictitious response variable that assigns a value of 1 to testing set inputs and 0 to training set inputs. From this first step, the β coefficients are obtained and used to compute weights for the second step, which uses the response variable of interest. These weights are calculated as the inverse of the absolute values of the β coefficients from the first stage.
The proposed method demonstrated superior performance across various datasets. Specifically, WENet outperformed Enet by 2.196% in terms of NRMSE. Similarly, WLasso exceeded Lasso by 2.63% in terms of NRMSE. Lastly, WRidge surpassed Ridge with a 1.309% improvement in NRMSE. We acknowledge that the margin of improvement is modest; however, the proposed method provides empirical evidence of its ability to enhance prediction accuracy in contexts with significant mismatches between training and testing sets. Therefore, we encourage further exploration and refinement of this method to boost its predictive power. Since in each environment of all datasets, we predict a complete family using the remaining families, we cannot expect good prediction performance when the family to be predicted is genetically distant from those in the training set. Additionally, since families are often genetically distinct from each other, achieving good prediction performance for new families is challenging. Furthermore, predicting a complete family is complicated because markers do not capture all the necessary information for optimal predictions.
Finally, we emphasize that GS is revolutionizing plant and animal breeding by enabling the selection of individuals without the need to measure their phenotypic values. However, its successful implementation faces numerous challenges that must be addressed. This paper underscores the critical importance of understanding that when training and testing materials have significant genetic distance, the predictive quality of GS methodologies deteriorates significantly. Therefore, investigating statistical approaches to mitigate this issue is crucial for enhancing the robustness of GS methodologies.

4.1. Why Weighting Markers Might Be Useful for Enhancing Genomic Prediction

The usefulness of weighting markers in genomic prediction can vary depending on the characteristics of the datasets and the traits being studied. There are several reasons why weighting markers might be beneficial for some datasets and traits, but not for others: (1) For traits controlled by many genes with small effects, weighting markers can help identify and prioritize the most influential markers, enhancing prediction accuracy, whereas traits controlled by a few genes with large effects benefit significantly from weighting these key markers, as they have a disproportionate impact on the trait. However, when a trait is controlled by a very large number of genes with very small effects, the benefit of weighting individual markers diminishes as the effects are too dispersed. (2) When marker density is high, weighting can help manage the large volume of data by emphasizing markers with greater predictive power. (3) Weighting markers in regions with high LD can reduce redundancy and focus on the most informative markers; in populations with clear substructures or relatedness, weighting can help correct for population stratification and relatedness, leading to more accurate predictions.
Furthermore, for traits with significant genotype-by-environment interactions, weighting markers that show consistent effects across environments can improve prediction robustness, and for traits with high heritability, certain markers will have strong associations with the trait, making weighting beneficial to capture this strong genetic signal. When a trait is controlled by a very large number of genes with very small effects, the benefit of weighting individual markers diminishes as the effects are too dispersed.
In summary, weighting markers in genomic prediction can enhance model performance by emphasizing the most informative genetic markers, correcting for population structure, and managing high-dimensional data. However, its effectiveness depends on the genetic architecture of the trait, marker density, population characteristics, and environmental stability. Understanding these factors can guide the decision on whether to apply marker weighting in genomic prediction studies.

4.2. Specific Considerations for Genomic Prediction

In the context of genomic prediction, predicting traits based on genetic markers is complex and often challenging. For example, an NRMSE of 0.015 (see Table 2 and Table A1) is particularly impressive. It indicates that the model predictions are very close to the true values, with only a 1.5% error, that is, with a very small deviation from its target. The larger and more complex the dataset, the more impressive a low NRMSE becomes. Highly variable traits typically result in higher errors, so a low NRMSE here is significant. Simpler models achieving low NRMSEs suggest a robust underlying pattern, whereas complex models might fit noise if overfitting. For example, in a genomic prediction study for plant breeding where breeders are predicting crop yield or resistance to diseases, an NRMSE of around 0.015 means the model’s predictions are very accurate compared to the variability in the observed data.

5. Conclusions

The lack of strong relatedness between training and testing sets is a common issue in the practical implementation of GS, significantly affecting its accuracy. To address this problem, we propose a method aimed at improving accuracy under these conditions. Our findings indicate that the proposed method enhances accuracy when there is a significant mismatch between training and testing sets. However, while the method provides improvement, it does not completely resolve the issue. In addition to addressing this issue, we emphasize the significance of the problem within the GS methodology, as it often compromises its efficiency. We encourage other scientists to enhance the proposed framework to further increase the robustness of the GS methodology.

Author Contributions

Conceptualization, O.A.M.-L. and A.M.-L.; methodology, O.A.M.-L. and A.M.-L.; formal analysis, O.A.M.-L. and A.M.-L.; writing—original draft preparation, O.A.M.-L., A.M.-L. and J.C.; writing—review and editing, O.A.M.-L., C.D.P.-C., A.M.-L., J.A.L.T., J.C.M.-L., A.A. and J.C.; project administration, all authors. All authors have read and agreed to the published version of the manuscript.

Funding

We are thankful for the financial support provided by the Bill & Melinda Gates Foundation [INV-003439, BMGF/FCDO, Accelerating Genetic Gains in Maize and Wheat for Improved Livelihoods (AG2MW)].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original data presented in the study are openly available at https://github.com/osval78/Mistmach_Penalized_Regression (accessed on 21 July 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Prediction performance in terms of normalized root mean square error (NRMSE) for datasets Maize_2, Maize_4, Sobybean_3, and Sobybean_4. ENet denotes Elastic net regression, Lasso denotes Lasso regression, and Ridge denotes Ridge regression, while WENet, WLasso, and WRidge denote the weighted counterparts of ENet, Lasso, and Ridge. NRMSE_SD denotes the standard deviation of the NRMSE.
Table A1. Prediction performance in terms of normalized root mean square error (NRMSE) for datasets Maize_2, Maize_4, Sobybean_3, and Sobybean_4. ENet denotes Elastic net regression, Lasso denotes Lasso regression, and Ridge denotes Ridge regression, while WENet, WLasso, and WRidge denote the weighted counterparts of ENet, Lasso, and Ridge. NRMSE_SD denotes the standard deviation of the NRMSE.
DatasetModelTraitNRMSENRMSE_SD
Maize_2ENetGDD_ASI1.6690.012
Maize_2ENetGDD_DTT1.6690.012
Maize_2LassoGDD_ASI1.6740.012
Maize_2LassoGDD_DTT1.6740.012
Maize_2RidgeGDD_ASI1.6470.013
Maize_2RidgeGDD_DTT1.6470.013
Maize_2WENetGDD_ASI1.6620.013
Maize_2WENetGDD_DTT1.6620.013
Maize_2WLassoGDD_ASI1.6790.013
Maize_2WLassoGDD_DTT1.6790.013
Maize_2WRidgeGDD_ASI1.5710.012
Maize_2WRidgeGDD_DTT1.5710.012
Maize_4ENetGDD_ASI1.5440.011
Maize_4ENetGDD_DTT1.5460.011
Maize_4LassoGDD_ASI1.5500.011
Maize_4LassoGDD_DTT1.5470.011
Maize_4RidgeGDD_ASI1.5440.012
Maize_4RidgeGDD_DTT1.5460.012
Maize_4WENetGDD_ASI1.5010.010
Maize_4WENetGDD_DTT1.5020.010
Maize_4WLassoGDD_ASI1.5130.011
Maize_4WLassoGDD_DTT1.5130.011
Maize_4WRidgeGDD_ASI1.4650.011
Maize_4WRidgeGDD_DTT1.4640.011
Soybean_3ENetR81.2500.028
Soybean_3ENetHeight1.1040.009
Soybean_3LassoR81.2760.030
Soybean_3LassoHeight1.0970.009
Soybean_3RidgeR81.2640.034
Soybean_3RidgeHeight1.0930.009
Soybean_3WENetR81.2780.033
Soybean_3WENetHeight1.0970.015
Soybean_3WLassoR81.2640.032
Soybean_3WLassoHeight1.0690.009
Soybean_3WRidgeR81.2920.037
Soybean_3WRidgeHeight1.0850.011
Soybean_4ENetR81.0900.010
Soybean_4ENetHeight1.1760.019
Soybean_4LassoR81.1460.010
Soybean_4LassoHeight1.1800.020
Soybean_4RidgeR81.0660.008
Soybean_4RidgeHeight1.1210.018
Soybean_4WENetR81.1270.010
Soybean_4WENetHeight1.1560.013
Soybean_4WLassoR81.1160.010
Soybean_4WLassoHeight1.1320.013
Soybean_4WRidgeR81.0690.008
Soybean_4WRidgeHeight1.1290.015
Figure A1. Dataset Maize 2. Normalized root mean square error (NRMSE) for the Maize_2 dataset. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge) and the 2 traits (GDD_ASI and GDD_DTT).
Figure A1. Dataset Maize 2. Normalized root mean square error (NRMSE) for the Maize_2 dataset. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge) and the 2 traits (GDD_ASI and GDD_DTT).
Genes 15 00969 g0a1
Figure A2. Dataset Maize 4. Normalized root mean square error (NRMSE) for the Maize_4 dataset. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge) and the 2 traits (GDD_ASI and GDD_DTT).
Figure A2. Dataset Maize 4. Normalized root mean square error (NRMSE) for the Maize_4 dataset. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge) and the 2 traits (GDD_ASI and GDD_DTT).
Genes 15 00969 g0a2
Figure A3. Dataset Soybean 3. Normalized root mean square error (NRMSE) for the Soybean_3 dataset. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge) and the 2 traits (Height and R8).
Figure A3. Dataset Soybean 3. Normalized root mean square error (NRMSE) for the Soybean_3 dataset. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge) and the 2 traits (Height and R8).
Genes 15 00969 g0a3
Figure A4. Dataset Soybean 4. Normalized root mean square error (NRMSE) for the Soybean_4 dataset. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge) and the 2 traits (Height and R8).
Figure A4. Dataset Soybean 4. Normalized root mean square error (NRMSE) for the Soybean_4 dataset. A comparison is presented for the 6 evaluated models (Enet, Lasso, Ridge, WEnet, WLasso, and WRidge) and the 2 traits (Height and R8).
Genes 15 00969 g0a4

References

  1. Montesinos López, O.A.; Montesinos-López, A.; Crossa, J. Multivariate statistical machine learning methods for genomic prediction. In Multivariate Statistical Machine Learning Methods for Genomic Prediction; Springer: Cham, Switzerland, 2022. [Google Scholar]
  2. Heffner, E.L.; Sorrells, M.E.; Jannink, J.L. Genomic selection for crop improvement. Crop Sci. 2009, 49, 1–12. [Google Scholar] [CrossRef]
  3. Montesinos-López, A.; Montesinos-López, O.A.; Crossa, J.; Toledo, F.H.; Pérez-Hernández, O.; Eskridge, K.M.; Rutkoski, J. A Genomic Bayesian Multi-trait and Multi-environment Model. G3 Genes Genomes Genet. 2016, 6, 2725–2744. [Google Scholar] [CrossRef] [PubMed]
  4. Crossa, J.; Beyene, Y.; Kassa, S.; Pérez, P.; Hickey, J.M.; Chen, C.; de los Campos, G.; Burgueño, J.; Windhausen, V.S.; Buckler, E.; et al. Genomic prediction in maize breeding populations with genotyping-by-sequencing. G3 Genes Genomes Genet. 2013, 3, 1903–1926. [Google Scholar] [CrossRef] [PubMed]
  5. Spindel, J.; Begum, H.; Akdemir, D.; Virk, P.; Collard, B.; Redoña, E.; Atlin, G.; Jannink, J.-L.; McCouch, S.R. Genomic selection and association mapping in rice (Oryza sativa): Effect of trait genetic architecture, training population composition, marker number and statistical model on accuracy of rice genomic selection in elite, tropical rice breeding lines. PLoS Genet. 2015, 11, e1004982. [Google Scholar]
  6. Bartholomé, J.; Prakash, P.T.; Cobb, J.N. Genomic prediction: Progress and perspectives for rice improvement. In Genomic Prediction of Complex Traits: Methods and Protocols; Humana: New York, NY, USA, 2022; pp. 569–617. [Google Scholar]
  7. Zhang, J.; Song, Q.; Cregan, P.B.; Jiang, G.L. Genome-wide association study, genomic prediction and marker-assisted selection for seed weight in soybean (Glycine max). Theor. Appl. Genet. 2016, 129, 117–130. [Google Scholar] [CrossRef] [PubMed]
  8. Jean, M.; Cober, E.; O’Donoughue, L.; Rajcan, I.; Belzile, F. Improvement of key agronomical traits in soybean through genomic prediction of superior crosses. Crop Sci. 2021, 61, 3908–3918. [Google Scholar] [CrossRef]
  9. Nielsen, N.H.; Jahoor, A.; Jensen, J.D.; Orabi, J.; Cericola, F.; Edriss, V.; Jensen, J. Genomic prediction of seed quality traits using advanced barley breeding lines. PLoS ONE 2016, 11, e0164494. [Google Scholar] [CrossRef] [PubMed]
  10. Lorenz, A.J.; Smith, K.P. Adding genetically distant individuals to training populations reduces genomic prediction accuracy in barley. Crop Sci. 2015, 55, 2657–2667. [Google Scholar] [CrossRef]
  11. Yonis, B.O.; Pino del Carpio, D.; Wolfe, M.; Jannink, J.L.; Kulakow, P.; Rabbi, I. Improving root characterisation for genomic prediction in cassava. Sci. Rep. 2020, 10, 8003. [Google Scholar] [CrossRef] [PubMed]
  12. Long, E.M.; Romay, M.C.; Ramstein, G.; Buckler, E.S.; Robbins, K.R. Utilizing evolutionary conservation to detect deleterious mutations and improve genomic prediction in cassava. Front. Plant Sci. 2023, 13, 1041925. [Google Scholar] [CrossRef] [PubMed]
  13. Adams, J.; de Vries, M.; van Eeuwijk, F. Efficient genomic prediction of yield and dry matter in hybrid potato. Plants 2023, 12, 2617. [Google Scholar] [CrossRef] [PubMed]
  14. Habyarimana, E.; Parisi, B.; Mandolino, G. Genomic prediction for yields, processing and nutritional quality traits in cultivated potato (Solanum tuberosum L.). Plant Breed. 2017, 136, 245–252. [Google Scholar] [CrossRef]
  15. de Los Campos, G.; Hickey, J.M.; Pong-Wong, R.; Daetwyler, H.D.; Calus, M.P. Whole-genome regression and prediction methods applied to plant and animal breeding. Genetics 2013, 193, 327–345. [Google Scholar] [CrossRef] [PubMed]
  16. Friedman, J.; Hastie, T.; Tibshirani, R. Regularization Paths for Generalized Linear Models via Coordinate Descent. J. Stat. Softw. 2010, 33, 1–22. [Google Scholar] [CrossRef] [PubMed]
  17. Montesinos-López, O.A.; Kismiantini Montesinos-López, A. Designing optimal training sets for genomic prediction using adversarial validation with probit regression. Plant Breed. 2023, 142, 594–606. [Google Scholar] [CrossRef]
  18. Christensen, P. Plane Answers to Complex Questions: The Theory of Linear Models; Springer Science+Business Media: New York, NY, USA, 2011. [Google Scholar]
  19. Wakefield, J. Bayesian and Frequentist Regression Methods; Springer Science+Business Media: New York, NY, USA, 2013. [Google Scholar]
  20. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2024; Available online: https://www.R-project.org/ (accessed on 21 July 2024).
  21. Hall, C.; Dawson, T.P.; Macdiarmid, J.I.; Matthews, R.B.; Smith, P. The impact of population growth and climate change on food security in Africa: Looking ahead to 2050. Int. J. Agric. Sustain. 2017, 15, 124–135. [Google Scholar] [CrossRef]
Table 1. Description of the datasets used to implement the proposed method.
Table 1. Description of the datasets used to implement the proposed method.
Dataset NumberDataset
(Crop and Number)
Cultivars
(Number)
Markers
(Number)
Environments
(Number)
Traits
(Number)
Families
(Number)
Dataset 1Maize_1100040851146
Dataset 2Maize_2100040851146
Dataset 3Maize_3100040851146
Dataset 4Maize_499940851146
Dataset 5Soybean_1104418108610
Dataset 6Soybean_269118088610
Dataset 7Soybean_37018098610
Dataset 8Soybean_45918038610
Data availability: The data are available at: https://github.com/osval78/Mistmach_Penalized_Regression (accessed on 21 July 2024).
Table 2. Prediction performance in terms of normalized root mean square error (NRMSE) for datasets Maize_1, Maize_3, Sobybean_1, and Sobybean_2. ENet denotes Elastic net regression, Lasso denotes Lasso regression, and Ridge denotes Ridge regression, while WENet, WLasso, and WRidge denote the weighted counterparts of ENet, Lasso, and Ridge. NRMSE_SD denotes the standard deviation of the NRMSE.
Table 2. Prediction performance in terms of normalized root mean square error (NRMSE) for datasets Maize_1, Maize_3, Sobybean_1, and Sobybean_2. ENet denotes Elastic net regression, Lasso denotes Lasso regression, and Ridge denotes Ridge regression, while WENet, WLasso, and WRidge denote the weighted counterparts of ENet, Lasso, and Ridge. NRMSE_SD denotes the standard deviation of the NRMSE.
DatasetModelTraitNRMSENRMSE_SD
Maize_1ENetGDD_ASI1.4270.010
Maize_1ENetGDD_DTT1.4290.010
Maize_1LassoGDD_ASI1.4330.011
Maize_1LassoGDD_DTT1.4330.010
Maize_1RidgeGDD_ASI1.3710.009
Maize_1RidgeGDD_DTT1.3710.009
Maize_1WENetGDD_ASI1.3070.005
Maize_1WENetGDD_DTT1.3070.005
Maize_1WLassoGDD_ASI1.3320.006
Maize_1WLassoGDD_DTT1.3320.006
Maize_1WRidgeGDD_ASI1.3030.006
Maize_1WRidgeGDD_DTT1.3030.006
Maize_3ENetGDD_ASI1.4050.011
Maize_3ENetGDD_DTT1.4040.011
Maize_3LassoGDD_ASI1.4090.011
Maize_3LassoGDD_DTT1.4070.011
Maize_3RidgeGDD_ASI1.3800.010
Maize_3RidgeGDD_DTT1.3810.010
Maize_3WENetGDD_ASI1.3690.007
Maize_3WENetGDD_DTT1.3690.007
Maize_3WLassoGDD_ASI1.3700.008
Maize_3WLassoGDD_DTT1.3700.008
Maize_3WRidgeGDD_ASI1.3310.008
Maize_3WRidgeGDD_DTT1.3330.008
Soybean_1ENetR81.2410.003
Soybean_1ENetHeight1.1930.003
Soybean_1LassoR81.2540.003
Soybean_1LassoHeight1.1950.003
Soybean_1RidgeR81.2700.004
Soybean_1RidgeHeight1.1430.002
Soybean_1WENetR81.2740.003
Soybean_1WENetHeight1.2390.003
Soybean_1WLassoR81.2550.003
Soybean_1WLassoHeight1.2270.003
Soybean_1WRidgeR81.2840.004
Soybean_1WRidgeHeight1.1910.003
Soybean_2ENetR81.5540.004
Soybean_2ENetHeight1.3610.002
Soybean_2LassoR81.5590.004
Soybean_2LassoHeight1.3570.002
Soybean_2RidgeR81.5450.004
Soybean_2RidgeHeight1.3420.002
Soybean_2WENetR81.4700.003
Soybean_2WENetHeight1.3160.002
Soybean_2WLassoR81.4730.003
Soybean_2WLassoHeight1.3130.002
Soybean_2WRidgeR81.5150.004
Soybean_2WRidgeHeight1.3320.002
Table 3. Prediction performance across all traits in terms of normalized root mean square error (NRMSE) across the eight datasets. ENet denotes Elastic net regression, Lasso denotes Lasso regression, and Ridge denotes Ridge regression, while WENet, WLasso, and WRidge denote the weighted counterparts of ENet, Lasso, and Ridge. NRMSE_SD denotes the standard deviation of the NRMSE.
Table 3. Prediction performance across all traits in terms of normalized root mean square error (NRMSE) across the eight datasets. ENet denotes Elastic net regression, Lasso denotes Lasso regression, and Ridge denotes Ridge regression, while WENet, WLasso, and WRidge denote the weighted counterparts of ENet, Lasso, and Ridge. NRMSE_SD denotes the standard deviation of the NRMSE.
ModelNRMSE_MeanNRMSE_Mean_SD
Enet1.4090.001
Lasso1.4130.001
Ridge1.3870.001
WENet1.3780.001
WLasso1.3770.001
WRidge1.3690.001
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Montesinos-López, O.A.; Pulido-Carrillo, C.D.; Montesinos-López, A.; Larios Trejo, J.A.; Montesinos-López, J.C.; Agbona, A.; Crossa, J. A Penalized Regression Method for Genomic Prediction Reduces Mismatch between Training and Testing Sets. Genes 2024, 15, 969. https://doi.org/10.3390/genes15080969

AMA Style

Montesinos-López OA, Pulido-Carrillo CD, Montesinos-López A, Larios Trejo JA, Montesinos-López JC, Agbona A, Crossa J. A Penalized Regression Method for Genomic Prediction Reduces Mismatch between Training and Testing Sets. Genes. 2024; 15(8):969. https://doi.org/10.3390/genes15080969

Chicago/Turabian Style

Montesinos-López, Osval A., Cristian Daniel Pulido-Carrillo, Abelardo Montesinos-López, Jesús Antonio Larios Trejo, José Cricelio Montesinos-López, Afolabi Agbona, and José Crossa. 2024. "A Penalized Regression Method for Genomic Prediction Reduces Mismatch between Training and Testing Sets" Genes 15, no. 8: 969. https://doi.org/10.3390/genes15080969

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop