Next Article in Journal
Animal as the Solution II: Phenotyping for Low Milk Urea Nitrogen A1PF Dairy Cows
Previous Article in Journal
Lameness Recognition of Dairy Cows Based on Compensation Behaviour Analysis by Swing and Posture Features from Top View Depth Image
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Immunoglobulin G Goat Colostrum Determination Using Color-Based Techniques and Data Science

by
Manuel Betancor-Sánchez
,
Marta González-Cabrera
,
Antonio Morales-delaNuez
,
Lorenzo E. Hernández-Castellano
,
Anastasio Argüello
* and
Noemí Castro
IUSA-ONEHEALTH 4, Animal Production and Biotechnology, Institute of Animal Health and Food Safety, Universidad de Las Palmas de Gran Canaria, Campus Montaña Cardones, 35413 Arucas, Spain
*
Author to whom correspondence should be addressed.
Animals 2025, 15(1), 31; https://doi.org/10.3390/ani15010031
Submission received: 30 October 2024 / Revised: 4 December 2024 / Accepted: 24 December 2024 / Published: 26 December 2024
(This article belongs to the Section Small Ruminants)

Simple Summary

Newborn goat kids rely on colostrum intake to gain immunity, as the circulating antibodies at birth are not enough to face infectious diseases. Colostrum contains vital proteins like immunoglobulin G (IgG), which protect newborn animals from diseases, as the immune system is not able to synthesize enough antibodies yet. Traditional methods used to determine IgG concentrations are often expensive and not accessible to many farmers. This study explores an affordable solution to predict IgG concentration in goat colostrum by using a color-based method combined with artificial intelligence. By measuring colostrum color and using machine learning models, such as decision trees and neural networks, an accurate prediction method was developed as a practical method to be used on farms. These novel models provide similar results to those obtained using expensive laboratory tests but at a much lower cost. This method could help farmers make better decisions to enhance newborn goat health status, and consequently improve animal welfare, reducing production costs and increasing economic profits.

Abstract

Circulating immunoglobulin G (IgG) concentrations in newborn goat kids are not sufficient to protect the animal against external agents. Therefore, consumption of colostrum, rich in immune components, shortly after birth is crucial. Traditional laboratory methods used to measure IgG concentrations, such as ELISA or RID, are reliable but costly and impractical for many farmers. This study proposes a more accessible alternative for farmers to predict IgG concentration in goat colostrum by integrating color-based techniques with machine learning models, specifically decision trees and neural networks, through the development of two regression models based on colostrum color data from Majorera dairy goats. A total of 813 colostrum samples were collected in a previous study (June 1997–April 2003) that utilized multiple regression analysis as a reference to verify that applying data science techniques improves accuracy and reliability. The decision tree model outperformed the neural network, achieving higher accuracy and lower error rates. Both models provided predictions that closely matched IgG concentrations obtained by ELISA. Therefore, this methodology offers a practical and affordable solution for the on-farm assessment of colostrum quality (i.e., IgG concentration). This approach could significantly improve farm management practices, ensuring better health outcomes in newborn animals by facilitating timely and accurate colostrum quality evaluation.

1. Introduction

Newborn goat kids are born with an immature immune system that seems to be unable to produce its own antibodies [1], mainly immunoglobulin G (IgG), to protect them against external agents. In addition, in this species, there is a limited transplacental transfer of maternal antibodies [2]. In ruminants, the transfer of IgG and other antibodies is performed through colostrum intake during the period immediately postpartum [1]. This process, known as transfer of passive immunity (TPI), is vital for providing immediate protection to newborn goat kids until their immune system starts producing endogenous antibodies [3,4].
Colostrum is a secretion synthesized by the mammary gland during the last eight weeks of gestation [1]. Besides its nutritional function, colostrum contains high concentration of immunoglobulins, and other bioactive compounds such as oligosaccharides, lactoferrin, lysozyme, and some other minor proteins [5] that enhance the immunological development and gastrointestinal health of the offspring. Previous research shows that colostrum composition changes rapidly, within the first 10 to 24 hours following parturition. Thus, the prompt determination of colostrum quality to ensure proper TPI to the newborn goat kid becomes essential [6].
For the quantification of IgG, different laboratory techniques can be used, such as ELISA (enzyme-linked immunosorbent assay) or RID (radial immunodiffusion) [7]. These methods provide a concentration value for IgG that is valid for controlling the quality of colostrum, but the associated cost and complexity are not suitable for most goat farmers.
On the other hand, the most used techniques on the farm are the colostrometer and the Brix refractometer [8,9,10,11]. Although the Brix refractometer and colostrometer can provide an approximate indication of colostrum quality, they lack precision. Direct measurement of IgG concentration remains the most reliable method for assessing the passive immunity provided by colostrum. This is particularly crucial due to external factors, such as the difficulty of maintaining a consistent 20°C when using the colostrometer, which may not always be feasible during colostrum evaluation on farms [12].
Argüello et al. [13] described a novel farm method for assessing IgG concentration in goat colostrum, based on color analysis. They identified a correlation between colostrum color and IgG concentration using conventional statistical techniques. This new method is an important step forward that adds to current farming techniques, though it has not yet reached lab-level precision.
Since 2005, artificial intelligence (AI) has advanced exponentially. Algorithmic advancements, the exponential increases in computing power and storage, and an explosion of data, as highlighted in the McKinsey AI guide [14], have evolved synergistically. This evolution has facilitated access to computational and statistical tools, offering new possibilities for the scientific community.
The utilization of machine learning and deep learning techniques in the field of healthcare and animal production has yielded highly promising results, as evidenced by several authors in recent years [15,16,17]. These techniques have added value to the datasets held by scientific laboratories, enriching knowledge with a fresh perspective from an alternative point of view.
The objective of this study is to integrate AI techniques to predict IgG concentration in goat colostrum by evaluating a color-based method for testing IgG concentration. It will employ both machine learning and deep learning methodologies, employing regression based on decision trees and neural networks, respectively. The goal is to find models that can provide more reliable support in goat farms, using a method that can be implemented at a significantly lower cost than traditional laboratory techniques, without sacrificing a high level of accuracy and reliability.

2. Materials and Methods

2.1. Dataset

The data used in the analyses were collected between June 1997 and April 2003, as detailed in the article published in 2005 [13]. A total of 813 colostrum samples were analyzed. Color and IgG concentration were measured from the first and second milking of the dams, which were performed using a milking machine. The samples were collected from Majorera goats located on four dairy farms in the Canary Islands, Spain.
Serum IgG concentration was determined using the single radial immunodiffusion method [7].
Color data were acquired using a Minolta CR200 Chromameter (Aquatecnis, Madrid, Spain). The color data registered with the Chromameter was CIE L*a*b*, where L* represents relative lightness, a* indicates relative redness, and b* represents relative yellowness [18]. These data were then transformed into CIE L*C*h format for the dataset, which includes the Chroma indicating color purity or intensity, and the hue represented by an angle in a circular color space. CIE L*C*h is a cylindrical representation of the CIE L*a*b* color space, which facilitates a more intuitive understanding of color, in addition to allowing a more direct and meaningful comparison in perceptual terms [19].
For the transformation of values into Chroma, the following formula was used:
C * = ( a * ) 2 + ( b * ) 2
Similarly, the arctangent h = a t a n 2 ( a * , b * ) was used for the calculation of hue angle value [20].
The dataset was randomly shuffled, and subsequently, two new datasets were generated: one for training and one for testing. The training set comprised 80% of the data, with a total of 650 records, while the test set comprised 20%, with a total of 163 records. A random seed was used during the data split to ensure reproducibility of the results. The use of a seed guarantees that the same training and testing subsets are generated each time the experiment is run This allows for consistent comparisons across different models and runs, as well as facilitating debugging and replication in future studies [21].

2.2. Models

In this study, two regression models were selected, one from the field of machine learning and the other from deep learning, to compare their performance against traditional statistical methods. The goal was to assess whether these advanced techniques offered improvements in predictive accuracy and model robustness. By focusing on widely used approaches in each category, a balanced evaluation of their effectiveness can be provided.
For the machine learning approach, decision trees were chosen, as they represent one of the most established and interpretable models in this category [22]. On the deep learning side, a neural network was selected, as it is one of the most used architectures for handling complex data patterns in regression tasks [23]. These models were selected to ensure that the comparison reflects the most representative techniques in each field [24].
The selected models are regression-based, even though a subsequent factorization will be performed to allow for a comparison similar to that conducted in the original study.

2.2.1. Decision Trees

Decision Trees are a widely used machine learning technique for decision-making and predictive modelling [22]. The term “Decision Tree” arises from the graphical representation of the model, which resembles an inverted tree. In this structure, each internal node represents a decision based on a specific attribute, each branch corresponds to the outcome of that decision, and each terminal node (leaf) represents a predicted outcome or class.
Decision trees can be divided into two main types: classification trees and regression trees. Classification trees are used to assign data points to discrete categories, whereas regression trees are employed to predict continuous values.
The construction of a decision tree begins with a dataset composed of input variables (features), which can be either numerical or categorical, and output variables (labels), representing the predicted outcome. The algorithm’s first step is to identify the optimal feature to split the data. This optimal feature is the one that results in the most homogeneous grouping of data points within the branches.
The specific splitting criterion depends on the type of decision tree. For classification tasks, common criteria include Gini impurity or entropy, while for regression tasks, the criterion typically used is mean squared error (MSE). Once the best feature has been selected, the dataset is divided into two or more branches, and this process is applied recursively to each branch.
The tree continues to split the data until a predefined stopping criterion is met. These criteria may include reaching a point where all data points in a branch belong to the same class (in classification tasks), the number of data points in a branch falling below a minimum threshold, or the tree reaching its maximum allowed depth. When further splitting is no longer possible, the branch becomes a leaf node, which provides the final output of the decision tree.
For this study, regression trees were applied, and the tree splitting was performed using mean squared error (MSE) as the criterion to minimize prediction errors.
Given that the dataset used in this study was small and well-balanced, no maximum tree depth was set, allowing for the algorithm to determine the optimal tree depth. However, a minimum depth of two branches was specified to ensure sufficient complexity. Additionally, data balancing techniques were not deemed necessary due to the balanced nature of the dataset.
Lastly, pre-pruning and post-pruning techniques [25] were not applied, as they were not necessary for the scale and characteristics of the data used in this study.

2.2.2. Neural Network

Neural networks are a deep learning technique inspired by the architecture and functioning of the human brain [26]. These computational models are composed of basic units called artificial neurons, which are organized into layers. Each neuron receives inputs, processes the information through a predefined mathematical function (known as an activation function), and generates an output.
Neural networks are organized into three distinct types of layers. The input layer serves as the entry point for the raw data, passing it along to the subsequent layers for further processing. Following the input layer are the hidden layers, which are responsible for the bulk of the network’s processing. Each hidden layer takes the output of the previous layer as its input, applying a series of transformations to capture patterns in the data. Finally, the output layer produces the network’s final result, which can either be a predicted value (in the case of regression tasks) or a classification outcome (for classification tasks).
For this study, a feedforward neural network (FNN) was selected, as it was determined to be the most appropriate model given the structure and characteristics of the dataset [27]. The feedforward architecture ensures efficient forward propagation of data without the complexity of recurrent connections, making it well-suited for tabular and structured datasets like the one used in this study.
Additionally, in deep learning, a normalization function was generated using standard scaling from the training data, as there were no outlier values significantly affecting the mean or standard deviation of the sample [28]. This normalization function was subsequently applied to both the training and test datasets before being fed into the neural network model, ensuring that the data were properly scaled for effective training and performance.
The neural network was designed with several specific characteristics tailored to the dataset. Kernel initialization was performed using a normal distribution, as this method was deemed most appropriate given the dataset’s properties. The network architecture includes four hidden layers, structured as follows: the first layer contains 64 neurons, the second and third layers contain 128 neurons each, and the fourth layer consists of 64 neurons.
To enable the network to learn complex, non-linear relationships in the data, a rectified linear unit (ReLU) activation function was used for all hidden layers [29]. Additionally, L2 regularization was applied to layers two, three, and four, with a regularization factor of 0.01, to reduce the risk of overfitting [30]. The output layer consists of a single neuron, responsible for generating the regression value for the predictive task.
The network’s performance was optimized using MSE as the loss function, which is appropriate for regression tasks, as it calculates the average of the squared differences between predicted and actual values. RMSprop, a gradient-based optimization algorithm, was employed to adjust the model’s weights. This algorithm incorporates a moving average of squared gradients to prevent large oscillations during training, and the learning rate was set to 0.001 for gradual and efficient weight updates.
The model was trained for 2000 epochs, with each epoch representing a full pass over the training data. During each epoch, the network’s weights were updated to minimize the loss function. An early stopping mechanism was also implemented, halting training if no improvement in performance was observed over 200 consecutive epochs. This prevented overfitting and reduced unnecessary computation time.
Figure 1 illustrates a schematic representation of the different layers in the feedforward neural network used in this study. Each layer is labeled with the number of neurons it contains and the activation function applied.

2.3. Performance Evaluation

To assess model performance, different techniques were employed during the training phase and in the final analysis of the results. Specifically, MSE and root mean squared error (RMSE) were used during the initial phase, while mean absolute error (MAE) and the coefficient of determination (r2) were applied to the final predictions [31].
The combination of these two techniques provides a more comprehensive view of model performance. During training, MSE was used to penalize larger errors more heavily, which helped both models, decision trees and neural networks, focus on minimizing significant deviations, thus enhancing the optimization process. Once the predictive values were obtained, MAE allowed for a more straightforward interpretation of the results, as it is expressed in the same units as the predicted values. This combination is effective in providing a more balanced and detailed understanding of model quality.
For comparison with the previous study, a factorization of the target and predicted values was also performed, following the protocol outlined in the original research [13]. Specifically, IgG values exceeding 20 mg/mL were categorized as HIGH, while the remaining values were classified as LOW.
Once the factorization was performed, several comparative metrics were utilized to assess the performance of the current study in comparison to the original. These included the contingency tables, which provides a summary of prediction results, and accuracy, a measure of the overall correctness of the model. Sensitivity (or recall) and specificity were also analyzed to evaluate the model’s ability to correctly identify positive and negative cases, respectively. Additionally, the negative predictive value (NPV) was used to assess the likelihood that negative predictions were correct [32]. Finally, the ROC-AUC (area under the curve) was employed to evaluate the model’s discriminative ability across different thresholds [33].

2.4. Tools and Development Environment

Computational experiments were conducted using Python 3.9.6 [34] in a locally hosted environment. Python was chosen for its versatility and the availability of extensive libraries that facilitated the implementation of both machine learning and deep learning algorithms, as well as data processing tasks. Scikit-learn 1.4.2 was used for the machine learning model [35], and TensorFlow 2.16.2 [36] alongside Keras 3.4.1 [37] was employed for the deep learning model. This setup ensured efficient handling of the computational workload, providing a reliable and stable environment for model development and testing.

3. Results

The results of this study are presented in two sections: The first section focuses on the outcomes derived from two predictive methods: decision trees and neural networks. These models were applied to the objective data, and their performance was evaluated using quantitative metrics, including MAE, MSE, RMSE, and r2. These metrics provide a comprehensive assessment of how accurately each model predicts the given data.
In the second section, the results are factorized to generate data that can be compared to those presented in the original study. This factorization allows for a more refined analysis of the classification outcomes. Key metrics such as accuracy, precision, recall, NPV, and the ROC-AUC are presented to evaluate the performance of the models in a classification context.

3.1. Predictive Model Performance Evaluation

In Table A1, the final color values in the CIE L*C*h format (L, Cr, and Hue) for each of the test dataset records are provided, along with the original IgG values [13] and the predicted values from the regression models. Specifically, the predicted IgG values from the decision tree model are shown in the IgG_p_DT column, while the predicted values from the neural network model are presented in the IgG_p_NN column.
The performance of the regression models was initially evaluated using the MSE and RMSE. The decision tree model achieved an MSE of 3.6571, corresponding to an RMSE of 1.9124. In comparison, the neural network model produced a higher MSE of 5.1804, resulting in an RMSE of 2.2761.
Additionally, the models were assessed using the MAE. The decision tree model achieved a MAE of 0.3206, while the neural network model showed a considerably higher MAE of 1.1076.
Furthermore, the r2 coefficients were calculated to assess the proportion of variance explained by each model. The decision tree model achieved an r2 of 0.9644, indicating that it explains 96.44% of the variance in the data. The neural network model obtained a slightly lower r2 of 0.9541, explaining 95.41% of the variance.
Figure 2, Figure 3 and Figure 4 visually represent the original and predicted data values from the regression models. The scatter plots illustrate the relationship between the original IgG values and the predicted values generated by both the decision tree and neural network models, using the final color values in the CIE L*C*h format (L, Cr, and Hue) as coordinates. These figures provide a clear comparison of model performance and predictive accuracy.

3.2. Classification Metrics After Factorization

In this section, both the target and predicted values were factorized, classifying any value greater than 20 mg/mL of colostrum IgG as HIGH and the remaining values as LOW. This transformation allows for the application of a series of classification metrics, facilitating a direct comparison with the results presented in the original article. By categorizing the values into these two distinct classes. The aim is to assess the models’ performance in distinguishing between high and low IgG levels, providing a more detailed evaluation of their predictive capabilities in line with the classification approach used in previous studies.
Table 1 and Table 2 display the contingency tables generated from the factorization process for both the decision tree and neural network models, respectively. These matrixes provide a detailed overview of the models’ classification performance, illustrating the number of correct and incorrect predictions for the HIGH and LOW categories. The values obtained from these contingency tables will be used in the subsequent calculation of performance metrics, allowing for a more comprehensive comparison of the models’ classification accuracy based on the established threshold.
The decision tree model achieved an overall accuracy of 0.9816. The model demonstrated a sensitivity of 1.0 and a specificity of 0.9741. Additionally, NPV was 1.0, indicating that all predictions classified as LOW were accurate.
The neural network model achieved an overall accuracy of 0.9632. The model demonstrated a sensitivity of 0.94 and a specificity of 0.9735. Additionally, NPV was 0.9735.
Figure 5 and Figure 6 show the ROC-AUCs for the decision tree and neural network models, respectively.

4. Discussion

4.1. Results Overview

The evaluation of the models’ performance, as detailed in Table A1, reveals notable differences between the decision tree and neural network models in terms of error metrics and explanatory power. Analyzing the MSE and RMSE, the decision tree model demonstrated better performance with an MSE of 3.6571 and an RMSE of 1.9124, compared to the neural network’s higher MSE of 5.1804 and RMSE of 2.2761. These values indicate that the decision tree model makes more accurate predictions overall, producing lower deviations from the actual IgG values.
The MAE further supports this observation, with the decision tree model showing lower MAE of 0.3206, while the neural network model yielded a higher MAE of 1.1076. This disparity suggests that the decision tree model consistently produces smaller prediction errors, while the neural network model exhibits greater variability in its predictions.
Importantly, considering the MAE in relation to the expected IgG concentration range in colostrum, the relatively low error rates of both models indicate that their precision approaches that of laboratory techniques. The decision tree’s low MAE suggests that its predictions are highly accurate, making it a practical tool for being used in farm settings where quick and reliable IgG estimates are needed. Although the neural network model has a higher MAE, its accuracy is still within a range that could be useful for practical applications in livestock management.
In terms of the coefficient of determination, which measures the proportion of variance explained by the models, both models performed well, although the decision tree model again outperformed the neural network. The decision tree model achieved an r2 of 0.9644, while the neural network model explained 95.41% (r2 = 0.9541). While this difference is small, it indicates that the decision tree model captures the underlying patterns in the data more effectively.
In addition to the evaluation of the continuous predictions, the target and predicted values were factorized into two categories (i.e., HIGH and LOW) setting as a threshold IgG concentration in colostrum (i.e., 20 mg/mL). This binary classification enabled the application of various classification metrics, aligning with the approach used in the original study and facilitating direct comparisons. The transformation provided clearer assessment of the models’ ability to distinguish between high and low IgG concentrations, which is particularly important for practical decision-making in farm environments.
The contingency tables generated for both the decision tree and neural network models offer a detailed view of their classification performance. These tables summarize the correct and incorrect classifications for each category (i.e., HIGH and LOW), allowing for a straightforward calculation of the models’ accuracy, sensitivity, specificity, and other key metrics. The decision tree model demonstrated near-perfect classification performance, with an accuracy of 98.16%, meaning that almost all predictions were correct. Additionally, its high sensitivity (1.0) underscores its ability to identify high IgG cases without error, while its specificity of 0.9741 indicates the ability to identify low IgG cases. The negative predictive value (NPV) of 1.0 further highlights the reliability of the decision tree model in predicting LOW cases, suggesting that no false negatives occurred in this category.
The neural network model also performed well, achieving an overall accuracy of 96.32%. While its sensitivity (0.94) was slightly lower than that of the decision tree, it still correctly identified 94% of the high IgG cases. Its specificity (0.9735) and NPV (0.9735) were comparable to the decision tree, indicating a similar performance in identifying low IgG cases and avoiding false negatives. Despite the marginal differences between the two models, both demonstrated strong predictive power, with very few misclassifications across both categories.
The ROC-AUCs, presented in Figure 5 and Figure 6, provide an additional layer of evaluation for these models. The decision tree model achieved an AUC of 0.97, while the neural network model followed closely with an AUC of 0.96. These high AUC values reflect the models’ excellent ability to distinguish between the HIGH and LOW categories across a range of thresholds, further reinforcing the robustness of the classification models.
Overall, the results from both regression and classification analyses suggest that the decision tree model offers better predictive accuracy and consistency compared to the neural network model. The decision tree’s lower error rates and higher r2 value indicate its superior ability to capture the nuances of the dataset, while the neural network’s higher error rates suggest a greater susceptibility to overfitting. Nevertheless, both models demonstrate strong predictive power, explaining over 95% of the variance and showing remarkable precision, particularly in terms of MAE, making them viable for practical implementation in farm environments, where their predictive accuracy closely mirrors that of laboratory methods.
Furthermore, the factorization of values and subsequent application of classification metrics reinforce the decision tree model’s edge in performance, particularly in terms of sensitivity and NPV, which are critical for accurately identifying high IgG cases. Despite this advantage, the neural network model also performed admirably, offering comparable specificity and overall strong classification capabilities. Both models present reliable and robust tools for distinguishing between high and low IgG levels, a crucial task for optimizing farm management and improving animal health outcomes.
The dataset used in this study exclusively comprises data from Majorera goats located on four dairy farms in the Canary Islands, Spain. This specificity was necessary to ensure a direct and reliable comparison with previous studies using similar methodologies. However, the methods presented in this article lay the groundwork for future analyses involving different breeds, diets, and environmental conditions. By applying these techniques to other goat populations, researchers could further explore the adaptability and generalizability of the models, thereby expanding their applicability to a wider range of farming contexts.

4.2. Comparison with the Previous Study

While the previous study employed regression models, access to the predictions generated by the linear regression models used in that study was not available. This limitation prevented the calculation of key performance metrics, such as MAE or MSE, which are the focus of Section 3.1. Consequently, a direct comparison based on these criteria is not feasible, restricting the ability to perform a detailed analysis of the regression models’ predictive performance under identical conditions.
However, examining the results obtained from factored data, corresponding to those in Section 3.2, the original study reported moderate classification performance. Accuracy and sensitivity were relatively high, but specificity and NPV were notably lower. These metrics serve as a useful reference point to assess the improvements observed with the methods applied in this study, despite the lack of error measures from the original models.
In contrast to the original study’s regression models [13], the decision tree model employed here demonstrated marked improvements in classification performance. Specifically, the decision tree achieved an accuracy of 0.9816, a sensitivity of 1.0, a specificity of 0.9741, and an NPV of 1.0. These results represent a substantial enhancement in both the identification of positive and negative cases compared to the original study, which achieved an accuracy of 0.8745, a sensitivity of 0.9303, a specificity of 0.7143, and an NPV of 0.78125. The high sensitivity and NPV values in the decision tree model underscore its superior performance, particularly in avoiding false negatives.
Similarly, the neural network model showed notable improvements, with an accuracy of 0.9632, a sensitivity of 0.94, a specificity of 0.9735, and an NPV of 0.9735. While its sensitivity is slightly lower than that of the decision tree model, it still surpasses the original regression model across all metrics. The high specificity and NPV values, combined with strong accuracy, further highlight the effectiveness of machine learning techniques such as neural networks in handling complex biological data. These advancements underscore the reliability of these models in more effectively predicting both positive and negative cases, leading to more accurate and consistent outcomes.

4.3. Advancements Through Machine Learning (ML) and Deep Learning (DL) in Biological Data

The application of data science, particularly through ML and DL, has enhanced the analysis of biological data, offering significant advancements over traditional statistical techniques. While classical methods such as linear regression or logistic models have long been used to model biological processes, their limitations in handling complex, high-dimensional datasets have become increasingly evident. In contrast, ML and DL algorithms excel at identifying intricate patterns and relationships within large datasets, providing more accurate and nuanced predictions. These techniques enable models to capture non-linear interactions that traditional approaches often overlook, resulting in improved precision and reliability in biological research.
Recent publications have consistently demonstrated the benefits of employing ML and DL in the animal sciences, with numerous examples illustrating their capacity to enhance predictive accuracy and uncover hidden insights [38,39,40]. From genomics and proteomics to disease prediction and animal health monitoring, these advanced methods have allowed for researchers to tackle previously intractable problems [41]. The use of ML and DL has led to breakthroughs in areas such as biomarker discovery, diagnostic tool development, and precision farming, showcasing their transformative potential across diverse biological fields [42,43]. This shift towards data-driven methodologies marks a significant evolution in the way biological data are analyzed and applied, offering new opportunities for innovation and efficiency.
A growing body of research demonstrates the better performance of ML and DL algorithms compared to traditional statistical approaches like multiple linear regression (MLR) in the analysis of complex biological datasets. For instance, Chen et al. [44] compared the prediction of nitrogen (N) excretion in lactating dairy cows found that ML or DL techniques such as artificial neural networks (ANNs), random forest regression (RFR), and support vector regression (SVR) outperformed MLR models in both accuracy and precision. In this study, the ANN model achieved significantly lower RMSE and higher concordance correlation coefficient (CCC) compared to the MLR model, underscoring the capacity of DL algorithms to handle the non-linear relationships and complex interactions present in biological systems. Thus, Chen et al. [44] highlights an important challenge inherent to MLR models: their reliance on assumptions such as linearity, homoscedasticity, and normality of residuals, which may not always hold true in complex datasets. The authors note that MLR models, while useful in certain contexts, can lead to biased results or fail to provide satisfactory predictions when these assumptions are violated. In contrast, ML models, and particularly ANN models, demonstrated the ability to explore deeper relationships between the variables and outputs, improving the predictive power without the need for strict assumptions about the data. These findings align with broader trends in biological research, where the shift towards data-driven methodologies has become essential for dealing with increasingly complex datasets.
Moreover, the study developed new ANN models for the prediction of manure nitrogen excretion. These models were able to predict with greater accuracy than traditional MLR methods. Notably, the ANN model produced a lower RMSE and higher CCC compared to the MLR, reflecting its superior ability to generalize across the dataset. This is a critical advantage when applied to practical situations in dairy farms, where accurate predictions of nitrogen excretion are necessary to mitigate environmental impacts and enhance economic sustainability.
The study by Hansen et al. [45] explores how modern ML, DL, and statistical methods can enhance the forecasting of milk deliveries to dairy plants compared to traditional techniques. Using historical data from Norwegian dairy farms, the authors evaluated several models, including seasonal ARIMA, LASSO, Group LASSO, the Prophet model, Boosting, and neural network autoregression (NNAR). The primary goal was to improve prediction accuracy over different time horizons, ranging from 1 to 24 months. The results demonstrated that ML and DL models, such as Boosting and NNAR, significantly outperformed traditional statistical methods in long-term forecasting, especially by identifying key features that influence future milk deliveries, such as the number of cows, inseminations, and calvings per month.
In particular, the NNAR model produced the best results for 12-month forecasts, while the SARIMAX model was more accurate for 24-month predictions. The study highlights that data-driven models like Boosting provided greater flexibility and accuracy, especially in long-term predictions, due to their ability to handle non-linear relationships and complex features in the data.
Similar to the present study, Hansen et al. [45] demonstrates that integrating ML and DL techniques offers significant improvements over traditional multiple linear regression models. While conventional approaches like ARIMA and the Prophet model remain useful for short-term predictions, more advanced models such as Boosting and neural networks deliver superior predictive capability, particularly when dealing with complex, multivariate datasets.
Hu et al. [43] investigated the application of ML techniques in combination with hyperspectral imagery from unmanned aerial vehicles (UAVs) to estimate the biomass of milk vetch, a winter-growing cover crop known for its ability to enhance soil fertility. The authors compared several regression models—random forest (RF), MLR, support vector machine (SVM), and deep neural network (DNN)—to evaluate their performance in biomass prediction. The research found that RF, which is an ensemble method based on decision trees, achieved the highest coefficient of determination (R2 = 0.950) and the lowest relative root mean square error (RRMSE = 14.86%) among all models. Notably, the DNN model also performed well on the test set, slightly surpassing RF in some respects, such as its performance in the second year of data collection.
The study highlights the significant potential of combining UAV-based hyperspectral data with ML techniques to perform large-scale, non-invasive biomass estimations. Hyperspectral imagery allowed for the researchers to compute vegetation indices (VIs) that served as inputs to the ML models. Through feature selection using Pearson correlation and principal component analysis (PCA), the VIs most strongly correlated with biomass were identified, ensuring that the models were built with the most relevant predictors. The results demonstrated that ML models, particularly RF and DNN, provide accurate predictions, significantly outperforming traditional methods like MLR in biomass estimation.
The parallels between the findings from these studies and our research further reinforce the growing consensus that ML and DL techniques offer a transformative approach to biological data analysis. By surpassing the limitations of traditional MLR, these advanced methods not only improve predictive accuracy but also provide deeper insights into the complex relationships between variables, such as those observed in animal health, agricultural productivity, and environmental management. As the integration of ML and DL models becomes increasingly prevalent in biological research, it is evident that these tools are indispensable for advancing both scientific understanding and practical applications in fields that rely on accurate, large-scale data analysis.

4.4. Future Implications

As described above, the determination of IgG concentration in goat colostrum is crucial, for ensuring the survival of newborn kids. High circulating IgG concentration is vital for providing adequate immunity to these animals during their early stages of life. However, while accurate IgG measurements are essential, precise laboratory methods are often costly and inaccessible for many farmers, who are forced to rely on less accurate farm-based techniques [8]. These alternative methods, while more affordable, do not offer the same level of precision, leading to potential risks in the health and survival of young animals.
The results of this study demonstrated that both the decision tree and neural network models are able to predict the IgG concentration that closely approximate to those obtained through laboratory testing. This means that by implementing these models through a simple and cost-effective system, farmers could achieve reliability levels exceeding 95%. Such an approach would provide farmers with a practical solution that bridges the gap between accuracy and affordability, without the need for expensive lab process.
Moreover, these models can be continuously refined and improved by integrating more data over time, further enhancing their reliability. They are designed to be user-friendly and can be implemented in practical farm settings through integration into portable devices or user-friendly software. This would enable farmers to measure colostrum quality effectively with minimal technical expertise. The data required to operate these models can be readily collected using accessible tools, ensuring that even small-scale farmers can benefit from their use. This scalability and adaptability make these models a highly valuable tool for ensuring livestock health and well-being, contributing to more sustainable and efficient farm management practices in the future.

5. Conclusions

In this study, AI techniques were successfully integrated to evaluate IgG concentration in goat colostrum using a low-cost color method. By employing both ML and DL methodologies, specifically regression based on decision trees and neural networks, models were developed to provide reliable support for goat farms without incurring the high costs associated with traditional laboratory techniques. The results indicate that the decision tree regression model outperformed the neural network model across multiple metrics.
These findings suggest that the decision tree regression model provides a highly accurate and reliable method for assessing IgG concentration, offering a cost-effective solution for goat farms.

Author Contributions

Conceptualization, M.B.-S., A.A. and N.C.; Methodology, M.B.-S.; Software, M.B.-S.; Investigation, M.B.-S.; Data curation, M.B.-S.; Writing—original draft, M.B.-S.; Writing—review and editing, M.G.-C., A.M.-d., L.E.H.-C. and N.C.; Supervision, M.G.-C., A.M.-d., L.E.H.-C. and N.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in Appendix A of this article.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Original and predicted values from decision tree and neural network regression models.
Table A1. Original and predicted values from decision tree and neural network regression models.
LCrHueIgGIgG_p_DTIgG_p_NN
83.8628.40103.5628.7128.7130.98
94.097.68113.233.973.973.59
90.3423.28107.1424.3024.3024.78
94.626.48106.671.781.781.45
77.8821.5499.7638.1938.1939.41
89.3318.90105.5032.9632.9633.40
90.1711.12111.0319.6919.6919,00
90.9810.59114.241.921.920.90
88.668.48107.0715.4815.4815.49
92.598.70101.8012.8012.8012.45
91.008.00108.005.005.004.40
89.8921.57106.1024.6224.6225.03
87.9718.06108.1916.8716.8716.82
87.0120.36105.1137.4037.4038.65
88.7324.22106.3833.3033.3034.01
91.1018.47108.8320.7120.7120.39
88.4515.92105.5516.8016.8016.81
93.1211.32108.488.328.327.29
93.5010.03105.264.604.603.93
84.6816.76105.8233.9433.9433.87
90.368.32100.3911.4211.4211.54
91.0811.01107.175.085.084.59
90.316.56109.404.754.754.26
89.719.95110.606.657.5220.94
92.009.00108.003.003.002.89
94.529.95111.902.552.551.49
91.0018.00108.0018.0018.0018.03
90.0015.00110.0022.0022.0022.23
88.0015.00106.0022.0022.0021.65
93.037.30110.776.006.004.64
92.0011.00113.009.009.008.15
92.898.90110.597.527.527.20
87.3320.62101.9524.4424.4425.02
89.9411.12108.6817.3817.3816.08
91.6512.39108.9814.4714.4714,00
90.9810.59114.241.921.920.90
88.8913.97101.236.166.165.58
91.469.25111.240.500.500.52
91.6811.21111.394.914.914.02
91.289.24110.4018.7918.7917.41
92.007.00111.005.005.003.90
92.0010.00111.0015.0013.3312.65
87.9718.06108.1916.8716.8716.82
91.668.66115.5011.3711.3710.28
90.0015.00110.0022.0022.0022.23
90.0020.00107.0022.0022.0022.47
83.7621.91102.2333.6433.6434.89
75.0029.0098.0032.0032.0034.56
90.459.62111.9210.8310.8310.51
94.529.95111.902.552.551.49
90.8715.93105.6211.5111.5112.84
92.608.67114.0211.4211.428.59
91.758.54114.044.264.263.75
84.6218.87105.1511.2111.2111.86
94.237.87113.602.342.342.19
92.459.05107.8113.0513.0511.07
85.0813.08101.346.676.675.92
89.719.95110.606.657.5220.94
93.927.53109.233.053.052.75
93.366.30113.681.031.032.37
91.168.84109.3010.7210.729.13
89.719.95110.606.657.5220.94
88.3411.48103.819.899.898.96
88.668.48107.0715.4815.4815.49
88.0718.24105.0024.3624.3625.37
88.5815.82106.6814.8514.8514.86
92.567.74116.631.301.300.98
92.618.69109.644.084.083.98
93.047.00113.497.267.267.29
90.6114.80109.5913.2313.2314.38
89.3611.86108.1810.8410.849.11
86.6618.03102.1323.4523.4523.87
92.569.35110.224.464.465.48
88.6919.18105.3336.3536.3537.29
92.975.41107.872.602.602.43
88.977.55114.432.562.562.46
88.3620.44105.5822.7922.7923.76
85.019.32101.770.820.82-0.45
87.3614.40103.998.038.037.59
91.127.63114.898.468.468.47
88.0020.00106.0022.0022.0023.04
88.949.4697.1614.3114.3113.85
84.6816.76105.8233.9433.9433.87
90.8313.84107.235.445.445.15
92.007.00115.006.006.003.92
86.0720.07109.2023.4423.4423.58
87.2621.4105.0628.1028.1031.99
93.008.11117.076.146.144.22
86.2411.53104.774.724.723.38
92.756.49110.284.484.483.80
91.6223.87105.8332.3732.3732.73
84.6218.87105.1511.2111.2111.86
88.5513.03108.025.515.514.65
93.3110.57107.517.047.046.75
89.836.93123.236.926.925.15
91.7115.88111.0722.548.7721.10
93.927.53109.233.053.052.75
93.509.58107.6811.4011.4010.91
88.5513.03108.025.515.514.65
88.6823.52105.1122.3522.3523.11
92.009.00108.003.003.002.89
92.569.35110.224.464.465.48
96.7521.11104.1723.0623.0618.93
92.7213.89109.3010.6010.6010.76
83.7621.91102.2333.6433.6434.89
89.5018.99103.4923.9823.9823.67
94.6811.24108.477.277.275.22
93.248.08115.1113.8013.808.19
87.227.59111.665.875.875.60
91.3211.58109.5820.2720.2719.79
91.089.32106.455.095.095.44
93.8217.25106.9925.0625.0623.91
89.8018.5107.2622.3822.3822.54
98.539.27105.273.513.511.40
90.9917.46107.6025.3525.3525.33
88.3718.39103.3017.5117.5118.40
88.5311.77102.476.516.516.10
93.879.90108.071.031.030.82
91.3213.65106.0718.3818.3817.65
83.2815.55103.728.438.438.94
93.5811.45109.3414.5114.5112.83
91.2811.33107.8922.4522.4521.75
89.5018.99103.4923.9823.9823.67
82.1923.73100.2734.0034.0036.05
89.3318.90105.5032.9632.9633.40
92.0010.00111.0010.0013.3312.65
92.0010.88110.229.539.539.08
92.0011.00113.009.009.008.15
91.7115.88111.0722.548.7721.10
91.177.89112.832.382.382.53
92.005.00110.003.003.002.30
91.397.54123.585.415.412.51
88.0017.00101.0028.0028.0029.07
90.0923.46107.8525.9725.9726.55
88.8913.97101.236.166.165.584
92.0315.42109.958.778.778.59
91.808.75106.883.833.833.77
89.2518.46107.1128.4928.4928.13
91.0218.30106.9013.0713.0713.49
92.808.24109.864.074.073.65
91.127.63114.898.468.468.47
92.317.85112.788.988.989.34
79.9922.0598.5622.8522.8523.76
90.2410.20110.4219.7319.7319.05
93.4610.48109.334.684.683.93
88.1327.58106.9231.6631.6632.91
85.2410.87105.150.900.900.33
90.9012.23109.0015.4615.4615.13
94.698.77107.737.437.436.35
93.387.99115.574.674.675.12
93.8217.25106.9925.0625.0623.91
91.7412.05109.347.227.227.64
90.316.56109.404.754.754.26
91.526.39120.380.700.70-0.11
90.0014.00105.0020.0020.0019.71
91.7012.93110.1317.2017.2017.07
75.0029.0098.0032.0032.0034.56
62.5428.8298.9431.3931.3932.34
92.9510.73114.2711.2711.279.60
90.0010.00100.0015.0015.0014.56
91.7115.88111.0722.548.7721.10
83.8628.40103.5628.7128.7130.98
92.0010.00111.0010.0013.3312.65

References

  1. Castro, N.; Capote, J.; Bruckmaier, R.M.; Argüello, A. Management Effects on Colostrogenesis in Small Ruminants: A Review. J. Appl. Anim. Res. 2011, 39, 85–93. [Google Scholar] [CrossRef]
  2. Constant, S.B.; LeBlanc, M.M.; Klapstein, E.F.; Beebe, D.E.; Leneau, H.M.; Nunier, C.J. Serum Immunoglobulin G Concentration in Goat Kids Fed Colostrum or a Colostrum Substitute. J. Am. Vet. Med. Assoc. 1994, 205, 1759–1762. [Google Scholar] [CrossRef] [PubMed]
  3. Argüello, A.; Castro, N.; Capote, J.; Tyler, J.W.; Holloway, N.M. Effect of Colostrum Administration Practices on Serum IgG in Goat Kids. Livest. Prod. Sci. 2004, 90, 235–239. [Google Scholar] [CrossRef]
  4. Castro, N.; Capote, J.; Morales-Delanuez, A.; Rodríguez, C.; Argüello, A. Effects of Newborn Characteristics and Length of Colostrum Feeding Period on Passive Immune Transfer in Goat Kids. J. Dairy Sci. 2009, 92, 1616–1619. [Google Scholar] [CrossRef]
  5. Hernández-Castellano, L.E.; Almeida, A.M.; Renaut, J.; Argüello, A.; Castro, N. A Proteomics Study of Colostrum and Milk from the Two Major Small Ruminant Dairy Breeds from the Canary Islands: A Bovine Milk Comparison Perspective. J. Dairy Res. 2016, 83, 366–374. [Google Scholar] [CrossRef]
  6. Moreno-Indias, I.; Sánchez-Macías, D.; Castro, N.; Morales-delaNuez, A.; Hernández-Castellano, L.E.; Capote, J.; Argüello, A. Chemical Composition and Immune Status of Dairy Goat Colostrum Fractions during the First 10h after Partum. Small Rumin. Res. 2012, 103, 220–224. [Google Scholar] [CrossRef]
  7. Mancini, G.; Carbonara, A.O.; Heremans, J.F. Immunochemical Quantitation of Antigens by Single Radial Immunodiffusion. Immunochemistry 1965, 2, 235–486. [Google Scholar] [CrossRef]
  8. Bartier, A.; Windeyer, M.; Doepel, L. Evaluation of On-Farm Tools for Colostrum Quality Measurement. J. Dairy Sci. 2015, 98, 1878–1884. [Google Scholar] [CrossRef]
  9. Buranakarl, C.; Thammacharoen, S.; Nuntapaitoon, M.; Semsirmboon, S.; Katoh, K. Validation of Brix Refractometer to Estimate Immunoglobulin G Concentration in Goat Colostrum. Vet. World 2021, 14, 3194–31999. [Google Scholar] [CrossRef]
  10. Castro, N.; Gómez-González, L.A.; Earley, B.; Argüello, A. Use of Clinic Refractometer at Farm as a Tool to Estimate the Igg Content in Goat Colostrum. J. Appl. Anim. Res. 2018, 46, 1505–1508. [Google Scholar] [CrossRef]
  11. Zobel, G.; Rodriguez-Sanchez, R.; Hea, S.; Weatherall, A.; Sargent, R. Validation of Brix Refractometers and a Hydrometer for Measuring the Quality of Caprine Colostrum. J. Dairy Sci. 2020, 103, 9277–9289. [Google Scholar] [CrossRef] [PubMed]
  12. Mechor, G.D.; Gröhn, Y.T.; McDowell, L.R.; Van Saun, R.J. Specific Gravity of Bovine Colostrum Immunoglobulins as Affected by Temperature and Colostrum Components. J. Dairy Sci. 1992, 75, 3131–3135. [Google Scholar] [CrossRef] [PubMed]
  13. Argüello, A.; Castro, N.; Capote, J. Short Communication: Evaluation of a Color Method for Testing Immunoglobulin g Concentration in Goat Colostrum. J. Dairy Sci. 2005, 88, 1752–1754. [Google Scholar] [CrossRef] [PubMed]
  14. Chui, M.; McCarthy, B.; McKinsey. An Executive’s Guide to AI. Available online: https://www.mckinsey.com/capabilities/quantumblack/our-insights/an-executives-guide-to-ai (accessed on 23 December 2024).
  15. Ali AlZubi, A. Artificial Intelligence and Its Application in the Prediction and Diagnosis of Animal Diseases: A Review. Indian J. Anim. Res. 2023, 57, 1265–1271. [Google Scholar] [CrossRef]
  16. Denholm, S.J.; Brand, W.; Mitchell, A.P.; Wells, A.T.; Krzyzelewski, T.; Smith, S.L.; Wall, E.; Coffey, M.P. Predicting Bovine Tuberculosis Status of Dairy Cows from Mid-Infrared Spectral Data of Milk Using Deep Learning. J. Dairy Sci. 2020, 103, 9355–9367. [Google Scholar] [CrossRef]
  17. Neto, H.A.; Tavares, W.L.F.; Ribeiro, D.C.S.Z.; Alves, R.C.O.; Fonseca, L.M.; Campos, S.V.A. On the Utilization of Deep and Ensemble Learning to Detect Milk Adulteration. BioData Min. 2019, 12, 13. [Google Scholar] [CrossRef]
  18. Schanda, J. Colorimetry: Understanding the CIE System; Schanda, J., Ed.; John Wiley & Sons: Hoboken, NJ, USA, 2007; ISBN 9780470049044. [Google Scholar]
  19. Berns, R.S. Billmeyer and Saltzman’s: Principles of Color Technology, 4th ed.; John Wiley & Sons: Hoboken, NJ, USA, 2019; ISBN 9781119367314. [Google Scholar]
  20. Konica Minolta Sensing. Color Spaces. Available online: https://sensing.konicaminolta.us/us/learning-center/color-measurement/color-spaces/ (accessed on 12 June 2024).
  21. Ziegler, A. An Introduction to Statistical Learning with Applications; James, R.G., Witten, D., Hastie, T., Tibshirani, R., Eds.; Springer: Berlin, Germany, 2013; p. 440. ISBN 978-1-4614-7138-7. [Google Scholar]
  22. McClarren, R.G. Decision Trees and Random Forests for Regression and Classification. In Machine Learning for Engineers: Using Data to Solve Problems for Physical Systems; McClarren, R.G., Ed.; Springer International Publishing: Cham, Switzerland, 2021; pp. 55–82. ISBN 978-3-030-70388-2. [Google Scholar]
  23. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT: Cambridge, MA, USA, 2016. [Google Scholar]
  24. Molnar, C. Interpretable Machine Learning, 2nd ed.; Leanpub: Victoria, BC, Canada, 2022; ISBN 978-1-09-812017-0. [Google Scholar]
  25. Quinlan, J.R. Pruning Decision Trees. In C4.5; Elsevier: Amsterdam, The Netherlands, 1993; pp. 35–43. ISBN 978-1-55860-238-0. [Google Scholar]
  26. Haykin, S.S. Neural Networks and Learning Machines; Prentice Hall/Pearson: London, UK, 2009; ISBN 9780131471399. [Google Scholar]
  27. Bebis, G.; Georgiopoulos, M. Feed-Forward Neural Networks. IEEE Potentials 1994, 13, 27–31. [Google Scholar] [CrossRef]
  28. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Springer: Berlin/Heidelberg, Germany, 2009; ISBN 9780387848570. [Google Scholar]
  29. Agarap, A.F. Deep Learning Using Rectified Linear Units (ReLU). arXiv 2018, arXiv:1803.08375. [Google Scholar]
  30. Cortes, C.; Research, G.; York, N.; Mohri, M.; Rostamizadeh, A. L 2 Regularization for Learning Kernels. arXiv 2012, arXiv:1205.2653. [Google Scholar]
  31. Hodson, T.O. Root-Mean-Square Error (RMSE) or Mean Absolute Error (MAE): When to Use Them or Not. Geosci. Model. Dev. 2022, 15, 5481–5487. [Google Scholar] [CrossRef]
  32. Sokolova, M.; Lapalme, G. A Systematic Analysis of Performance Measures for Classification Tasks. Inf. Process Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
  33. Fawcett, T. An Introduction to ROC Analysis. Pattern Recognit. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
  34. Python 3.9.6 Documentation. Available online: https://docs.python.org/release/3.9.6/ (accessed on 23 December 2024).
  35. API Reference—Scikit-Learn 1.4.2 Documentation. Available online: https://scikit-learn.org/1.4/modules/classes.html (accessed on 13 June 2024).
  36. API Documentation. Available online: https://www.tensorflow.org/api_docs (accessed on 23 December 2024).
  37. Keras 3 API Documentation. Available online: https://keras.io/api/ (accessed on 23 December 2024).
  38. Zhang, L.; Han, G.; Qiao, Y.; Xu, L.; Chen, L.; Tang, J. Interactive Dairy Goat Image Segmentation for Precision Livestock Farming. Animals 2023, 13, 3250. [Google Scholar] [CrossRef] [PubMed]
  39. Gonçalves, P.; Marques, M.d.R.; Belo, A.T.; Monteiro, A.; Morais, J.; Riegel, I.; Braz, F. Exploring the Potential of Machine Learning Algorithms Associated with the Use of Inertial Sensors for Goat Kidding Detection. Animals 2024, 14, 938. [Google Scholar] [CrossRef]
  40. Solis, I.L.; de Oliveira-Boreli, F.P.; de Sousa, R.V.; Martello, L.S.; Pereira, D.F. Using Thermal Signature to Evaluate Heat Stress Levels in Laying Hens with a Machine-Learning-Based Classifier. Animals 2024, 14, 1996. [Google Scholar] [CrossRef]
  41. Pedrosa, V.B.; Chen, S.-Y.; Gloria, L.S.; Doucette, J.S.; Boerman, J.P.; Rosa, G.J.M.; Brito, L.F. Machine Learning Methods for Genomic Prediction of Cow Behavioral Traits Measured by Automatic Milking Systems in North American Holstein Cattle. J. Dairy Sci. 2024, 107, 4758–4771. [Google Scholar] [CrossRef]
  42. García-Infante, M.; Castro-Valdecantos, P.; Delgado-Pertíñez, M.; Teixeira, A.; Guzmán, J.L.; Horcada, A. Effectiveness of Machine Learning Algorithms as a Tool to Meat Traceability System. A Case Study to Classify Spanish Mediterranean Lamb Carcasses. Food Control 2024, 164, 110604. [Google Scholar] [CrossRef]
  43. Hu, H.; Zhou, H.; Cao, K.; Lou, W.; Zhang, G.; Gu, Q.; Wang, J. Biomass Estimation of Milk Vetch Using UAV Hyperspectral Imagery and Machine Learning. Remote Sens. 2024, 16, 2183. [Google Scholar] [CrossRef]
  44. Chen, X.; Zheng, H.; Wang, H.; Yan, T. Can Machine Learning Algorithms Perform Better than Multiple Linear Regression in Predicting Nitrogen Excretion from Lactating Dairy Cows. Sci. Rep. 2022, 12, 12478. [Google Scholar] [CrossRef]
  45. Hansen, B.G.; Li, Y.; Sun, R.; Schei, I. Forecasting Milk Delivery to Dairy—How Modern Statistical and Machine Learning Methods Can Contribute. Expert Syst. Appl. 2024, 248, 123475. [Google Scholar] [CrossRef]
Figure 1. Architecture of the feedforward neural network.
Figure 1. Architecture of the feedforward neural network.
Animals 15 00031 g001
Figure 2. Three-dimensional plot of IgG values based on L, Cr, and Hue.
Figure 2. Three-dimensional plot of IgG values based on L, Cr, and Hue.
Animals 15 00031 g002
Figure 3. Three-dimensional plot of decision tree prediction values based on L, Cr, and Hue.
Figure 3. Three-dimensional plot of decision tree prediction values based on L, Cr, and Hue.
Animals 15 00031 g003
Figure 4. Three-dimensional plot of neural network prediction values based on L, Cr, and Hue.
Figure 4. Three-dimensional plot of neural network prediction values based on L, Cr, and Hue.
Animals 15 00031 g004
Figure 5. ROC-AUCs for the decision tree regression model.
Figure 5. ROC-AUCs for the decision tree regression model.
Animals 15 00031 g005
Figure 6. ROC-AUC curves for the neural network regression model.
Figure 6. ROC-AUC curves for the neural network regression model.
Animals 15 00031 g006
Table 1. Contingency tables for decision tree regression model.
Table 1. Contingency tables for decision tree regression model.
HIGHLOW
HIGH47350
LOW0113113
47116
Table 2. Contingency tables for neural network regression model.
Table 2. Contingency tables for neural network regression model.
HIGHLOW
HIGH47350
LOW3110113
50113
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Betancor-Sánchez, M.; González-Cabrera, M.; Morales-delaNuez, A.; Hernández-Castellano, L.E.; Argüello, A.; Castro, N. Enhancing Immunoglobulin G Goat Colostrum Determination Using Color-Based Techniques and Data Science. Animals 2025, 15, 31. https://doi.org/10.3390/ani15010031

AMA Style

Betancor-Sánchez M, González-Cabrera M, Morales-delaNuez A, Hernández-Castellano LE, Argüello A, Castro N. Enhancing Immunoglobulin G Goat Colostrum Determination Using Color-Based Techniques and Data Science. Animals. 2025; 15(1):31. https://doi.org/10.3390/ani15010031

Chicago/Turabian Style

Betancor-Sánchez, Manuel, Marta González-Cabrera, Antonio Morales-delaNuez, Lorenzo E. Hernández-Castellano, Anastasio Argüello, and Noemí Castro. 2025. "Enhancing Immunoglobulin G Goat Colostrum Determination Using Color-Based Techniques and Data Science" Animals 15, no. 1: 31. https://doi.org/10.3390/ani15010031

APA Style

Betancor-Sánchez, M., González-Cabrera, M., Morales-delaNuez, A., Hernández-Castellano, L. E., Argüello, A., & Castro, N. (2025). Enhancing Immunoglobulin G Goat Colostrum Determination Using Color-Based Techniques and Data Science. Animals, 15(1), 31. https://doi.org/10.3390/ani15010031

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop