Next Article in Journal
Investigating the Efficacy of Kidney-Protective Lactobacillus Mixture-Containing Pet Treats in Feline Chronic Kidney Disease and Its Possible Mechanism
Next Article in Special Issue
Open-Set Recognition of Individual Cows Based on Spatial Feature Transformation and Metric Learning
Previous Article in Journal
An Evaluation of the Impact of an OPEN Stewardship Generated Feedback Intervention on Antibiotic Prescribing among Primary Care Veterinarians in Canada and Israel
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

In Vivo Prediction of Breast Muscle Weight in Broiler Chickens Using X-ray Images Based on Deep Learning and Machine Learning

1
College of Information and Electrical Engineering, China Agricultural University, Beijing 100083, China
2
Scientific Research Base for Integrated Technologies of Precision Agriculture (Animal Husbandry), The Ministry of Agriculture, Beijing 100083, China
3
College of Animal Science and Technology, China Agricultural University, Beijing 100193, China
*
Authors to whom correspondence should be addressed.
Animals 2024, 14(4), 628; https://doi.org/10.3390/ani14040628
Submission received: 15 January 2024 / Revised: 9 February 2024 / Accepted: 12 February 2024 / Published: 16 February 2024

Abstract

:

Simple Summary

The breast muscle weight of broilers is a key indicator in poultry production. The accurate and nondestructive measurement of broiler breast muscle weight can improve breeding and the precision management level of broiler farms. Therefore, this study proposed an efficient method for predicting the breast muscle weight of broilers in vivo which can automatically predict broiler breast muscle weight. The experimental results demonstrate the method’s accuracy and superiority. The proposed method streamlines the data collection process, improves measurement efficiency, and provides crucial data support for broiler breeding and precision management.

Abstract

Accurately estimating the breast muscle weight of broilers is important for poultry production. However, existing related methods are plagued by cumbersome processes and limited automation. To address these issues, this study proposed an efficient method for predicting the breast muscle weight of broilers. First, because existing deep learning models struggle to strike a balance between accuracy and memory consumption, this study designed a multistage attention enhancement fusion segmentation network (MAEFNet) to automatically acquire pectoral muscle mask images from X-ray images. MAEFNet employs the pruned MobileNetV3 as the encoder to efficiently capture features and adopts a novel decoder to enhance and fuse the effective features at various stages. Next, the selected shape features were automatically extracted from the mask images. Finally, these features, including live weight, were input to the SVR (Support Vector Regression) model to predict breast muscle weight. MAEFNet achieved the highest intersection over union (96.35%) with the lowest parameter count (1.51 M) compared to the other segmentation models. The SVR model performed best (R2 = 0.8810) compared to the other prediction models in the five-fold cross-validation. The research findings can be applied to broiler production and breeding, reducing measurement costs, and enhancing breeding efficiency.

1. Introduction

Chicken breast has gained increasing popularity among consumers in the market due to its attributes of high protein content, low-fat calories, and affordable price [1,2,3]. Therefore, breeding broiler chickens with a high yield of breast muscle holds significant economic value, which requires accurate measurement of the breast muscle weight [4,5,6]. The conventional approach to broiler breeding infers the carcass performance of an individual to be cultivated by using the slaughter value from individuals within the same family [7,8]. However, such methods may consume a large amount of time, compromise animal welfare, and incur high economic costs [9]. Furthermore, this approach may be influenced by individual differences in the group and cannot retain the individuals with high breeding value among the slaughtered population [10]. Therefore, it is imperative to employ a nondestructive, accurate, and efficient broiler phenotypic estimation method to select individuals with elevated breeding values.
With the development of medical imaging technology, it is easy to nondestructively visualize the intricate internal structures of organisms with enhanced precision and resolution [11,12,13]. In early studies, researchers manually determined the cross-sectional area of the pectoral muscle in ultrasound images and combined it with live weight to construct a regression equation for predicting the breast muscle weight of broilers [9,14,15]. However, this manual method for determining the area is time-consuming and highly depends on the operator. Two recent studies on the breast muscle weight of broilers employed external body traits and breast muscle thickness determined by ultrasound images to develop machine learning models for breast muscle weight [7,8]. Nevertheless, the process of ultrasonic measurement is tedious, including feather separation, gel application, and probe positioning, which is not suitable for the automation of data collection. In contrast, X-ray technology can complete the scan within seconds after the broiler is simply fixed, without the need for additional operations. It is often employed for assessing the physiological condition of animal organisms [16,17,18,19]. In addition, many phenotypic traits of most organisms can also be measured or analyzed from X-ray images [20,21,22]. However, the level of automation in the aforementioned studies remains relatively low, as they still require human intervention for image analysis.
Recently, advancements in deep learning technology have provided a novel solution for medical image recognition [23]. Numerous deep learning-based medical image diagnostic assistance technologies have been developed to help physicians diagnose diseases and predict drug interactions [24,25,26,27,28]. Simultaneously, deep learning-based segmentation networks are widely utilized in the study of biological phenotypic traits. Melo et al. [29] replaced the tedious task of manually delimiting cattle rib-eye areas in ultrasound images using UNet++. Jeong and Sung [30] utilized the improved Attention U-Net to acquire mask images of the heart and thoracic vertebra from dog chest X-ray images, allowing for the automatic measurement of their length and area to calculate the adjusted heart volume index. Tempelaere et al. [31] employed U-Net to accurately classify every pixel in the X-ray images of pears, thereby obtaining the type, size, and location of internal fruit lesions. Chen et al. [32] developed MCC-Net to precisely segment the internal structure of rice seedling stems in X-ray scans and generate high-resolution images for phenotypic analysis. However, the running of deep learning models requires substantial storage and computational resources. The memory consumption of the aforementioned models is relatively high, making them less suitable for deployment on edge devices with limited computational resources in real farm environments. Hence, a lightweight model that preserves high segmentation accuracy will be more advantageous for practical application scenarios.
To tackle these problems, this study proposed a nondestructive, accurate, and efficient method for predicting the in vivo breast muscle weight of broilers using X-ray images and live weight. The method primarily comprises image segmentation, feature extraction, and weight prediction components. Firstly, a novel lightweight segmentation model called multistage attention enhancement fusion segmentation network (MAEFNet), was proposed for the automatic acquisition of pectoral muscle mask images from X-ray images of broilers. The encoder of the proposed model is the pruned MobileNetV3 [33], which is a lightweight convolutional neural network designed for mobile and embedded devices. In the decoder, two efficient attention modules, the Attention Refinement Module (ARM) [34] and the Coordinate Attention Module (CAM) [35], were introduced to enhance the model’s understanding of channel importance and spatial information, thus facilitating the forward propagation of effective features and excluding irrelevant features within the images. In addition, the Feature Fusion Module (FFM) [34] was integrated into the model for the fusion of features from different stages, which will improve the model’s capability to recognize target regions of varying sizes. Next, 15 related shape features were extracted from the pectoral muscle mask images obtained with MAEFNet. Meanwhile, inspired by the previous studies on breast muscle weight [7,8], live weight was also included as one of the features for breast muscle weight prediction. To further enhance the model’s accuracy and reliability, the recursive feature elimination (RFE) method was employed to screen these features. Finally, these selected features were inputted into the Support Vector Regression (SVR) model to complete the prediction of breast muscle weight. The proposed method enables the cost-effective and efficient in vivo prediction of breast muscle weight in broilers, providing crucial informational references for broiler breeding and precision management.
The principal objectives of this study were as follows: (1) to design a segmentation model with high accuracy and low memory consumption to automatically acquire pectoral muscle mask images of broiler chickens; (2) to select the optimal feature subset and validate the superiority of the SVR model for predicting breast muscle weight; and (3) to validate the feasibility of combining deep learning and machine learning methods with X-ray imaging for predicting in vivo pectoral muscle weight.

2. Materials and Methods

2.1. Overall Process

The overall process of the proposed method for breast muscle weight prediction is depicted in Figure 1. First, the X-ray images of the broilers were scanned, and then we weighed their live weight. Second, the X-ray images were preprocessed, including image cropping and scaling. Third, pectoral muscle mask images were obtained using MAEFNet. Fourth, relevant shape features were extracted from these mask images and integrated with live weight as inputs to the SVR model. Finally, the weight of the breast muscle was predicted.

2.2. Data Acquisition

The broilers used in this study were collected from a broiler farm in Guangze Country, Nanping City, Fujian Province. All broilers were reared in floor pens and freely administered water and food under the same recommended environmental and nutritional conditions (Feeding Standard of Chicken, China, NY/T 33-2004). The X-ray images were acquired with a customed mobile X-ray scanning device made by Aoying Testing Technology Co., Ltd. (Shanghai/Ningbo, China).
The broilers were positioned with their breasts directly facing the flat-panel detector (front-facing) during scanning. Concurrently, to increase the number of original images, data collection in the latter stages also included scans of broilers oriented with their backs facing the flat-panel detector. In addition, the scanning postures of broilers were checked before actually confirming the image. A total of 622 original X-ray images with a resolution of 1536 pixels × 1536 pixels were collected. These images were randomly divided into a training set, a validation set, and a test set in a ratio of 7:2:1. These images constituted the broiler X-ray breast muscle segmentation dataset and were used in the segmentation experiments. To annotate the broiler breast muscle region, the people involved in labeling learned the in vivo shape of the broiler breast muscle and the pixel distribution characteristics in the X-ray images. In addition, relevant professionals reviewed and modified the contents of the initial labeling to ensure the accuracy of the annotation. LabelMe software was employed to annotate the breast muscle regions in these 622 X-ray images. Figure 2 displays the annotation results of two broilers.
Data augmentation was performed using Albumentations [36] to enhance the robustness and generalization of the segmentation model. Figure 3 shows an example of augmented images.
In addition, a total of 100 front-facing acquisition broiler images were selected from these X-ray images. These images came from the validation and testing sets of the segmented dataset and corresponded one-to-one to actual broiler samples, meaning that 100 images represented 100 different broilers. After scanning these 100 broilers, their live weights were weighed, and they were subsequently dissected to obtain their actual breast muscle weights. These data were utilized in the breast muscle weight prediction experiments. All of the animal procedures were approved by the Experimental Animal Welfare and Animal Experiment Ethics Review Committee of China Agricultural University.

2.3. Data Preprocessing

Owing to the irrelevant background pixels within the original X-ray images and the high resolution of the primitive images, they would increase the recognition difficulty, memory consumption, and training time. Therefore, the X-ray images were cropped to remove irrelevant regions and scaled to a resolution of 512 pixels × 512 pixels.

2.4. MAEFNet Segmentation Model

To maintain excellent segmentation accuracy while reducing model memory consumption, a lightweight multistage attention enhancement fusion segmentation network (MAEFNet) was proposed for the automatic acquisition of pectoral muscle mask images. This network employs an encoder–decoder architecture and offers easy deployment and applicability in real-world environments. The pruned MobileNetV3 was employed as the encoder to efficiently capture features. Simultaneously, a novel decoder architecture was designed to achieve superior performance in semantic segmentation tasks. The detailed architecture of MAEFNet is depicted in Figure 4.

2.4.1. Encoder

MobileNetV3 [33] is a lightweight deep neural network architecture designed for efficient feature extraction from images. It utilizes the inverted residual structure as the core component to capture and process features while reducing the computational cost. In MAEFNet, to alleviate the loss of local information caused by frequent downsampling, the stride of the last convolutional downsampling in MobilNetV3 was modified, which resulted in a reduction in the downsampling ratio from 32 to 16. In addition, the structure after this layer was discarded, further reducing the memory consumption of the model.

2.4.2. Decoder

A novel decoder architecture was designed for the efficient enhancement and fusion of features at various stages of the encoder. The essential components of this decoder included the Attention Refinement Module (ARM) [34] for channel importance enhancement, the Coordinate Attention Module (CAM) [35] for spatial information enhancement, and the Feature Fusion Module (FFM) [34] for multistage features fusion. The design of these modules focused on achieving effective feature processing while maintaining a relatively low computational cost. The detailed structures of these three modules are shown in Figure 5.
Initially, feature maps at different stages in the encoder were compressed and refined by the ARM. This process adaptively recalibrated the importance of distinct channels within each feature map, thus enhancing the model’s sensitivity and discriminative capability towards critical information.
Specifically, first, a 3 × 3 convolutional transformation layer was applied to the input feature map X i n p u t , yielding a channel dimension-reduced feature map denoted as X C × H × W . This convolutional transformation layer was composed of convolution, batch normalization, and an activation function. Then, a global average pooling layer aggregated each channel of X into a vector to squeeze the global spatial information. Next, a 1 × 1 convolutional conversion layer was employed on the vector to derive the channel-wise attention weight denoted as α c C × 1 × 1 . This weight enabled the network to focus more on effective features without increasing excessive computational resources. Last, the final output X o u t p u t C × H × W was obtained as the product of X and α c . The detailed structure is depicted in Figure 5a and the computational process is shown in Equation (1).
X = σ ( B N ( f 3 × 3 ( X i n p u t ) ) ) α c = δ ( B N ( f 1 × 1 ( P A v g ( X ) ) ) ) X o u t p u t = X × α c
where f 3 × 3 denotes a convolution operator with a kernel size of 3 × 3, f 1 × 1 denotes a convolution operator with a kernel size of 1 × 1, BN denotes the batch normalization operator, σ is the ReLU activation function, δ is the Sigmoid activation function, and P a v g denotes a global average pooling operator.
Subsequently, the CAM, as an efficient attention mechanism, can enhance the spatial position information in feature maps with little computation, which is particularly suitable for embedding into a lightweight network to improve performance. It was introduced to compensate for the neglect of spatial details by the ARM.
Specifically, first, the input feature map, X i n p u t C × H × W , was globally averaged in the vertical and horizontal directions using two separate global average pooling layers, yielding two direction-aware feature maps. This step enabled the CAM to collect long-range dependencies in one spatial direction while preserving exact positioning information in the other direction, allowing the network to locate the regions of interest more precisely. Then, these two aggregated feature maps were concatenated in the spatial dimension and fed into a 1 × 1 convolutional conversion layer, generating an intermediate feature map F C / r × ( H + W ) × 1 . F encodes spatial information in two directions. Next, F was decomposed in the spatial dimension into two components, F h C / r × H × 1 and F w C / r × 1 × W . Afterward, each of these components underwent individual processing through two 1 × 1 convolutional conversion layers, restoring the number of channels to match the input. This operation yielded the vertical and horizontal coordinate attention weights, α H C × H × 1 and α W C × 1 × W , which can enable the model to focus more on the target areas. Last, these two attention weights were multiplied with the original input feature map, X i n p u t , to produce the final output, X o u t p u t C × H × W . The detailed structure is depicted in Figure 5b, and the computational process is shown in Equation (2).
F = ς h ( B N ( f 1 × 1 ( C o n c a t s ( P A v g H ( X i n p u t ) , P A v g W ( X i n p u t ) ) ) ) ) F h , F w = S p l i t ( F ) α H = σ ( f 1 × 1 ( F h ) α W = σ ( f 1 × 1 ( F w ) ) X o u t p u t = X i n p u t × α H × α W
where ς h is the Hard Swish activation function, P A v g H denotes the global average pooling operator with a kernel size of H × 1, P A v g W denotes the global average pooling operator with a kernel size of 1 × W, C o n c a t s denotes the concatenation operation along the spatial dimension, and S p l i t denotes the decomposition operation along the spatial dimension.
Finally, fusing feature maps from different stages allows the model to acquire multiscale information, which is crucial for recognizing objects of varying scales. Meanwhile, it can partially compensate for the feature loss during the downsampling process, thus improving the model’s performance. Therefore, the FFM was introduced to efficiently integrate features from different hierarchical levels, resulting in a more comprehensive feature representation.
Specifically, the process began by concatenating the input feature maps, X 1 C 1 × H × W and X 2 C 2 × H × W , along the channel dimension, yielding a fusion feature map, X ( C 1 + C 2 ) × H × W , that combines information from two stages. Then, a channel dimension-reduced feature map, X C × H × W , was generated through a 3 × 3 convolutional transformation layer. Next, a global average pooling operator and two 1 × 1 convolution operators, each followed by an activation function, were applied to produce the channel weight α c C × 1 × 1 , which fully collected the channel-wise dependencies. Last, the final fused feature map, X o u t p u t C × H × W , was obtained by adding the product of X and α c to the original X . The detailed structure is depicted in Figure 5c, and the computational process is shown in Equation (3).
X = σ ( B N ( f 3 × 3 ( C o n c a t c ( X 1 , X 2 ) ) ) ) α c = δ ( f 1 × 1 ( σ ( f 1 × 1 ( P A v g ( X ) ) ) ) ) X o u t p u t = X + X × α c
where C o n c a t c denotes the concatenation operation along the channel dimension.

2.4.3. Setup

MAEFNet was trained from scratch. The parameters for network training are outlined in Table 1.
The cross-entropy loss function was employed on the training set to guide the model’s iterative updates. Meanwhile, the dice loss function was used to calculate the model’s loss on the validation set for a more intuitive interpretation of segmentation performance. The equations for these loss functions are provided as follows:
C r o s s E n t r o p y L o s s = 1 N i = 1 N c = 1 C y i , c l o g ( p i , c )
D i c e L o s s = 1 2 | A B | | A | + | B |
where N represents the total number of samples in the dataset; C represents the number of classes; y i , c is a binary indicator of whether class c is the correct classification for the i-th sample, with values of either 0 or 1; and p i , c is the predicted probability that the i-th sample belongs to class c , with values in the range [0-1]. A and B represent the number of pixels in the predicted and real target areas, respectively.

2.5. Breast Muscle Weight Prediction Model

The pectoral muscle mask images can be automatically obtained through MAEFNet, but they cannot be directly used to predict breast muscle weight. As a result, the extraction of shape features from the mask images and the construction of a machine learning model are needed to achieve the prediction of breast muscle weight. In addition, live weight is included as one of the features in the breast muscle weight prediction model. The reasons for this are listed as follows: (1) it is frequently employed in research on prediction methods for broiler breast muscle weight [7,8,15]; (2) it is highly correlated with breast muscle weight in poultry [37]; and (3) it is relatively easy to obtain.
Inspired by research that utilized computer vision technology to estimate weight [38], a multitude of shape features were extracted from pectoral muscle mask images. Table 2 provides detailed descriptions of the extracted shape features and explains their calculation principles, where A and Ca are the number of pixels in the pectoral muscle region and the convex hull of the pectoral muscle region, respectively.
The initial features of the breast muscle weight prediction model include live weight and 15 extracted pectoral muscle mask image shape features. Directly applying these features may lead to a decrease in the accuracy and reliability of the prediction model. Selecting appropriate features not only reduces data dimensionality to save computational resources but also suppresses data noise to alleviate model overfitting. Recursive feature elimination (RFE) is a common feature selection method, typically used to enhance the model’s accuracy and interpretability. The fundamental idea behind RFE is to iteratively build models and eliminate the least important feature at each step until reaching the predefined stop criterion. The ranking criteria of MRE is the coefficient of determination for the SVR model. RFE is firstly used to filter out the optimal subset containing different numbers of features. Then, the highest result among them is selected, including live weight, area, height, width, and diameter.
SVR (Support Vector Regression) is a variant of the SVM (Support Vector Machine) for regression tasks, falling under the category of supervised machine learning algorithms. The fundamental concept behind SVR is to map the input data into a high-dimensional feature space using a kernel function [39]. Subsequently, within this space, a hyperplane is constructed to be as close as possible to all data points while maintaining a certain tolerance for errors between predicted and observed values. The SVR model exhibits robustness against outliers and noise, and its inherent regularization parameters mitigate the risk of overfitting, making it suitable for data of varying dimensions [40]. Compared to other machine learning models, this model demonstrates exceptional generalization, and it excels in adapting to small-sample data [41]. Meanwhile, some related studies have achieved promising experimental results using this model [7,42,43]. Furthermore, the SVR model with linear kernel function outperformed both polynomial and radial kernel functions in preliminary experimental results. Therefore, SVR with a linear kernel was selected for predicting the breast muscle weight. The other parameters of the SVR model were determined through a grid search method with 5-fold cross-validation. Specifically, the penalty parameter was set to 10 and epsilon was set to 0.3, while the other parameters were maintained at their default values.

2.6. Experimental Settings and Evaluation Metrics

All experiments were conducted on the same experimental platform. The specific configuration of the experimental platform is displayed in Table 3. PyTorch was used to construct the segmentation models. OpenCV was employed to preprocess the raw images and extract features from mask images. Scikit-learn was adopted to build the machine learning models.
In the segmentation experiments, the precision (Pre), recall (Rec), intersection over union (IoU), and Dice similarity coefficient (DSC) were employed to evaluate the accuracy of the model. Pre measures the accuracy of the model in predicting target regions. Recall assesses the model’s ability to capture all parts of the actual target areas. The IoU is the ratio of the intersection to the union between the predicted results and the actual labels, reflecting the model’s capability for accurate identification of target areas. The DSC quantifies the degree of overlap between the model’s predictions and the actual labels. Additionally, the memory consumption and speed of the model were assessed according to the number of parameters (Params) and the inference time for a single image (Latency). The equations for Pre, Rec, IoU, and DSC are defined as follows:
P r e = T P T P + F P × 100 %
R e c = T P T P + F N × 100 %
I o U = T P T P + F P + F N × 100 %
D S C = 2 T P 2 T P + F P + F N × 100 %
where TP is the number of pixels in the target area that were correctly predicted, FN is the number of pixels in the target area that were incorrectly predicted as the nontarget area, TN is the number of pixels in the nontarget area that were correctly predicted, and FP is the number of pixels in the nontarget area that were incorrectly predicted as the target area.
In the weight prediction experiments, the coefficient of determination (R2), root mean square error (RMSE), mean absolute error (MAE), and mean relative error (MRE) were used to evaluate the prediction models. R2 measures the degree of fit between the model’s predictions and the actual values. RMSE calculates the average magnitude of error between predicted and actual values. MAE calculates the average absolute difference between predicted and actual values. MRE measures the average relative difference between the predicted and actual values, expressed as a percentage. Their equations are defined as follows.
R 2 = 1 i = 1 N ( y ^ y i ) 2 i = 1 N ( y i y ¯ ) 2
R M S E = 1 N i = 1 N ( y ^ i y i ) 2
M A E = 1 N i = 1 N | y ^ i y i |
M R E = 1 N i = 1 N | y ^ i y i y i | × 100 %
where N denotes the total number of samples, y i denotes the actual weight for the i-th sample, y ^ i denotes the predicted weight for the i-th sample, and y ¯ denotes the average actual breast muscle weight of the broilers.

3. Results

3.1. Comparison of Segmentation Models

To better validate the overall performance of the proposed MAEFNet, this study compared the experimental results of this model with those of five other models, including UNet [44], TransUNet [45], DeepLabV3Plus [46], BiSeNet [34], and LR-ASPP [33]. The training parameters and loss function of the compared models were kept consistent with MAEFNet.
The results of the segmentation experiments are shown in Table 4. In terms of segmentation accuracy, MAEFNet achieved a Pre, IoU, Rec, and DSC of 97.89%, 96.35%, 98.40%, and 98.15%, respectively, all of which were higher than those of the other compared models. Compared to LR-ASPP, which used the same encoder, MAEFNet had an improvement of 1.01% in the IoU, indicating that the designed decoder can better enhance the model’s performance. The IoU increased by 0.65% compared to UNet, which had the second-highest accuracy. In terms of memory consumption, MAEFNet had the lowest number of parameters (1.51 M), decreasing by 53% and 87% compared to LR-ASPP and BiSeNet, respectively, and this was significantly lower than the other compared models. In terms of inference speed, a single image inference time of MAEFNet was approximately 4.75 milliseconds (ms), slightly longer than that of LR-ASPP (3.68 ms) and BiSeNet (4.01 ms) but significantly lower than other compared models.
The loss curves for different models on the training and validation sets are depicted in Figure 6. All models exhibited a gradual reduction in loss on the training set as the epoch increased, eventually reaching a stable level. For the validation set, MAEFNet converged faster compared to the other models. Furthermore, after the 10th epoch, MAEFNet exhibited lower loss values, which indicates that MAEFNet performs better on new images and possesses good generalization capabilities.
Figure 7 displays the visualized segmentation results of different models. LR-ASPP and DeepLabV3Plus exhibited rough predictions of the top boundary of the pectoral muscle, with some mis-segmentation issues in the bottom boundary. UNet, TransUNet, and BiSeNet performed slightly better but still had varying degrees of mis-segmentation at the top boundary. In contrast, MAEFNet can precisely identify the top and bottom boundaries of the pectoral muscle and predict a clear and complete contour. This demonstrates the model’s ability to effectively extract edge features from images, resulting in excellent segmentation results.
In summary, MAEFNet achieved precise segmentation of the breast muscle area in broiler X-ray images, with low memory consumption while maintaining an acceptable speed. These attributes significantly enhance the model’s adaptability and usability in real farm environments with constrained computational resources, providing robust support for its swift deployment and practical application.

3.2. Comparison of Weight Prediction Models

To adequately validate the superiority of the SVR model in predicting breast muscle weight, this study compared it with other prevalent machine learning models, including K-Nearest Neighbor (KNN), Decision Tree (DT), Random Forest (RF), Adaptive Boosting (AdaBoost), and Extreme Gradient Boosting (XGBoost). The hyperparameters of all of these machine learning models were determined through the grid search approach with 5-fold cross-validation.
The experimental results of the prediction of breast muscle weight in this study were the average of 5-fold cross-validation. Table 5 presents the results of different models before and after feature selection. It can be observed that the SVR model performed better than other models, regardless of whether feature selection was used. Notably, after the feature selection, all models exhibited varying degrees of optimization across all metrics, especially with SVR’s R2 showing an improvement of 0.0381. This amply illustrates the role of feature selection in enhancing model accuracy. Subsequent analysis is based on the feature selection.
The SVR model demonstrated both a satisfactory actual fitting degree and prediction error in predicting breast muscle weight, with an R2 of 0.8810, an RMSE of 25.38 g, an MAE of 20.48 g, and an MRE of 3.668%. Compared with KNN, DT, RF, ADBoost, and XGBoost, the SVR model achieved a higher R2 by 0.0919, 0.1271, 0.0401, 0.0452, and 0.0826, respectively. Similarly, in terms of the MAE indicator, the SVR model displayed reduced values of 6.25, 10.20, 3.72, 3.97, and 6.56 compared to these models, respectively.
Figure 8 shows the distribution of actual values and predicted values of breast muscle weight for different prediction models based on the same test set. The error between the actual value and the predicted value is the difference between the abscissa and ordinate of the sample point. The degree of clustering of the sample points near the 45-degree diagonal line (y = x) reflects the fitting performance of the model. It can be seen that most of the sample points in the SVR model are closer to this diagonal line, which illustrates the advantage of the SVR model in predicting breast muscle weight.
In general, in light of the results obtained, the SVR model demonstrated a commendable fitting effect and prediction error in translating the selected features into the actual breast muscle weight of broilers. This indicates that the combination of MAEFNet and SVR models can effectively predict in vivo breast muscle weight. The accuracy meets the requirements and expectations of practical farm applications.

4. Discussion

4.1. Discussion of Data Processing

In this study, cropping and scaling were employed to process the original images. Cropping allows the model to focus more on the breast muscle area in the image while scaling alleviates memory consumption and computational costs during training and prediction. These steps contribute to improving the accuracy and efficiency of the model. However, they may also lead to some loss of details in the images. Even so, given the constrained computational resources in actual farm environments and the ultimate prediction results, these losses are deemed acceptable.

4.2. Discussion of the Evaluation Metrics

In the context of segmenting breast muscles in X-ray images for shape feature extraction, all four metrics—Pre, Rec, IoU, and DSC—are important and complementary. Pre and Rec provide insights into the model’s accuracy at identifying breast muscle pixels and the completeness of muscle segmentation, respectively. The IoU and DSC offer a more comprehensive evaluation by measuring the overall agreement between the predicted segmentation and the ground truth, which directly influences the reliability of the subsequent shape feature extraction.
For the weight prediction, R2 is useful for assessing how well the model captures the variability in the breast muscle weight. However, it cannot directly reflect the actual error of the prediction model. The MAE offers a simple and intuitive measure of the average error, while the RMSE provides insight into the average error magnitude, with a higher penalty for large errors. The MRE provides a relative measure of a model’s prediction accuracy, ensuring that evaluation results are not influenced by the scale of measurement.
For a robust evaluation, it is best to consider all of these metrics together, because they offer a comprehensive evaluation of a model’s accuracy and reliability in different tasks, with each metric contributing unique and valuable insight into the model’s performance.

4.3. Discussion of MAEFNet

By comparing the results in Table 4, the proposed MAEFNet outperformed other models in both segmentation accuracy and parameter count. In addition, it can be observed in Figure 7 that MAEFNet exhibited a comprehensive recognition capability for pectoral muscles of different sizes, especially in the top or bottom boundary, which is challenging for other segmentation models. On the one hand, the accuracy of the segmentation model influences the reliability of feature extraction, subsequently impacting the precision of the subsequent prediction model. On the other hand, the quantity of parameters in the model directly influences its memory consumption. Models with lower memory consumption can be more easily deployed in edge devices with limited computational resources, thereby reducing the usage costs of the model. MAEFNet possesses the advantages of low memory consumption and high accuracy, making it more suitable for application in practical farms with constrained computational resources.
Additionally, it is worth noting that UNet has fewer parameters compared to DeepLabV3Plus and TransUNet, yet its actual inference time was longer. Similarly, the proposed MAEFNet, despite having fewer parameters, was slightly slower in inference time compared to BiSeNet and LR-ASPP. The model’s inference speed is influenced by various factors, such as memory access, parallelism, and the software and hardware environments [47]. MAEFNet discards the final layers of the MobileNetV3 encoder, a pivotal factor contributing to its model parameter count being less than that of the LR-ASPP. Simultaneously, the decoder incorporates many pooling, batch normalization, and activation operations that do not increase the model’s parameter count but consume computational time. In contrast, LR-ASPP has an optimized quantity of channels and filters to enhance the computational efficiency [33]. This targeted optimization is a crucial factor enabling LR-ASPP to maintain a lower inference time despite having a higher parameter count.
Table 6 shows the performance after integrating different backbone networks into the model. It is observed that, whether for ResNet50, VGG16, or Xception, the pruned MobileNetV3 outperforms these three backbone networks in various metrics. The IoU metric increases by 1.58%, 2.68%, and 1.71%, respectively, while the number of parameters decreases by 94.33%, 82.15%, and 93.62%. This improvement can be attributed to a series of architectural optimization methods employed in MobileNetV3 and the introduction of depthwise separable convolution and the SE attention mechanism in the inverted residual structure. Furthermore, the decoder in MAEFNet is specifically optimized for the pruned MobileNetV3, while such optimization may not be suitable for other backbone networks.
The rationale behind choosing these improvement strategies is to enable the model to achieve a better balance between memory consumption and accuracy. In addition, to better assess the contributions of various improvement strategies to the model, ablation experiments were conducted by incrementally incorporating strategies into the original network, which is founded on the raw MobileNetV3 encoder and multistage feature extraction. Table 7 presents the results of different combinations of strategies, where PMobileNetV3 denotes MobileNetV3 with a modified downsampling rate. Before adjusting the downsampling rate in the encoder, the model’s IoU and DSC were 95.68% and 97.80%, respectively. Following the modification of the downsampling rate, these metrics improved by 0.15% and 0.09%, respectively. Upon the integration of the ARM and CAM, reinforcing crucial information in the model’s feature maps, the IoU and DSC improved by 0.34% and 0.18%, respectively. Lastly, with the inclusion of the FFM optimizing the feature fusion process, the model exhibited a notable enhancement of 0.67% and 0.35% in the two metrics compared to the initial values, reaching 96.36% and 98.15%.
Figure 9 shows the IoU scores of MAEFNet after training with different batch sizes and learning rates. It can be observed that the optimal results were achieved when the batch size was set to 8, and the learning rate is set to 0.01. In other cases, fixing one parameter while varying the other, whether increasing or decreasing the value of the parameter, caused a decrease in the IoU.
To better validate the generalizability of MAEFNet, a publicly available dataset of human lung X-ray images was introduced [48,49]. Table 8 presents the results of the different models based on this dataset. It can be observed that, although MAEFNet exhibited relatively poor performance regarding the Rec metric, it achieved comprehensive superiority in the Pre, IoU, and DSC metrics. Specifically, the IoU showed an improvement of 0.2%, 0.61%, 1.54%, 1.58%, and 2.58% compared to UNet, TransUNet, DeepLabV3Plus, BiSeNet, and LR-ASPP, respectively. A high Pre and low Rec suggest that the model has high confidence in lung region determination but tends to be conservative, resulting in a decrease in false positives and an increase in false negatives. A higher IoU and DSC indicate better consistency between the segmentation results predicted by MAEFNet and the actual results.
To further improve the performance and reduce computational costs, model compression methods can be considered. These methods include pruning, quantization, and knowledge distillation. Simultaneously, the inference speed of the model can be optimized by judiciously reducing the number of filters and channels in certain structures of MAEFNet. Additionally, semi-supervised learning can leverage both labeled and unlabeled data to enhance model training efficiency and performance, possibly through iterative training using self-training or pseudo-labeling methods. Unsupervised learning, on the other hand, can uncover hidden patterns or features within the data that may improve the understanding of the underlying processes and potentially boost model performance. Possible avenues include the use of autoencoders for feature extraction or generative adversarial networks (GANs) for data augmentation. These techniques achieve commendable segmentation performance with minimal manual annotations. The segmentation dataset labels in this study are manually annotated, and these techniques can be integrated to further enhance the overall efficiency of the proposed method.

4.4. Discussion of Selected Features

In this study, the features selected by RFE include live weight, area, height, width, and diameter. Live weight represents the overall body weight of broilers in a live state. Area directly reflects the size of the breast muscle. They are highly correlated with breast muscle weight and have been used in previous research [9,15]. The length and width of the breast muscle in vivo are similar to the extracted height and width from X-ray images, and their correlation with breast muscle weight has been validated in a previous study [5]. As for diameter, it represents the maximum distance from the center point of the breast muscle to the farthest edge, reflecting the shape of the breast muscle from another perspective. The combination of these features better captures the overall size and shape characteristics of the breast muscle.

4.5. Discussion of Weight Prediction Models

As shown in Table 5, the SVR model achieved a better fitting effect and lower prediction error compared to other models. This may be ascribed to the use of an epsilon-insensitive loss function, allowing the model to ignore errors within a small range and focus more on structural errors rather than random noise [50]. The poorer performance of the XGBoost model based on the test set may be attributed to its better suitability for large-scale datasets, which can entail a risk of overfitting or becoming trapped in local optima when applied to small datasets [51].
From Figure 8, it can be observed that the SVR model is more accurate in predicting the weight of samples within the range of 400–750 g. This may be attributed to the higher consistency in the developmental characteristics of the breast muscle within this weight range. Notably, all models exhibit similar prediction biases for samples with weights below 400 g and above 750 g. This may be caused by the imbalanced breast muscle weight distribution within the dataset [52]. Figure 10 illustrates the distribution range of breast muscle weights for the 100 broilers used in the weight prediction experiments. Samples weighing below 400 g and above 750 g were notably scarce, while the number of samples in the middle range was considerably higher. Therefore, the training set naturally contained more data belonging to the middle range, leading the models to become biased toward the middle range values when predicting these lighter or heavier breast muscle weight samples [53]. Furthermore, the developmental characteristics of breast muscle appearance may differ for samples with excessively high or low weights, and the limited quantity of samples hinders the model from adequately learning relevant features. Therefore, appropriately increasing the number of samples below 400 g and above 750 g would improve the performance of the prediction model.

4.6. Discussion of the Proposed Method

The proposed method may be the first attempt to predict the breast muscle weight of live broilers using conventional two-dimensional X-ray images. The X-rays generated by the machine with a kinetic energy of up to 5 MeV do not induce radioactivity in the exposed food [54]. Therefore, the use of X-ray detection for broiler chickens does not pose food safety concerns. Compared with the traditional measurement method, this method eliminates the need for slaughtering broilers, saving a lot of time and the cost of broilers and manpower required for slaughtering. In addition, the memory consumption of the proposed MAEFNet is very low, which further reduces the deployment cost of the model.
Previous studies have typically used computed tomography (CT), dual-energy X-ray absorptiometry (DEXA), and ultrasound techniques to assess the body composition of live broilers [7,8,9,15,55]. CT and DEXA provide more features for the target region, which may lead to more accurate prediction results. However, the machines are costly, the scanning process is time-consuming, and anesthesia is typically necessary before the broilers are scanned, which is not suitable for practical production. The advantages of ultrasound technology lie in its relatively low cost and absence of radiation exposure. Yet, the measurement process is cumbersome, including feather removal, gel application, and precise probe positioning. In addition, outcomes are susceptible to the influence of different operators [15]. Moreover, such methods frequently necessitate some additional body measurements [7,8], which can also be a time-consuming and laborious process.
In contrast, the X-ray technology used in this study can complete the acquisition of raw image data through simple fixation. It does not require anesthesia or other preparatory procedures, and the operation’s process is also simpler and faster. Compared to other medical imaging approaches, X-ray technology is more convenient, which greatly simplifies the acquisition process of raw data and improves the overall efficiency of measurement. Certainly, this technology has limitations, such as radiation exposure or image artifacts. Additionally, conventional X-ray technology may not be able to clearly observe certain visceral tissues within the animal’s body.
The time consumed for each step of our method is shown in Table 9. The first and second steps both contain the time for fixing broilers. The actual X-ray scanning duration for a single broiler is approximately 2 seconds (s). The fifth step contains the time for feature extraction. It can be seen that the breast muscle weight prediction of a single broiler can be completed in around 30 s. This is credited to the convenience of X-ray scanning and the autonomous learning capability of deep learning methods on the features of raw image data. Moreover, the majority of the time is taken up by the first two steps. In the future, the integration of a weight measurement device into the image acquisition equipment may be considered to shorten the overall time.
The samples in this study were randomly selected from broilers of different genders in the same broiler farm. Their weight distribution was close to a normal distribution. This is a common occurrence under natural conditions, indicating no deliberate selection bias. However, it should be acknowledged that the samples are all from the same farm, which may introduce potential biases.
The limitation of this study lies in solely verifying the effectiveness of the proposed method in predicting the weight of breast muscle in broilers. This may result in the selected five features being more suitable for predicting breast muscle weight. To better extend the proposed method to other slaughter traits in broilers, such as wings and legs, a similar roadmap to this study will be adopted. Firstly, MAEFNet will be utilized to obtain masks corresponding to different traits. Then, RFE will be employed to determine the optimal feature subsets for various slaughter traits. Finally, the SVR model will be used for weight prediction. In addition, future work also includes further training and validating the models involved in the method in a wider variety and quantity of poultry, which will further improve the generalizability of the proposed method.
Despite these shortfalls, the convenience and efficiency of the proposed method render it more practical for the large-scale nondestructive measurement of breast muscle weight in broilers, which is of great value for the breeding and precision management of broilers. In general, the proposed method strikes a favorable balance between cost, efficiency, and accuracy, catering to the needs of practical farm environments.

5. Conclusions

This study proposed a simple and effective method for the automated prediction of in vivo breast muscle weight in broilers from X-ray images. It integrates image segmentation, feature extraction, and weight prediction as its core components. Initially, a novel segmentation network, MAEFNet, was designed. This network achieved a satisfactory balance between accuracy and memory consumption (IoU = 96.35%, Params = 1.51 M) while preserving an acceptable inference speed (Latency = 4.75 ms). Subsequently, RFE was applied to refine features, including live weight and the extracted features from the mask image, effectively reducing prediction errors of breast muscle weight. Finally, the SVR model was adopted to transform the selected features into breast muscle weight. This method, ultimately, achieved an R2 of 0.8810, an RMSE of 25.38 g, an MAE of 20.48 g, and an MRE of 3.668% under five-fold cross-validation. The findings suggest that the integration of deep learning and machine learning techniques offers a viable approach for predicting the in vivo breast muscle weight of broilers through X-ray imagery. It can be applied to the production and breeding work in practical broiler farms.
The application of this method can assist farm caretakers In monitoring the nutritional status of breast muscle in a large-scale broiler population in less time, enabling the formulation of more precision management strategies over time. Furthermore, by accurately and swiftly measuring breast muscle weight, farm breeders can more easily cultivate individuals with superior breast muscle characteristics at reduced costs, thereby accelerating genetic gains and promoting the sustainable development of the poultry industry.

Author Contributions

Conceptualization, R.Z. and R.S.; methodology, R.Z. and K.Y.; software, R.Z. and J.L.; supervision, R.S.; validation, R.Z.; formal analysis, R.Z. and J.L.; data curation, R.Z., J.Y., and K.Y.; visualization, R.Z., J.Y., and J.L.; writing—original draft, R.Z.; writing—review and editing, R.Z., R.S., and K.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China (grant number: 2021YFD1300101) and the National Key R&D Program of China (grant number: 2021ZD0113805).

Institutional Review Board Statement

The animal study protocol was approved by the Experimental Animal Welfare and Animal Experiment Ethics Review Committee of China Agricultural University (protocol code: AW52903202-3-1).

Informed Consent Statement

Not applicable.

Data Availability Statement

The code can be requested from the corresponding authors. The data are not publicly available due to being part of an ongoing study and privacy.

Acknowledgments

The authors thank all editors and reviewers for their suggestions and comments.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Fu, L.; Wang, A.; Zhang, H.; Zhou, Q.; Chen, F.; Su, W.; Yu, A.; Ji, Z.; Liu, Q. Analysis of Chicken Breast Meat Freshness with an Electrochemical Approach. J. Electroanal. Chem. 2019, 855, 113622. [Google Scholar] [CrossRef]
  2. Hailemariam, A.; Esatu, W.; Abegaz, S.; Urge, M.; Assefa, G.; Dessie, T. Effect of Genotype and Sex on Breast Meat Quality Characteristics of Different Chickens. J. Agric. Food Res. 2022, 10, 100423. [Google Scholar] [CrossRef]
  3. Soro, A.B.; Botinestean, C.; Shokri, S.; Juge, A.; Hannon, S.; Whyte, P.; Bolton, D.J.; Bourke, P.; Poojary, M.M.; Tiwari, B.K. Comparison of the Impact of UV-Light Emitting Diode and UV Lamp at Pilot-Plant Scale Level on Quality Parameters and Consumer Perception of Fresh Chicken Meat. Food Chem. 2024, 434, 137397. [Google Scholar] [CrossRef] [PubMed]
  4. Van Loo, E.J.; Caputo, V.; Nayga, R.M.; Meullenet, J.-F.; Ricke, S.C. Consumers’ Willingness to Pay for Organic Chicken Breast: Evidence from Choice Experiment. Food Qual. Prefer. 2011, 22, 603–613. [Google Scholar] [CrossRef]
  5. Erensoy, K.; Noubandiguim, M.; Cilavdaroglu, E.; Sarica, M.; Yamak, U.S. Correlations between Breast Yield and Morphometric Traits in Broiler Pure Lines. Braz. J. Poult. Sci. 2020, 22, 1–8. [Google Scholar] [CrossRef]
  6. Velleman, S.G. Broiler Breast Muscle Myopathies: Association with Satellite Cells. Poult. Sci. 2023, 102, 102917. [Google Scholar] [CrossRef] [PubMed]
  7. Chen, J.-T.; He, P.-G.; Jiang, J.-S.; Yang, Y.-F.; Wang, S.-Y.; Pan, C.-H.; Zeng, L.; He, Y.-F.; Chen, Z.-H.; Lin, H.-J.; et al. In Vivo Prediction of Abdominal Fat and Breast Muscle in Broiler Chicken Using Live Body Measurements Based on Machine Learning. Poult. Sci. 2023, 102, 102239. [Google Scholar] [CrossRef]
  8. Li, Z.; Zheng, J.; An, B.; Ma, X.; Ying, F.; Kong, F.; Wen, J.; Zhao, G. Several Models Combined with Ultrasound Techniques to Predict Breast Muscle Weight in Broilers. Poult. Sci. 2023, 102, 102911. [Google Scholar] [CrossRef] [PubMed]
  9. Remignon, H.; Seigneurin, F.; Moati, F. In Vivo Assessment of the Quantity of Breast Muscle by Sonography in Broilers. Meat Sci. 2000, 56, 133–138. [Google Scholar] [CrossRef]
  10. Pewan, S.B.; Otto, J.R.; Huerlimann, R.; Budd, A.M.; Mwangi, F.W.; Edmunds, R.C.; Holman, B.W.B.; Henry, M.L.E.; Kinobe, R.T.; Adegboye, O.A.; et al. Next Generation Sequencing of Single Nucleotide Polymorphic DNA-Markers in Selecting for Intramuscular Fat, Fat Melting Point, Omega-3 Long-Chain Polyunsaturated Fatty Acids and Meat Eating Quality in Tattykeel Australian White MARGRA Lamb. Foods 2021, 10, 2288. [Google Scholar] [CrossRef]
  11. Franchi, M.V.; Raiteri, B.J.; Longo, S.; Sinha, S.; Narici, M.V.; Csapo, R. Muscle Architecture Assessment: Strengths, Shortcomings and New Frontiers of in Vivo Imaging Techniques. Ultrasound Med. Biol. 2018, 44, 2492–2504. [Google Scholar] [CrossRef] [PubMed]
  12. Yadav, P.; Mandal, C.C. Bioimaging: Usefulness in Modern Day Research. In Practical Approach to Mammalian Cell and Organ Culture; Mukherjee, T.K., Malik, P., Mukherjee, S., Eds.; Springer Nature: Singapore, 2023; pp. 1205–1230. ISBN 978-981-19173-1-8. [Google Scholar]
  13. Li, J.; Jiang, P.; An, Q.; Wang, G.-G.; Kong, H.-F. Medical Image Identification Methods: A Review. Comput. Biol. Med. 2024, 169, 107777. [Google Scholar] [CrossRef] [PubMed]
  14. Silva, S.R.; Pinheiro, V.M.; Guedes, C.M.; Mourão, J.L. Prediction of Carcass and Breast Weights and Yields in Broiler Chickens Using Breast Volume Determined in Vivo by Real-Time Ultrasonic Measurement. Br. Poult. Sci. 2006, 47, 694–699. [Google Scholar] [CrossRef] [PubMed]
  15. Oviedo-Rondón, E.O.; Parker, J.; Clemente-Hernández, S. Application of Real-Time Ultrasound Technology to Estimate in Vivo Breast Muscle Weight of Broiler Chickens. Br. Poult. Sci. 2007, 48, 154–161. [Google Scholar] [CrossRef] [PubMed]
  16. Yaroshenko, A.; Hellbach, K.; Yildirim, A.Ö.; Conlon, T.M.; Fernandez, I.E.; Bech, M.; Velroyen, A.; Meinel, F.G.; Auweter, S.; Reiser, M.; et al. Improved In Vivo Assessment of Pulmonary Fibrosis in Mice Using X-ray Dark-Field Radiography. Sci. Rep. 2015, 5, 17492. [Google Scholar] [CrossRef] [PubMed]
  17. Hümmelchen, H.; Wagner, H.; Brügemann, K.; Wenisch, S.; König, S.; Wehrend, A. Frequency and Characterisation of Anomalies and Fractures of the Caudal Spine in Sheep with Undocked Tails. Animals 2023, 13, 1419. [Google Scholar] [CrossRef] [PubMed]
  18. Pereira, K.H.N.P.; Fuchs, K.d.M.; Terçariol, L.A.A.; Silva, R.C.; Camargo, G.d.A.; Mendonça, J.C.; Paulino, N.T.; Zone, M.A.; Oba, E.; Lourenço, M.L.G. Two Types of Management for the Noninvasive Treatment of Pectus Excavatum in Neonatal Puppies—Case Reports. Animals 2023, 13, 906. [Google Scholar] [CrossRef] [PubMed]
  19. Zheng, J.; Liu, D.; Zhang, G.; Ding, J.; He, Z.; Li, Q.; Liu, R.; Wen, J.; Zhao, G. Improve the Accuracy of Chicken Leg Disease Assessment by Combination of Gait Score and X-ray Scan. J. Appl. Poult. Res. 2023, 32, 100339. [Google Scholar] [CrossRef]
  20. Zhao, J.H.; Liu, M.H.; Shen, J.; Tu, D.C.; Yu, F. Study on Mass and Thickness Measurement of Chicken Meat Using X-ray Image. Appl. Mech. Mater. 2011, 39, 182–185. [Google Scholar] [CrossRef]
  21. Ide, Y.; Nakahara, T.; Nasu, M.; Matsunaga, S.; Iwanaga, T.; Tominaga, N.; Tamaki, Y. Postnatal Mandibular Cheek Tooth Development in the Miniature Pig Based on Two-Dimensional and Three-Dimensional X-ray Analyses. Anat. Rec. 2013, 296, 1247–1254. [Google Scholar] [CrossRef]
  22. Senol, D.; Bodur, F.; Secgın, Y.; Bakıcı, R.S.; Sahin, N.E.; Toy, S.; Öner, S.; Oner, Z. Sex Prediction with Morphometric Measurements of First and Fifth Metatarsal and Phalanx Obtained from X-ray Images by Using Machine Learning Algorithms. Folia. Morphol. 2023, 82, 704–711. [Google Scholar] [CrossRef]
  23. Pan, X.; Zhu, J.; Tai, W.; Fu, Y. An Automated Method to Quantify the Composition of Live Pigs Based on Computed Tomography Segmentation Using Deep Neural Networks. Comput. Electron. Agric. 2021, 183, 105987. [Google Scholar] [CrossRef]
  24. Kundu, R.; Das, R.; Geem, Z.W.; Han, G.-T.; Sarkar, R. Pneumonia Detection in Chest X-ray Images Using an Ensemble of Deep Learning Models. PLoS ONE 2021, 16, e0256630. [Google Scholar] [CrossRef]
  25. El Asnaoui, K.; Chawki, Y. Using X-ray Images and Deep Learning for Automated Detection of Coronavirus Disease. J. Biomol. Struct. Dyn. 2021, 39, 3615–3626. [Google Scholar] [CrossRef]
  26. Hardalaç, F.; Uysal, F.; Peker, O.; Çiçeklidağ, M.; Tolunay, T.; Tokgöz, N.; Kutbay, U.; Demirciler, B.; Mert, F. Fracture Detection in Wrist X-ray Images Using Deep Learning-Based Object Detection Models. Sensors 2022, 22, 1285. [Google Scholar] [CrossRef]
  27. Tran, T.-O.; Vo, T.H.; Le, N.Q.K. Omics-Based Deep Learning Approaches for Lung Cancer Decision-Making and Therapeutics Development. Brief Funct Genom. 2023, elad031. [Google Scholar] [CrossRef] [PubMed]
  28. Le, N.Q.K. Predicting Emerging Drug Interactions Using GNNs. Nat. Comput. Sci. 2023, 3, 1007–1008. [Google Scholar] [CrossRef] [PubMed]
  29. de Melo, M.J.; Goncalves, D.N.; Bonin Gomes, M.d.N.; Faria, G.; Silva, J.d.A.; Marques Ramos, A.P.; Osco, L.P.; Garcia Furuya, M.T.; Marcato Junior, J.; Goncalves, W.N. Automatic Segmentation of Cattle Rib-Eye Area in Ultrasound Images Using the UNet plus plus Deep Neural Network. Comput. Electron. Agric. 2022, 195, 106818. [Google Scholar] [CrossRef]
  30. Jeong, Y.; Sung, J. An Automated Deep Learning Method and Novel Cardiac Index to Detect Canine Cardiomegaly from Simple Radiography. Sci. Rep. 2022, 12, 14494. [Google Scholar] [CrossRef] [PubMed]
  31. Tempelaere, A.; Minh Phan, H.; Van De Looverbosch, T.; Verboven, P.; Nicolai, B. Non-Destructive Internal Disorder Segmentation in Pear Fruit by X-ray Radiography and AI. Comput. Electron. Agric. 2023, 212, 108142. [Google Scholar] [CrossRef]
  32. Chen, M.; Liao, J.; Zhu, D.; Zhou, H.; Zou, Y.; Zhang, S.; Liu, L. MCC-Net: A Class Attention-Enhanced Multi-Scale Model for Internal Structure Segmentation of Rice Seedling Stem. Comput. Electron. Agric. 2023, 207, 107717. [Google Scholar] [CrossRef]
  33. Howard, A.; Sandler, M.; Chen, B.; Wang, W.; Chen, L.-C.; Tan, M.; Chu, G.; Vasudevan, V.; Zhu, Y.; Pang, R.; et al. Searching for MobileNetV3. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
  34. Yu, C.; Wang, J.; Peng, C.; Gao, C.; Yu, G.; Sang, N. BiSeNet: Bilateral Segmentation Network for Real-Time Semantic Segmentation. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; Springer International Publishing: Cham, Switzerland, 2018; pp. 334–349. [Google Scholar]
  35. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Los Alamitos, CA, USA, 20–25 June 2021; pp. 13708–13717. [Google Scholar]
  36. Buslaev, A.; Iglovikov, V.I.; Khvedchenya, E.; Parinov, A.; Druzhinin, M.; Kalinin, A.A. Albumentations: Fast and Flexible Image Augmentations. Information 2020, 11, 125. [Google Scholar] [CrossRef]
  37. Laudadio, V.; Tufarelli, V.; Dario, M.; D’Emilio, F.P.; Vicenti, A. Growth Performance and Carcass Characteristics of Female Turkeys as Affected by Feeding Programs. Poult. Sci. 2009, 88, 805–810. [Google Scholar] [CrossRef] [PubMed]
  38. Zhang, L.; Wang, J.; Duan, Q. Estimation for Fish Mass Using Image Analysis and Neural Network. Comput. Electron. Agric. 2020, 173, 105439. [Google Scholar] [CrossRef]
  39. Ma, X.; Zhang, Y.; Wang, Y. Performance Evaluation of Kernel Functions Based on Grid Search for Support Vector Regression. In Proceedings of the 2015 IEEE 7th International Conference on Cybernetics and Intelligent Systems (CIS) and IEEE Conference on Robotics, Automation and Mechatronics (RAM), Siem Reap, Cambodia, 15–17 July 2015; pp. 283–288. [Google Scholar]
  40. Awad, M.; Khanna, R. Support Vector Regression. In Efficient Learning Machines: Theories, Concepts, and Applications for Engineers and System Designers; Awad, M., Khanna, R., Eds.; Apress: Berkeley, CA, USA, 2015; pp. 67–80. ISBN 978-1-4302-5990-9. [Google Scholar]
  41. Meng, M.; Zhao, C. Application of Support Vector Machines to a Small-Sample Prediction. Adv. Petrole. Explor. Dev. 2015, 10, 72–75. [Google Scholar]
  42. Faridi, A.; Sakomura, N.K.; Golian, A.; Marcato, S.M. Predicting Body and Carcass Characteristics of 2 Broiler Chicken Strains Using Support Vector Regression and Neural Network Models. Poult. Sci. 2012, 91, 3286–3294. [Google Scholar] [CrossRef] [PubMed]
  43. Fan, N.; Liu, G.; Zhang, C.; Zhang, J.; Yu, J.; Sun, Y. Predictability of Carcass Traits in Live Tan Sheep by Real-Time Ultrasound Technology with Least-Squares Support Vector Machines. Anim. Sci. J. 2022, 93, e13733. [Google Scholar] [CrossRef]
  44. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2015; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  45. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. TransUNet: Transformers Make Strong Encoders for Medical Image Segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar] [CrossRef]
  46. Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; Springer International Publishing: Cham, Switzerland, 2018; pp. 833–851. [Google Scholar]
  47. Canziani, A.; Paszke, A.; Culurciello, E. An Analysis of Deep Neural Network Models for Practical Applications. arXiv 2017, arXiv:1605.07678. [Google Scholar] [CrossRef]
  48. Jaeger, S.; Karargyris, A.; Candemir, S.; Folio, L.; Siegelman, J.; Callaghan, F.; Xue, Z.; Palaniappan, K.; Singh, R.K.; Antani, S.; et al. Automatic Tuberculosis Screening Using Chest Radiographs. IEEE Trans. Med. Imaging 2014, 33, 233–245. [Google Scholar] [CrossRef] [PubMed]
  49. Candemir, S.; Jaeger, S.; Palaniappan, K.; Musco, J.P.; Singh, R.K.; Xue, Z.; Karargyris, A.; Antani, S.; Thoma, G.; McDonald, C.J. Lung Segmentation in Chest Radiographs Using Anatomical Atlases with Nonrigid Registration. IEEE Trans. Med. Imaging 2014, 33, 577–590. [Google Scholar] [CrossRef]
  50. Sotiropoulou, K.F.; Vavatsikos, A.P.; Botsaris, P.N. A Hybrid AHP-PROMETHEE II Onshore Wind Farms Multicriteria Suitability Analysis Using kNN and SVM Regression Models in Northeastern Greece. Renew. Energy 2024, 221, 119795. [Google Scholar] [CrossRef]
  51. Huiping, J.; Yaning, C.; Gonghuan, F.; Zhi, L.; Weili, D.; Qifei, Z. Adaptability of machine learning methods and hydrological models to discharge simulations in datasparse glaciated watersheds. J. Arid Land 2021, 13, 549–567. [Google Scholar] [CrossRef]
  52. Munguía Mondragón, J.C.; Rendón Lara, E.; Alejo Eleuterio, R.; Granda Gutirrez, E.E.; Del Razo López, F. Density-Based Clustering to Deal with Highly Imbalanced Data in Multi-Class Problems. Mathematics 2023, 11, 4008. [Google Scholar] [CrossRef]
  53. Alam, T.; Ahmed, C.F.; Zahin, S.A.; Khan, M.A.H.; Islam, M.T. An Effective Recursive Technique for Multi-Class Classification and Regression for Imbalanced Data. IEEE Access 2019, 7, 127615–127630. [Google Scholar] [CrossRef]
  54. Farkas, J.; Mohácsi-Farkas, C. History and Future of Food Irradiation. Trends Food Sci. Technol. 2011, 22, 121–126. [Google Scholar] [CrossRef]
  55. Schallier, S.; Li, C.; Lesuisse, J.; Janssens, G.P.J.; Everaert, N.; Buyse, J. Dual-Energy X-ray Absorptiometry Is a Reliable Non-Invasive Technique for Determining Whole Body Composition of Chickens. Poult. Sci. 2019, 98, 2652–2661. [Google Scholar] [CrossRef]
Figure 1. Overall process of the proposed method.
Figure 1. Overall process of the proposed method.
Animals 14 00628 g001
Figure 2. Annotated images of two broilers.
Figure 2. Annotated images of two broilers.
Animals 14 00628 g002
Figure 3. X-ray image augmented example. (a) Original; (b) Random rotation; (c) Vertical flip; (d) Horizontal flip after vertical flip; (e) Perspective transform; (f) Elastic transform after horizontal flip.
Figure 3. X-ray image augmented example. (a) Original; (b) Random rotation; (c) Vertical flip; (d) Horizontal flip after vertical flip; (e) Perspective transform; (f) Elastic transform after horizontal flip.
Animals 14 00628 g003
Figure 4. Architecture of MAEFNet. The thickness of the encoder block represents the spatial size, while the length indicates the number of channels. The thickness of the hollow arrow represents the magnification of upsampling.
Figure 4. Architecture of MAEFNet. The thickness of the encoder block represents the spatial size, while the length indicates the number of channels. The thickness of the hollow arrow represents the magnification of upsampling.
Animals 14 00628 g004
Figure 5. Architecture of (a) ARM; (b) CAM; and (c) FFM.
Figure 5. Architecture of (a) ARM; (b) CAM; and (c) FFM.
Animals 14 00628 g005
Figure 6. Training and validation loss curves of different segmentation models.
Figure 6. Training and validation loss curves of different segmentation models.
Animals 14 00628 g006
Figure 7. Visualized segmentation results of different segmentation models.
Figure 7. Visualized segmentation results of different segmentation models.
Animals 14 00628 g007
Figure 8. Scatter plots of predicted and actual values from different weight prediction models.
Figure 8. Scatter plots of predicted and actual values from different weight prediction models.
Animals 14 00628 g008
Figure 9. Comparison of the results with different hyperparameters.
Figure 9. Comparison of the results with different hyperparameters.
Animals 14 00628 g009
Figure 10. Distribution of broiler chicken breast muscle weights.
Figure 10. Distribution of broiler chicken breast muscle weights.
Animals 14 00628 g010
Table 1. The training parameters of MAEFNet.
Table 1. The training parameters of MAEFNet.
ParametersValue
Epochs100
Batch size8
Learning rate0.01
OptimizerSGD
Momentum0.9
Weight decay0.0001
Table 2. The list of shape features extracted from the pectoral muscle mask image.
Table 2. The list of shape features extracted from the pectoral muscle mask image.
FeatureDescriptionSymbol
AreaSquare root of the number of pixels in the pectoral muscle A r = A
Convex areaSquare root of the number of pixels in the convex hull of the pectoral muscle C o = C a
PerimeterNumber of pixels in the pectoral muscle boundaryP
Major axis lengthMajor axis length of an ellipse enclosing the pectoral muscleMal
Minor axis lengthMinor axis length of an ellipse enclosing the pectoral muscleMil
HeightHeight of the pectoral muscleH
WidthWidth of the pectoral muscleW
Rectangle areaSquare root of the product of the pectoral muscle height and widthRa = H W
EllipticityRatio of the major axis length to the minor axis lengthEl = Mal/Mil
DiameterDiameter of a circumscribed circle of the pectoral muscleD
Equivalent diameterDiameter of a circle equal to the area of the pectoral muscleEd = 2 A / π
Heywood circularity factorDegree to which the pectoral muscle is close to the circleHcf = P/(2 π A )
AspectRatio of the height to the width of the pectoral muscleAs = H/W
CurvatureRatio of the pectoral muscle area to the circumscribed circle areaCu = 4∗A/(πD2)
ComplexityRatio of a square of the perimeter to the pectoral muscle areaCl = P2/A
Table 3. The configuration of the experimental platform.
Table 3. The configuration of the experimental platform.
ConfigurationParameter
Operating systemWindows 10
CPUIntel Core i9-13900K
GPUNVIDIA GeForce RTX 4090
Development languagePython 3.10
FrameworkPyTorch 1.12 + OpenCV 4.8+ Scikit-learn 1.2
CUDA versionCUDA 12.1
Table 4. Performance of different segmentation models.
Table 4. Performance of different segmentation models.
ModelPre (%)IoU (%)Rec (%)DSC (%)Params (M)Latency (ms)
UNet97.6695.7097.9597.8417.2613.26
TransUNet97.4295.5298.0097.7377.3711.19
DeepLabV3Plus97.0995.4498.2697.6939.769.68
BiSeNet97.5795.5697.8997.7611.894.01
LR-ASPP97.3895.3497.8597.653.223.68
Ours97.8996.3598.4098.151.514.75
Table 5. Comparison of different prediction models.
Table 5. Comparison of different prediction models.
ModelWithout Feature SelectionWith Feature Selection
R2RMSEMAEMRE (%)R2RMSEMAEMRE (%)
KNN0.783335.1127.024.8880.789134.8426.734.738
DT0.737838.3831.485.6800.753937.5430.685.601
RF0.827631.3225.384.5620.840930.2224.204.320
ADBoost0.800333.8227.664.9680.835830.7524.454.374
XGBoost0.795434.2827.594.9000.798433.8527.044.794
SVR0.842929.6824.204.3230.881025.3820.483.668
Table 6. Performance after integrating different backbone networks.
Table 6. Performance after integrating different backbone networks.
BackbonePre (%)IoU (%)Rec (%)DSC (%)Params (M)Latency (ms)
ResNet5096.6594.7797.9897.3426.6513.77
VGG1695.8893.6797.6096.828.468.66
Xception96.8394.6497.6797.2923.665.92
Ours97.8996.3598.4098.151.514.75
Table 7. Performance of different combinations of improvement strategies.
Table 7. Performance of different combinations of improvement strategies.
MobileNetV3PMobileNetV3ARMCAMFFMIoU (%)DSC (%)
----95.6897.80
----95.8397.89
---96.0698.00
--96.1798.07
-96.3598.15
Table 8. Performance of different segmentation models on the new dataset.
Table 8. Performance of different segmentation models on the new dataset.
ModelPre (%)IoU (%)Rec (%)DSC (%)
UNet94.6991.4296.3695.40
TransUNet94.2691.0196.3595.15
DeepLabV3Plus93.4890.0896.1294.47
BiSeNet92.8790.0496.7394.61
LR-ASPP92.5689.0495.8994.01
Ours96.6091.6294.6795.53
Table 9. The detailed time consumption of the method.
Table 9. The detailed time consumption of the method.
StepTime
Scan X-ray imageapproximately 20 s
Get live weightapproximately 10 s
Preprocess image9.34 ms
Segment image4.75 ms
Predict breast muscle weight1.30 ms
All proceduresapproximately 30 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, R.; Li, J.; Yang, J.; Sun, R.; Yu, K. In Vivo Prediction of Breast Muscle Weight in Broiler Chickens Using X-ray Images Based on Deep Learning and Machine Learning. Animals 2024, 14, 628. https://doi.org/10.3390/ani14040628

AMA Style

Zhu R, Li J, Yang J, Sun R, Yu K. In Vivo Prediction of Breast Muscle Weight in Broiler Chickens Using X-ray Images Based on Deep Learning and Machine Learning. Animals. 2024; 14(4):628. https://doi.org/10.3390/ani14040628

Chicago/Turabian Style

Zhu, Rui, Jiayao Li, Junyan Yang, Ruizhi Sun, and Kun Yu. 2024. "In Vivo Prediction of Breast Muscle Weight in Broiler Chickens Using X-ray Images Based on Deep Learning and Machine Learning" Animals 14, no. 4: 628. https://doi.org/10.3390/ani14040628

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop