Next Article in Journal
Molecular Characterization and Functional Analysis of Hypoxia-Responsive Factor Prolyl Hydroxylase Domain 2 in Mandarin Fish (Siniperca chuatsi)
Next Article in Special Issue
Sheep Face Recognition Model Based on Deep Learning and Bilinear Feature Fusion
Previous Article in Journal
Developing a Preference Scale for a Bear: From “Bearly Like” to “Like Beary Much”
Previous Article in Special Issue
FedAAR: A Novel Federated Learning Framework for Animal Activity Recognition with Wearable Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Study on the Influence of PCA Pre-Treatment on Pig Face Identification with Random Forest

1
College of Information Science and Engineering, Shanxi Agricultural University, Jinzhong 030801, China
2
Science & Technology Information and Strategy Research Center of Shanxi, Taiyuan 030024, China
*
Author to whom correspondence should be addressed.
Animals 2023, 13(9), 1555; https://doi.org/10.3390/ani13091555
Submission received: 22 March 2023 / Revised: 30 April 2023 / Accepted: 4 May 2023 / Published: 6 May 2023

Abstract

:

Simple Summary

Pig face recognition plays an important role in the intelligent breeding and accurate management of pigs, and the mobile and embedded applications of this technology in the management of small and medium-sized pig farms are in great demand; therefore, in order to make the model more suitable for use in small and medium-sized pig farms, in this study, PCA pre-treatment was added to the traditional method. The experiment shows that the model is suitable for small and medium-sized pig farms, and it can promote the intelligent transformation of pig breeding management.

Abstract

To explore the application of a traditional machine learning model in the intelligent management of pigs, in this paper, the influence of PCA pre-treatment on pig face identification with RF is studied. By this testing method, the parameters of two testing schemes, one adopting RF alone and the other adopting RF + PCA, were determined to be 65 and 70, respectively. With individual identification tests carried out on 10 pigs, accuracy, recall, and f1-score were increased by 2.66, 2.76, and 2.81 percentage points, respectively. Except for the slight increase in training time, the test time was reduced to 75% of the old scheme, and the efficiency of the optimized scheme was greatly improved. It indicates that PCA pre-treatment positively improved the efficiency of individual pig identification with RF. Furthermore, it provides experimental support for the mobile terminals and the embedded application of RF classifiers.

1. Introduction

As a trend in the development of the pig breeding industry, intelligent breeding and precise management of pigs promote the research and development of relative technologies. During the process, research focus groups in the fields of individual identification, automatic inspection, detection of food and drinking water, gesture measuring, walking behavior detection, and abnormal behavior detection of pigs have been formed, among which individual identification and behavior identification of pigs are the basis for many studies. So far, the research techniques for individual identification and behavior identification of pigs can be classified into three schools. In the early stage, RFID technology [1] was mainly applied to study the efficiency of readers and construct the traceability system for the pig industry, which played a huge role in pig automatic feeding, pork quality management, slaughter supervision, and safety production. As new technologies were developed, the cost of RFID technology was still high, which is why its research and application declined. The second school focuses on the application and research of traditional machine learning models. Many scholars have studied the application of models such as SVM [2,3], KMEANS [4], KNN [5,6], LDA, RANDOMFOREST [7,8], and others and how they can be improved in the field of intelligent management of pigs. Substantial research achievements have been made in individual identification, huddling to keep warm, fast walking and slow walking behavior, statistical calculation of exercise amount, statistical calculation of sleep, and alike. The third school refers to the computer vision technology that mainly focuses on application and improvement of deep learning models. Scholars have used AlexNet, GoogleNet [9], VGG series, YOLO series [10,11,12,13,14], Transformer, and other models to carry out research on individual identification, feeding of food and drinking water, climbing and crossing behavior, and aggressive behavior. An enormous hashrate is needed by this school. At the moment, it is mostly used in laboratories. In the management of small and medium-sized pig farms, mobile terminals and embedded applications are in great need of intelligence technologies. To promote the intelligent development of small and medium-sized pig farms, it is necessary to provide technology and equipment that suits their intelligent management. The application of the first school is limited in the field, while the third school needs a large enough budget to support its hashrate and to have the equipment deployed, which means that it is difficult to spread in small and medium-sized pig farms. The second school, whose identification accuracy and operation efficiency can be better improved, becomes the first choice for the intelligent development of small and medium-sized pig farms due to its low cost as well as abundant research bases.
Therefore, this study explored the influence of PCA pre-treatment on pig face identification with RF. With RF + PCA pre-treatment, though identification accuracy was lost to some degree, training time and testing time decreased to 4.8% and 9% of the original values, respectively. Taking all these factors into consideration, after going through PCA pre-treatment, with the identification accuracy maintained in a range as a premise, the operating efficiency of the RF model improved greatly; thus, it is suitable for the intelligent management technologies to be applied in mobile terminals and embedded application in the management of small and medium-sized pig farms.
Some scholars designed a high-frequency radio frequency identification (HFRFID) system which was validated for its suitability to register individual pigs’ feeding patterns at a round trough in a group-housing context; this system showed good potential for measuring the feeding patterns of growing–finishing pigs in commercial pig houses [15]. Maselyne et al. [16] designed a high-frequency radio frequency identification (HFRFID) system to record the drinking behavior of individual pigs. The sensitivity of the system was 92%, the specificity was 93%, the precision was 90%, and the accuracy was 93%. It improves the productivity and economics of the swine industry as well as the health and welfare of the pigs. Kapun et al. [17] used the UHFRFID system to record the daily activity patterns of pigs. The sensitivity (true positive rate) of the UHFRFID system was about 80% at the feeding trough and playback equipment and about 60% at the drinker. The experimental results show that the system can record the time of pig visits and has a higher data density than video or direct observation. Maselyne et al. [18] proposed a method for constructing feeding visits based on the RFID registration of growing pigs in feeding tanks. The experimental results showed that when two tags were used for each pig, the average sensitivity of the method was 83%, the specificity was 98%, the precision was 97%, the accuracy was 75%, and it can automatically and accurately record the feeding information of growing pigs. Zhu Jun et al. [19] used RFID, intelligent control, network transmission, and other technologies to build a digital breeding pig breeding platform. The platform realizes the integration of automatic fine feeding, environmental control in the pig house, production management, and visual monitoring. Zhu Weixing et al. [20] designed an embedded pig behavior monitoring system based on RFID technology and ARM-Linux. The system is based on the ARM-LINUX platform, and the embedded device is installed in the pig breeding area to monitor the pigs and collect feeding data throughout the day. The experimental results show that the system has good real-time performance and stability. Chen T. Y. et al. [21] used RFID technology to collect pig diet information in the feed area of the pigsty. The experimental results showed that the sensitivity of this technology was 71.1%, the singularity was 87.1%, and the accuracy was 88.8%, realizing the remote monitoring of pig breeding and greatly reducing the labor cost for farmers. LASSO regression and a random forest model were used to predict the weight of pigs at 159 days to 166 days under four scenarios [22]; random forest and generalized linear regression were used to predict the physiological temperature of piglets. However, the prediction error was relatively high [23]; an auto-regression (AR) model and improved local linear embedding (LLE) were used to estimate pig weight in an actual farm environment [24], and SSD [25], F-rcnn [26], and other models were used to study the posture changes of pigs and then analyze their behavior. Win et al. [27] established an automatic pig size sorting system based on computer vision, which can realize personalized feeding according to pig size. Yan Hongwen [28] used Feature Pyramid Attention (FPA) combined with a Tiny-YOLO model to achieve multi-target detection of pigs in different scene groups, and Yan Hongwen [29], based on YOLOV3 as the basic model, introduced spatial attention and channel attention information to construct corresponding attention sub-models and realized the detection of different types of facial poses of group breeding pigs. Kim et al. [30] presented an algorithm that utilizes the major and minor axes of the pig detection box to associate the pig’s head with its corresponding body. They evaluated the detection performance of the YOLOv5 model with respect to the anchor box and demonstrated that the proposed method outperforms the previous method. Hu Zhiwei [31] used ResNet50 and ResNet101 as the backbone network to build a dual attention unit combining channel attention and spatial attention and used it in the feature pyramid network structure, achieving the goal of realizing instance detection of live pigs in different scenarios. Hansen et al. [32] combined the feature extraction results of the VGG-Face model [33] and Fisherfaces with the convolution network constructed by themselves to test a total of 1553 pictures of 10 individual pigs in the natural conditions of the farm and used Grad-CAM [34] activation-resembling mapping to distinguish the adhered pigs, achieving an identification precision rate of 96.7%.
With GPU applied, the high-precision advantage of computer vision technology has gradually been revealed. In the management of small and medium-sized farms, the need for mobile terminals and embedded apps has gradually been increasing. The deep learning model has a high requirement for hardware, and it is difficult to adapt to wide application, while the RFID technology is easy to simulate in the management [35,36] and requires the use of physical tags, such as ear cutting and ear tags, which can easily cause pain to pigs [37]. While it takes time and effort for workers to install tags, it also violates animal welfare breeding values; in addition, low-frequency RFID cannot receive signals from multiple pigs at the same time, and the identification area of high-frequency RFID is very small [38], so RFID technology has been gradually eliminated. Only the hardware requirements of the traditional machine learning model conform to the standards for mobile terminals and embedded apps, though there is still room for improvement in its identification accuracy and running time. PCA can extract the main features of pig faces [39,40], thus reducing the computation burden, improving operation efficiency, eliminating noise interference, and improving identification accuracy.
To promote the application of the traditional machine learning model in mobile terminals and embedded apps, RF is adopted for bettering the pig face identification efficiency in this study, and its influence on the efficiency of pig face identification is further studied by adding PCA pre-treatment, which provides experimental support for its application in both mobile terminals and embedded application. The terms used in this paper are listed in Table 1.

2. Materials and Methods

2.1. Sample Collection

The data in this study were collected two times. The first time, the data were collected in Dongsongjiazhuang Village, Jicun Town, Fenyang City, Shanxi Province, China (111°95′ E, 37°27′ N). In order to obtain live pig images of different pig house scenes, data were collected from 9:00 to 14:00 on June 1, 2019 (fine, strong light). Three pig farms were selected for video capture; each pig farm consisted of 10–30 pig pens, the number of pigs in each pen varied from 6 to 8, and the size of the pig pens was about 3.5 m × 2.5 m × 1 m. A total of 35 videos of 5 pens of breeding pigs aged 20 to 105 days were collected. The second time, data were collected from 10:30 to 12:00 on 13 October 2019 (cloudy, weak light), and the collection site was located in the Laboratory Animal Management Center of Shanxi Agricultural University, Taigu City, Shanxi Province, China (112°59′ E, 37°43′ N). A total of 15 pigs in 6 pens were selected for video collection. In this study, 10 pigs were selected as the research objects, as shown in Figure 1, including 768 training samples, 85 validation samples, and 250 test samples.
The computer used in the experiment is configured with 64-bit Windows system, Intel Core i7-6700, 8 GB memory, 6 GB video memory capacity, and program development uses Python V3.5 version language.

2.2. Principle of Pig Face Identification with RF

In order to ensure that the machine learning model finds the optimal result in each iteration, some random processes are added to most of the machine learning methods. Random forest (RF) [41] also adopts this design idea to construct a random decision tree. For each iteration, the algorithm usually generates the optimal predictive variable [42], with the basic idea of finding the average value of noise to construct sets of decision trees. Through a complex interaction tree, the decision tree in RF maps the complex input space to a simpler space. In reference [28], it is shown that sets of decision trees in RF are randomly trained, and RF reduces over-fitting through the effective use of data. Therefore, the method is an extension of the bagging classification tree, and it is a parallel learning process with fast training and high accuracy as its features. Its advantages are as follows [43]:
(1)
RF has both anti-over-fitting and anti-noise performances because random steps are included;
(2)
High-dimensional data can be processed;
(3)
Learning can be achieved in a parallel way;
(4)
The training time can be shortened.
With random selection methods adopted, RF performs better when there are many redundant features that cannot be distinguished [44,45]. In this study, RF was used for pig face identification. The complex pig face images were mapped to the category label space via the RF model. The steps for pig face identification with RF are as follows:
(1)
The n_tree training set is generated by sampling the pig face training samples for n_tree times, with m samples taken each time, for which the Bootstrapping method, a random sampling with replacement method, is used.
(2)
Every training set needs to train a decision tree model.
(3)
When splitting the decision tree according to the information gain or gini index, it is necessary to select an optimal feature among all the features.
(4)
Each decision tree is split in this way, and finally, all the training samples of this node are classified into the same category, and there is no pruning operation in this process.
(5)
In the end, multiple decision trees will be formed to generate the random forest. In the case of multi-classification tasks, the output of the random forest will be determined by voting.
In this study, 10 different pigs are classified with RF. When each decision tree selects the optimal feature, all the features in the attribute set are read and the feature with the minimum gini coefficient or maximum information gain is selected as the classification standard. Similarly to entropy, the gini coefficient here reflects the uncertainty of the data set. The process of determining is a process of entropy reduction. The gini coefficient is calculated according to Formula (1):
G i n i D = k C k D 1 C k D = 1 k C k D 2
where D represents data collection, [a];
k represents the category label, [dimensionless];
C k represents the kth sample subset, [dimensionless];
C k represents the number of samples included in the kth sample subset, [a].
When feature A is selected, the data set D is divided into subset D 1 and subset D 2 according to whether the value of feature A is a certain eigenvalue, as shown in the following formula:
D 1 = D | A = a
D 2 = D | A a
where a represents a certain characteristic value of characteristic A, [dimensionless];
D i represents the ith subset of data set D, [dimensionless].
In the case of feature A, the gini coefficient of data set D is calculated as shown in the formula below:
G i n i D , A = C k D G i n i D 1 + C k D G i n i D 2
If the formula listed above achieves the minimum value, then feature A is selected; information gain represents the change degree of entropy, i.e., the pre-classification information entropy minus the post-classification information entropy, as shown in Formula (5):
g D , A = H D H D | A
where H D represents the entropy of the data set D, [dimensionless];
H D | A represents the entropy of data set D divided by feature A, [dimensionless].
Each base learner classifies the pig samples according to the selected feature sequence, while the random forest algorithm uses the absolute majority voting method to make the final classification decision. The rules of the voting method are as follows:
H x =     c j ,       i f i = 1 T h i j x > 0.5 k = 1 N i = 1 T h i j x ; r e j e c t ,     o t h e r w i s e .
where h i represents base learner, [a];
c j represents the category label, [dimensionless];
T represents the number of samples to be tested, [a];
k represents the number of base classifiers, [dimensionless];
j represents the output of the base learner on the category label   c j , [dimensionless].
According to Formula (6), if the random forest has more than half of the predictions for a certain category voted, the final prediction will be this category; otherwise, the prediction will be rejected.

3. Comparison Process in the Experiment

3.1. Pig Face Identification Test Carried out with Random Forest Alone

3.1.1. Random Forest Model Parameter Determination

In order to determine the number of decision trees and parameters for the splitting quality performance function in the random forest, in this study, multiple tests were designed for the relationship between the number of decision trees, splitting quality performance function model and accuracy, recall rate, as well as f1-score when the pig face images were classified. Splitting quality function included gini and entropy. The number of decision trees was within the range of 0~100. The effects of each parameter combination on the performance of random forest are as shown in Figure 2.
Figure 2a shows the relationship between the number of decision trees within the random forest and the accuracy. The abscissa represents the number of decision trees and the ordinate represents the accuracy. The blue curve represents the gini impurity level “gini” and the green curve represents the information gain “entropy”. It did not matter if it was “gini” or “entropy” that was selected by splitting quality performance function; as the number of decision trees within the random forest increased, there was an obvious growth in the accuracy of forest classification. When the number of decision trees was less than 20, the accuracy grew rapidly; when the number of decision trees was over 20, the accuracy grew slowly; when the number of decision trees was over 65, the accuracy of random forest classification reached 90%; and with the increase in the number of decision trees, the classification accuracy kept stable. The more decision trees there are, the higher the computational complexity of the random forest; in this study, the number of decision trees in the forest was taken as 65, which could not only make the accuracy reach the expected requirement, but also make the computational complexity of random forest relatively low. As can be seen in Figure 2a, “gini” and “entropy” have a similar influence on the accuracy of the model. Therefore, for the splitting quality function of the random forest, either “gini” or “entropy” can be selected.
Figure 2b shows the relationship between the number of decision trees in the random forest and the recall rate. The abscissa represents the number of decision trees, and the ordinate represents the recall rate. Similarly to Figure 2a, the number of decision trees and the recall rate also show a logarithmic growth trend. Since the blue curve fluctuates less than the green recall curve, for the splitting quality performance function of random forest, adopting “gini” would be more stable than adopting “entropy”.
Figure 2c shows the relationship between the number of decision trees in the random forest and the f1-score. The abscissa represents the number of decision trees, and the ordinate represents the f1-score. The number of decision trees and the f1-score also showed a logarithmic growth trend. For the model with gini selected as the parameter, in most cases, the classification performance would be better than entropy. Considering the performance of the three indicators with different amounts of decision trees, the number of decision trees, which would be a parameter for subsequent tests, was taken as 65.

3.1.2. Evaluation Index of RF Model

According to the parameter determined in the above tests, the splitting quality function in the random forest was set as “gini”, and the number of decision trees was set as 65. Tests were conducted on the test set, and according to the test results, the confusion matrix was drawn, as shown in Figure 3.
The leftmost column of the confusion matrix represents the real category, the top row represents the predicted category, and the diagonal line represents the number of correct predictions. The precision, recall, and f1-score values of ten different pigs were obtained in accordance with the confusion matrix and Formulas (7)–(9), respectively, as shown in Table 2. The precision ratio was defined as
p r e c i s i o n = T P T P + F P
The recall ratio was defined as
r e c a l l = T P T P + F N
The recall ratio was also called the recall rate. The recall ratio and precision ratio changed in an opposite trend. The f1-score can measure the different preferences of these two indexes, and the formula was as follows:
f 1 - s c o r e = 2 × p r e c i s i o n × r e c a l l p r e c i s i o n + r e c a l l
where
TP represents the number of positive samples that are actually positive samples, [a];
FP represents the number of positive samples that are actually negative samples, [a];
FN represents the number of negative samples that are actually positive samples, [a].
As can be seen from Table 2, the average accuracy of classification and identification of pig face data with RF adopted reached 90.61%, with the recall rate and the f1-score reaching 89.76% and 89.79%, respectively.

3.2. Experiment of Pig Face Identification with RF + PCA Pre-Treatment

3.2.1. Determination of the k Value in Principal Component Analysis

At the first stage of the experiment, the number of principal components needs to be determined for principal component analysis. Here, the k value was taken as 300, with the variance explanation rate reaching over 95%.

3.2.2. Determination of RF Parameters in the Optimization Plan

The distribution of the data going through PCA dimension reduction may change to some extent; thus, the number of decision trees determined in the previous stage may not be the optimal value when used as the input for the RF model. Therefore, the number of decision trees in the RF model needs to be redetermined. With the same testing method adopted as in Section 3.1.1 ‘Random Forest Model Parameter Determination’, the relationships between the number of decision trees and the classification accuracy, recall rate, as well as f1-score of the RF model to pig face data were measured, and their relationship is as shown in Figure 4.
In Figure 4a, the abscissa represents the number of decision trees in the random forest, and the ordinate represents the precision. The green broken line represents entropy, and the blue broken line represents gini. When the number of decision trees was 1, the precision value was around 0.45, which meant that the classification effect was even lower than random guess. As the number of decision trees increased, the corresponding precision values grew rapidly. When the number of decision trees reached about 20, the precision value reached above 0.85. Then, as the number of decision trees continued to increase, the corresponding precision values grew slowly, accompanied by a small amplitude of oscillation. The overall performance of the green broken line, whose precision value reached the maximum of around 0.92 at 70, was better than that of the blue broken line. For the classification performance of the model, the accuracy, recall rate, and f1-score value need to be considered in a comprehensive way. According to the test results, the broken-line graphs that represent the relationship between the recall rate and the f1-score value for different numbers of decision trees were drawn, as shown in Figure 4b,c.
As shown in Figure 4b,c, if entropy was chosen as the parameter for the model, its performance would be better than gini. As the number of decision trees increased, entropy showed the same trend as precision, reaching a maximum of 0.92 and 0.90 when the number of decision trees was around 70. With the classification performance index for different numbers of decision trees considered, the optimal classification results were obtained when the number was 70.

3.2.3. Model Evaluation Index of Optimization Plan

The parameter determined in Section 3.2.2 ‘Determination of RF Parameters in the Optimization Plan’, “entropy”, was used for the splitting quality function in the random forest, and the number of decision trees was taken as 70. Tests were carried out on the test set, and the confusion matrix was drawn, as shown in Figure 5, in which the left-most column of the matrix refers to the actual category, the top row refers to the category in the actual prediction, while the leading diagonal shows the number correctly predicted, and the right-most column is a visual representation of the correct values for each category participating in the prediction.
The precision, recall, and f1-core values of 10 different pigs were obtained in accordance with the confusion matrix and Formulas (7)–(9), as shown in Table 3.
The RF classifier with PCA pre-treatment adopted was used for individual pig identification, which not only improved the identification accuracy but also reduced the testing time of the model. The specific test indexes for the two schemes are as shown in Table 4.
It can be seen from Table 4 that, with the optimization plan of PCA + RF pre-processing, except for the slight increase in training time, the other evaluation indicators have improved. In order to study the effect of the PCA preprocessing method on the efficiency of the machine learning model in pig individual recognition, the model running efficiency of pig individual recognition using SVM, KNN, PCA + SVM, and PCA + KNN is simultaneously compared through experiments. The results are shown in Table 5.
According to Table 5, the accuracy of the SVM, KNN, and RF models reached 83.66%, 91.46%, and 90.61%, respectively, before PCA pre-treatment; the accuracy of PCA + RF was the highest (93.22%), and the training time of the three models increased or decreased in the aspect of running efficiency; however, the testing time directly related to the practical application is reduced. Although the PCA + SVM and PCA + KNN methods reduce greatly, their accuracy is only 82.82%; PCA + RF is the main content of further research.
In addition, in other studies [28,29,31,46,47,48,49,50,51], the authors studied the role of modern neural networks, such as YOLO, AlexNet, and Tiny-YOLO, in pig identification, head and face postures, and behavioral analysis. However, the modern neural network has a large number of parameters and a deep hierarchical structure, which has become the bottleneck of its research results in the development of embedded applications; this is one of the starting points of this study.

4. Discussion

As can be seen in Table 4, in PCA + random forest optimizing test scheme, the precision reached 93.22%, the recall value was 92.52%, and the f1-score was 92.60%, which increased by 2.66, 2.76, and 2.81 percentage points, respectively, compared with the values obtained in the random forest alone scheme. In terms of the operating efficiency of the algorithm, the training time increased slightly from 1229 ms to 1340 ms, with an increase of 9%, though the testing time decreased from 8 ms to 6 ms. In the new scheme, only 75% of the testing time for the old scheme is needed. This improvement better suits actual production scenarios.
The accuracy, recall rate, and f1-score improved for the following two reasons: First, after the original pig sample data underwent PCA feature extraction, the model had the most identifiable features of the pigs extracted, which was conducive to the improvement of accuracy. Second, with PCA employed, the secondary features were ignored, and the noise in the data was filtered; thus, the model possessed better generalization ability, as was verified in the tests for the validation set. PCA needed some time to process, and the training time was increased, while the post-optimization testing time was greatly improved and even reduced to 75% of that of the old scheme. With the PCA + random forest test scheme used, the identification accuracy of the algorithm was improved; the operating efficiency also improved, which provides both theoretical and experimental support for further embedded application of the algorithm.

5. Conclusions

This paper studied the influence of the PCA pre-treatment method on the efficiency of identifying ten pigs with an RF classifier. The parameters of the classifier were determined through tests. By comparing the influence of the two testing schemes, in which one adopted the RF classifier alone and the other adopted PCA + RF, on identification efficiency, the following conclusions are drawn:
(1)
For individual identification of pigs, the RF classifier can be used, for which the parameter selection is relative to the pre-treatment method. If RF alone is used, the splitting quality function shall be “gini”, and the number of decision trees shall be 65; in the case of the PCA + RF optimization scheme, the corresponding parameters shall be “entropy” and 70.
(2)
PCA pre-treatment can increase the efficiency of individual pig identification with RF, and the accuracy, recall rate, and the f1-score are increased by 2.66, 2.76, and 2.81 percentage points, respectively, while the testing time is reduced to 75% of the original value.
(3)
The RF classifier that underwent PCA pre-treatment is more suitable for application in mobile terminals and embedded application, and it is suitable for the development of a portable and real-time pig face identification system; thus, the cost of intelligent breeding and management of small and medium-sized farms can be reduced, and the process of intellectualization of small and medium-sized farms can be promoted.

Author Contributions

Conceptualization, H.Y.; methodology H.Y. and Z.H.; software, H.Y. and J.L.; validation H.Y. and Z.H.; formal analysis, H.Y. and S.C.; investigation, Q.L. and E.L.; resources, H.W., Q.L., S.C. and J.L.; data curation, Z.H.; writing—original draft preparation, H.Y.; writing—review and editing, H.Y. and Z.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Plan of China (2016YFD0701801), the Shanxi Province Basic Research Program Project (Free Exploration) (Grant No. 20210302124523, 202103021224149, 202103021223141), and the Doctor Scientific Research Foundation of Shanxi Agricultural University (2020BQ14). The authors are grateful and honored to have obtained support from the Key Laboratory of Biomechanics.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors thank the editor and anonymous reviewers for providing helpful suggestions for improving the quality of this manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Voulodimos, A.S.; Patrikakis, C.Z.; Sideridis, A.B.; Ntafis, V.A.; Xylouri, E.M. A complete farm management system based on animal identification using RFID technology. Comput. Electron. Agric. 2010, 70, 380–388. [Google Scholar] [CrossRef]
  2. Ghosh, P.; Mandal, S.N. PigB: Intelligent pig breeds classification using supervised machine learning algorithms. Int. J. Artif. Intell. Soft Comput. 2022, 7, 242–266. [Google Scholar] [CrossRef]
  3. Bazi, Y.; Melgani, F. Convolutional SVM Networks for Object Detection in UAV Imagery. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3107–3118. [Google Scholar] [CrossRef]
  4. Smith, R.P.; Gavin, C.; Gilson, D.; Simons, R.R.L.; Williamson, S. Determining pig holding type from British movement data using analytical and machine learning approaches. Prev. Vet. Med. 2020, 178, 104984. [Google Scholar] [CrossRef]
  5. Shaik Mazhar, S.A.; Suseendran, G. Precision Pig Farming Image Analysis Using Random Forest and Boruta Predictive Big Data Analysis Using Neural Network and K-Nearest Neighbor. In Proceedings of the International Conference on Intelligent Engineering and Management (ICIEM), London, UK, 28–30 April 2021; pp. 260–264. [Google Scholar] [CrossRef]
  6. Adeniyi, E.A.; Ogundokun, R.O.; Gbadamosi, B.; Misra, S.; Kalejaiye, O. Classification of Swine Disease Using K-Nearest Neighbor Algorithm on Cloud-Based Framework. In Artificial Intelligence for Cloud and Edge Computing; Springer: Berlin/Heidelberg, Germany, 2022; pp. 71–90. [Google Scholar]
  7. Lee, W.; Ham, Y.; Ban, T.-W.; Jo, O. Analysis of growth performance in swine based on machine learning. IEEE Access 2019, 7, 161716–161724. [Google Scholar] [CrossRef]
  8. Kontschieder, P.; Fiterau, M.; Criminisi, A.; Bulo, S.R. Deep Neural Decision Forests. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 11–18 December 2015. [Google Scholar]
  9. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar] [CrossRef]
  10. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  11. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  12. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  13. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  14. Yang, X.; Chai, L.; Bist, R.B.; Subedi, S.; Wu, Z. A Deep Learning Model for Detecting Cage-Free Hens on the Litter Floor. Animals 2022, 12, 1983. [Google Scholar] [CrossRef]
  15. Maselyne, J.; Saeys, W.; De Ketelaere, B.; Mertens, K.; Vangeyte, J.; Hessel, E.F.; Millet, S.; Van Nuffel, A. Validation of a High Frequency Radio Frequency Identification (HF RFID) system for registering feeding patterns of growing-finishing pigs. Comput. Electron. Agric. 2014, 102, 10–18. [Google Scholar] [CrossRef]
  16. Maselyne, J.; Adriaens, A.; Huybrechts, T.; De Ketelaere, B.; Millet, S.; Vangeyte, J.; Van Nuffel, A.; Saeys, W. Measuring the drinking behaviour of individual pigs housed in group using radio frequency identification (RFID). Animal 2016, 10, 1557–1566. [Google Scholar] [CrossRef]
  17. Kapun, A.; Adiron, A.; Gallmann, E. Case Study on Recording Pigs’ Daily Activity Patterns with a UHF-RFID System. Agriculture 2020, 10, 542. [Google Scholar] [CrossRef]
  18. Maselyne, J.; Saeys, W.; Briene, P.; Mertens, K.; Vangeyte, J.; De Ketelaere, B.; Hessel, E.F.; Sonck, B.; Van Nuffel, A. Methods to construct feeding visits from RFID registrations of growing-finishing pigs at the feed trough. Comput. Electron. Agric. 2016, 128, 9–19. [Google Scholar] [CrossRef]
  19. Zhu, J.; Ma, S.S.; Bi, Y.G.; Cui, H.M. Construction of digital breeding platform for breeding pig. Trans. CSAE 2010, 26, 215–219. [Google Scholar]
  20. Zhu, W.; Zhong, F.; Li, X. Automated Monitoring System of Pig Behavior Based on RFID and ARM-LINUX. In Proceedings of the Third International Symposium on Intelligent Information Technology and Security Informatics, Jian, China, 2–4 April 2010. [Google Scholar] [CrossRef]
  21. Chen, T.Y.; Lin, Y.H.; Shieh, C.S.; Lo, C.C.; Guo, S.W.; Horng, M.F. A RFID-based Diet Estimation of Grower Pigs. In Proceedings of the 2021 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS), Hualien, Taiwan, 16–19 November 2021; pp. 1–2. [Google Scholar] [CrossRef]
  22. He, Y.; Tiezzi, F.; Howard, J.; Maltecca, C. Predicting body weight in growing pigs from feeding behavior data using machine learning algorithms. Comput. Electron. Agric. 2021, 184, 106085. [Google Scholar] [CrossRef]
  23. Gorczyca, M.T.; Milan, H.F.M.; Maia, A.S.C.; Gebremedhin, K.G. Machine learning algorithms to predict core, skin, and hair-coat temperatures of piglets. Comput. Electron. Agric. 2018, 151, 286–294. [Google Scholar] [CrossRef]
  24. Wongsriworaphon, A.; Arnonkijpanich, B.; Pathumnakul, S. An approach based on digital image analysis to estimate the live weights of pigs in farm environments. Comput. Electron. Agric. 2015, 115, 26–33. [Google Scholar] [CrossRef]
  25. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 8–16 October 2016; pp. 21–37. [Google Scholar] [CrossRef]
  26. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  27. Win, K.D.; Kawasue, K.; Yoshida, K.; Lee, G. Automatic pig selection system based on body size using a camera: Rotating mechanics for pig selection. Artif. Life Robot. 2020, 26, 155–161. [Google Scholar] [CrossRef]
  28. Yan, H.W.; Liu, Z.Y.; Cui, Q.L.; Hu, Z.W. Multi-target detection based on feature pyramid attention and deep convolution network for pigs. Trans. Chin. Soc. Agric. Eng. (Trans. CSAE) 2020, 36, 193–202. [Google Scholar]
  29. Yan, H.W.; Liu, Z.Y.; Cui, Q.L.; Hu, Z.W.; Li, Y.W. Detection of facial gestures of group pigs based on improved Tiny-YOLO. Trans. Chin. Soc. Agric. Eng. (Trans. CSAE) 2019, 35, 169–179. [Google Scholar]
  30. Kim, T.; Kim, Y.; Kim, S.; Ko, J. Estimation of Number of Pigs Taking in Feed Using Posture Filtration. Sensors 2022, 23, 238. [Google Scholar] [CrossRef] [PubMed]
  31. Hu, Z.W.; Yang, H.; Lou, T.T. Dual attention-guided feature pyramid network for instance segmentation of group pigs. Comput. Electron. Agric. 2021, 186, 106140. [Google Scholar] [CrossRef]
  32. Hansen, M.F.; Smith, M.L.; Smith, L.N.; Salter, M.G.; Baxter, E.M.; Farish, M.; Grieve, B. Towards on-farm pig face recognition using convolutional neural networks. Comput. Ind. 2018, 98, 145–152. [Google Scholar] [CrossRef]
  33. Parkhi, O.M.; Vedaldi, A.; Zisserman, A. Deep Face Recognition. In Proceedings of the British Machine Vision Conference (BMVC), Swansea, UK, 7–10 September 2015; pp. 41.1–41.12. [Google Scholar]
  34. Ramprasaath, R.S.; Das, A.; Vedantam, R. Grad-CAM: Why Did You Say That? arXiv 2016, arXiv:1610.07450. [Google Scholar]
  35. Shi, Y.W.; Zhang, Y.Q.; Liu, K.S. Research on Security Issues Based on RFID System. Comput. Sci. 2012, 39, 214–216. [Google Scholar]
  36. Zhou, J.; Lu, J.Z.; Tu, Y.Z. Discussion on the Use of Intelligent Recognition Terminal to Promote the Management of Animal Ear Tags. Chin. Livest. Poult. Breed. 2018, 14, 33–34. [Google Scholar]
  37. Adrion, F.; Kapun, A.; Eckert, F.; Holland, E.-M.; Staiger, M.; Götz, S.; Gallmann, E. Monitoring trough visits of growing-finishing pigs with UHF-RFID. Comput. Electron. Agric. 2018, 144, 144–153. [Google Scholar] [CrossRef]
  38. Wang, Z.; Liu, T. Two-stage method based on triplet margin loss for pig face recognition. Comput. Electron. Agric. 2022, 194, 106737. [Google Scholar] [CrossRef]
  39. Basak, H.; Kundu, R.; Chakraborty, S.; Das, N. Cervical Cytology Classification Using PCA and GWO Enhanced Deep Features Selection. SN Comput. Sci. 2021, 2, 369. [Google Scholar] [CrossRef]
  40. Smallman, L.; Underwood, W.; Artemiou, A. Simple Poisson PCA: An algorithm for (sparse) feature extraction with simultaneous dimension determination. Comput. Stat. 2019, 35, 559–577. [Google Scholar] [CrossRef]
  41. Salih, A.I.; Kardouchi, M.; Belacel, N. Fast and Efficient Face Recognition System Using Random Forest and Histogram of Oriented Gradients. In Proceedings of the International Conference of the Biometrics Special Interest Group (BIOSIG), Darmstadt, Germany, 6–7 September 2012; Volume P-196, pp. 1–11. [Google Scholar]
  42. Witten, I.H.; Frank, E.; Hall, M.A. Data Mining: Practical Machine Learning Tools and Techniques; Morgan Kaufmann: Los Angeles, CA, USA, 2011; pp. 62–65. [Google Scholar]
  43. Belle, V. Detection and Recognition of Human Faces Using Random Forests for a Mobile Robot. Ph.D. Thesis, Rheinisch-Westfälische Technische Hochschule Aachen, Aachen, Germany, 2008. [Google Scholar]
  44. Archer, K.J.; Kimes, R.V. Empirical Characterization of Random Forest Variable Importance Measures. Comput. Stat. Data Anal. 2008, 52, 2249–2260. [Google Scholar] [CrossRef]
  45. Fanelli, G.; Gall, J.; Van Gool, L. Real Time Head Pose Estimation with Random Regression Forests. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Colorado Springs, CO, USA, 21–23 June 2011; pp. 617–624. [Google Scholar] [CrossRef]
  46. Yan, H.W.; Cui, Q.L.; Liu, Z.Y. Pig Face Identification Based on Improved Alexnet Model. INMATEH-Agric. Eng. 2020, 61, 97–104. [Google Scholar] [CrossRef]
  47. Hu, Z.W.; Yang, H.; Lou, T.T.; Hu, G.; Xie, Q.Q.; Huang, J.J. Extraction of pig contour based on fully convolutional networks. J. South China Agric. Univ. 2018, 39, 111–119. [Google Scholar]
  48. Hu, Z.W.; Yang, H.; Lou, T.T. Instance detection of group breeding pigs using a pyramid network with dual attention feature. Trans. Chin. Soc. Agric. Eng. 2021, 37, 166–174. [Google Scholar]
  49. Hu, Z.W.; Yan, H.W.; Lou, T.T. Parallel channel and position attention-guided feature pyramid for pig face posture detection. Int. J. Agric. Biol. Eng. 2022, 15, 222–234. [Google Scholar] [CrossRef]
  50. Hu, Z.W.; Yang, H.; Lou, T.T.; Yan, H.W. Concurrent channel and spatial attention in Fully Convolutional Network for individual pig image segmentation. Int. J. Agric. Biol. Eng. 2023, 16, 232–242. [Google Scholar] [CrossRef]
  51. Yan, H.W.; Hu, Z.W.; Cui, Q.L. Study on feature extraction of pig face based on principal component analysis. INMATEH-Agric. Eng. 2022, 68, 333–340. [Google Scholar] [CrossRef]
Figure 1. Pig Samples.
Figure 1. Pig Samples.
Animals 13 01555 g001
Figure 2. The relationship between the evaluation index of RF model and its parameter. (a) Accuracy; (b) recall rate; (c) f1 value.
Figure 2. The relationship between the evaluation index of RF model and its parameter. (a) Accuracy; (b) recall rate; (c) f1 value.
Animals 13 01555 g002
Figure 3. RF prediction result confusion matrix.
Figure 3. RF prediction result confusion matrix.
Animals 13 01555 g003
Figure 4. The relationship between RF model performance and parameters after pre-PCA processing. (a) Accuracy; (b) recall rate; (c) f1 value.
Figure 4. The relationship between RF model performance and parameters after pre-PCA processing. (a) Accuracy; (b) recall rate; (c) f1 value.
Animals 13 01555 g004
Figure 5. RF + PCA prediction result confusion matrix.
Figure 5. RF + PCA prediction result confusion matrix.
Animals 13 01555 g005
Table 1. Named terms list.
Table 1. Named terms list.
AbbreviationMeaning
PCAPrincipal Component Analysis
RFRandom Forest
RFIDRadio Frequency IDentification
HFRFIDHigh-Frequency Radio Frequency IDentification
UHFRFIDUltra-High Frequency Radio Frequency IDentification
SVMSupport Vector Machine
KNNK-Nearest Neighbors
LDALinear Discriminant Analysis
VGGVisual Geometry Group
SSDSingle Shot Detector
F-rcnnFaster Region-based Convolutional Neural Networks
YOLOYou Only Look Once
Table 2. RF model prediction performance table.
Table 2. RF model prediction performance table.
CategoryPrecision (%)Recall (%)f1-Score (%)Count [a]
1721008431
296899328
397919432
495848925
587969228
6100798824
791888924
878948531
9100839118
1092929213
average90.6189.7689.7925
Table 3. RF + PCA model prediction performance table.
Table 3. RF + PCA model prediction performance table.
CategoryPrecision (%)Recall (%)f1-Score (%)Count [a]
181978831
2100939628
394979532
4100929625
596869128
695838924
792929224
883979031
9100899418
1010010010013
average93.2292.5292.6025
Table 4. The optimization result of the RF model by PCA pre-processing.
Table 4. The optimization result of the RF model by PCA pre-processing.
ModelPrecision (%)Precision ChangeTest_Time (ms)Test_New/Old (%)Train_Time (ms)Traintest_New/Old (%)
RF90.61081001229100
PCA + RF93.22+2.616751340109
Table 5. Comparison of generalization test results.
Table 5. Comparison of generalization test results.
ModelPrecision (%)Precision ChangeTest_Time (ms)Test_New/Old (%)Train_Time (ms)Traintest_New/Old (%)
SVM83.66032910012,823100
KNN91.4601306100187100
RF90.61081001229100
PCA + SVM88.85+5.196920.9386130.1
PCA + KNN82.82−8.6493794.8
PCA + RF93.22+2.616751340109
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yan, H.; Cai, S.; Li, E.; Liu, J.; Hu, Z.; Li, Q.; Wang, H. Study on the Influence of PCA Pre-Treatment on Pig Face Identification with Random Forest. Animals 2023, 13, 1555. https://doi.org/10.3390/ani13091555

AMA Style

Yan H, Cai S, Li E, Liu J, Hu Z, Li Q, Wang H. Study on the Influence of PCA Pre-Treatment on Pig Face Identification with Random Forest. Animals. 2023; 13(9):1555. https://doi.org/10.3390/ani13091555

Chicago/Turabian Style

Yan, Hongwen, Songrui Cai, Erhao Li, Jianyu Liu, Zhiwei Hu, Qiangsheng Li, and Huiting Wang. 2023. "Study on the Influence of PCA Pre-Treatment on Pig Face Identification with Random Forest" Animals 13, no. 9: 1555. https://doi.org/10.3390/ani13091555

APA Style

Yan, H., Cai, S., Li, E., Liu, J., Hu, Z., Li, Q., & Wang, H. (2023). Study on the Influence of PCA Pre-Treatment on Pig Face Identification with Random Forest. Animals, 13(9), 1555. https://doi.org/10.3390/ani13091555

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop