Next Article in Journal
Phenolic Profile and the Antioxidant, Anti-Inflammatory, and Antimicrobial Properties of Açaí (Euterpe oleracea) Meal: A Prospective Study
Previous Article in Journal
Comparison of Binding Properties of a Laccase-Treated Pea Protein–Sugar Beet Pectin Mixture with Methylcellulose in a Bacon-Type Meat Analogue
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Double-Camera Fusion System for Animal-Position Awareness in Farming Pens

1
Department of Electrical Engineering, Eindhoven University of Technology, 5612 AP Eindhoven, The Netherlands
2
Department of Mathematics and Computer Science, Eindhoven University of Technology, 5612 AZ Eindhoven, The Netherlands
3
Department of Animal Sciences, Wageningen University & Research, 6700 AH Wageningen, The Netherlands
*
Author to whom correspondence should be addressed.
Foods 2023, 12(1), 84; https://doi.org/10.3390/foods12010084
Submission received: 17 October 2022 / Revised: 6 December 2022 / Accepted: 9 December 2022 / Published: 23 December 2022
(This article belongs to the Section Food Engineering and Technology)

Abstract

:
In livestock breeding, continuous and objective monitoring of animals is manually unfeasible due to the large scale of breeding and expensive labour. Computer vision technology can generate accurate and real-time individual animal or animal group information from video surveillance. However, the frequent occlusion between animals and changes in appearance features caused by varying lighting conditions makes single-camera systems less attractive. We propose a double-camera system and image registration algorithms to spatially fuse the information from different viewpoints to solve these issues. This paper presents a deformable learning-based registration framework, where the input image pairs are initially linearly pre-registered. Then, an unsupervised convolutional neural network is employed to fit the mapping from one view to another, using a large number of unlabelled samples for training. The learned parameters are then used in a semi-supervised network and fine-tuned with a small number of manually annotated landmarks. The actual pixel displacement error is introduced as a complement to an image similarity measure. The performance of the proposed fine-tuned method is evaluated on real farming datasets and demonstrates significant improvement in lowering the registration errors than commonly used feature-based and intensity-based methods. This approach also reduces the registration time of an unseen image pair to less than 0.5 s. The proposed method provides a high-quality reference processing step for improving subsequent tasks such as multi-object tracking and behaviour recognition of animals for further analysis.

1. Introduction

Animals are one of the primary food sources for society, supplementary to vegetables and cereals [1]. Pigs are one of the most commonly raised livestock in the world, and they are the primary protein supplier for millions of people [2]. The world population is projected to be 9.2 billion in 2050 [3] and food security is facing more severe global challenges. However, traditional pig farms produce many greenhouse gasses and cause water pollution [4], which urgently need to transform into precision livestock farming (PLF) [5]. PLF has resulted in the development of intelligent tools to manage animals. These tools take the precise measurement, monitoring, and tracking of individual and group animals, yielding to automated status and condition estimations of the animals [6,7,8,9]. Nowadays, computer vision technologies are increasingly employed for livestock monitoring [7,8,9]. The application of computer vision to monitor animal farms (e.g., pigs) can address a broad range of monitoring and analysis tasks [7]. Cowton et al. [10] have designed a complete system for recognizing, tracking, and extracting metrics related to individual pigs via RGB cameras. Specifically, they combined a faster region-based convolutional neural network with DeepSORT and used a transfer learning strategy during the training. The authors have evaluated the method on their pig detection dataset and achieved an average precision of 0.901 for recognizing pigs in frames. The result has enabled to yield 92% for Multi-object tracking accuracy (MOTA) and 73.4% for identity F1-score (IDF1). The above-mentioned study is based on image data captured by a single-camera system in a fixed position. In large-scale farming scenarios, higher object density typically causes more occlusions. In addition, different lighting conditions, such as overexposure within a recording view, also generate challenges to subsequent data analysis. In this work, we propose to install two tilted cameras at the top of pig pens with an overlapping field of view. Compared with single-camera-based monitoring, a double-camera-based monitoring system can capture more appearance details of each pig from different viewpoints. However, this raises a challenge for the optimal fusion of the information from each individual camera.
Recently, many researchers have conducted studies on double-camera or multi-camera fusion, where most studies focused on pedestrian and vehicle tracking. In order to improve the detection performance of the multi-camera system under occlusion conditions, Wu et al. [11] have proposed a Multi-View Region Proposal Network (MVRPN) to localize the vehicles on the ground plane. The authors have used multiple side views to evaluate the ground-occupancy vector of the vehicles, where MVRPN is trained under the supervision of top-view ground-plane information. The method achieves an average precision of 0.789 and a multi-object detection precision of 0.708 on a dataset captured on the road in Richardson, TX, USA. Wang et al. [12] have unified information from multiple cameras by transferring object representations from multiple side views to the top view. First, the authors apply a detector to each side view to obtain the detection results. Then they use a neural network to map these results into top-view representations. Finally, the authors have applied a clustering algorithm to the mapped object representations in the top view to fuse the information from multiple cameras. Wang et al. have obtained a distance error of 0.019 in normalized coordinates on their dataset.
These research works are very ingenious at integrating information and performing well in their respective tasks. However, these works require feature engineering or complete detection and trajectory extraction using a single camera prior to information fusion. The single-camera feature engineering inevitably leads to the loss of information. Furthermore, providing the image information from only one camera for the receptive field of a convolutional neural network can potentially lead to under-utilizing the feature characterization. In addition, the use of a single camera has shortcomings in solving occlusions when animals are densely packed and there are quality limitations in the corner views. Therefore, we propose a method to find the spatial correspondence at the pixel level in a double-camera system. In this way, the image information of the same objects collected from different perspectives can be spatially aligned, providing more flexibility and accuracy for developing intelligent tools based on PLF.
The process of unifying objects recorded from different viewpoints into the same coordinate system is called image registration. Until recently, numerous image registration tasks with critical purposes such as in the medical domain had to be performed manually. Such registration tasks can be time-consuming and even require professional pre-job training for operators when registering complex images. In recent years, researchers have been working on developing automated registration methods to circumvent the limitations of manual registration. When applying artificial intelligence, deep learning can be used to automatically register images and has proven to be a valid upcoming solution [13].
Deep learning was originally applied to optimize the intermediate process of intensity-based iterative registration. Most of the existing research aims at learning a new similarity metric that can be inserted into a classical iterative registration framework [14,15]. However, the improvement of such approaches in multi-modal registration tasks is more evident than in single-modal registration tasks [16]. Moreover, the registration method based on iterative techniques is difficult to apply to real-time registration [17,18]. Therefore, researchers have developed several supervised methods to obtain optimal registration results in one step. Chee et al. [19] proposed a method for predicting rigid transformation parameters without similarity metrics. The samples used to train the model were synthesized from unlabelled data. Experiments on magnetic resonance images of the axial view of the human brain showed that their method outperformed MI optimization-based registration. Rohe et al. [20] used a U-net-inspired CNN [21] to evaluate the displacement field. The SSD between the registered image and the ground-truth image was considered the optimization objective of the network. The authors validated the proposed method on 3D brain MR volumes. The results showed that in uni-modal conditions, the proposed method outperformed LCC Demons-based registration [22].
Supervised learning-based approaches allow for fast registration across applications. However, these labels are usually challenging to obtain and the quality of the registrations using this framework relies on the quality of the ground-truth registrations. These challenges motivate the development of partially supervised or unsupervised approaches. De Vos et al. [23] proposed a multi-stage, multi-scale, mono-modal registration method. The authors used a CNN to perform an affine transformation and subsequently a coarse-to-fine transformation. They validated their method on the Cardiac Cine MRI and Chest CT datasets, where the results also showed that their method was superior to the Elasix toolbox [24]. Hering et al. [25] proposed a weakly supervised transformation estimation. The authors considered both edge-based normalized gradient fields distance and segmentation overlap in the loss function. The experimental results on 2D Cine-MR Images showed that the proposed method outperformed a multi-level registration approach mentioned in [26].
Although unsupervised image-similarity-based registration avoids the need for expert labels of any kind and outperforms conventional iterative techniques [27], the image-similarity metrics (ISMs) used to construct the loss function only evaluate the similarity between the image intensity fields. These similarity quantifications are almost meaningless without context constraints on sampling since they do not indicate the actual measurement error [28]. In other words, these networks are unaware of the displacement errors before and after the registration during training. Rohlfing et al. [29] elaborated on the limitations of using image-similarity metrics and region overlaps in the deformable registration tasks. They also proved that the quality of the registration result is not necessarily related to the magnitude of the quantized value. Although tedious and time-consuming, placing landmarks is still the gold standard for evaluating the registration results.
In this paper, the research objective is to improve visual monitoring of animals such that their behaviour can be identified at the group or individual level, which will enable the farmers to take action for social behaviour so that animals are not lost by accidents or disease. To this end, we propose a fine-tuned semi-supervised registration framework for spatial fusion in a double-camera system. It will fully utilize the complementary information between cameras to improve the performance of multi-object detection, multi-object tracking, and behaviour recognition. Specifically, a small number of manual landmarks are placed to serve as a control signal and provide global constraints during optimization. We validate this approach with pig datasets, which have been captured at real farms. The proposed framework uses a convolutional neural network for mapping all pixels from a moving image to a fixed image by sharing the same parameters for a set of image pairs. The procedure learns a common representation that can align any unseen frame pair from the same recording settings. Aligning a new image pair is achieved by simply applying the learned function made from the given pairs, which results in a fast registration of the new unseen image pair.
This article is organized as follows. Section 2 introduces the image dataset, preprocessing methods, deformation models, evaluation methods, and experimental details. Section 3 provides qualitative, quantitative, and running time evaluations. Section 4 discusses the experimental results, application scenarios, and future improvements.

2. Materials and Methods

This section first introduces the dataset used to train and test the proposed methods for estimating spatial correspondence between images from different viewpoints. The section also presents camera calibration, masking, grey-scale conversion, and global initial registration methods to preprocess the raw data. After these steps, the unsupervised and semi-supervised learning-based deformable models are elaborated. Finally, the evaluation method and the implementation details are described.

2.1. Dataset Description

A set of video data from pigs is used in this study, recorded at a farm in Volmer, Germany. Each farm enclosure contains 10–11 pigs of the same age and similar body size, where some pigs are artificially coloured on their backs. For our experiment, we have selected video data of 4 pens recorded by 2 cameras (LOREX 4KSDAI168) mounted at the top of the wall opposite the feeding station and at the top of the wall to the left of the feeding station. Figure 1 shows the positions of the cameras in the pig pens. The field of view of both cameras covers the entire living ground and is recorded continuously during the experiment. The data acquisition scenes include daylight changes, solid concrete floors, urine stains, manure, and other sundries. The videos have a resolution of 1280 × 720 pixels and are recorded at 15 fps. The recorded video data is saved as one file every hour. Specifically, we have selected video recordings from 2:00–3:00 p.m. for 7 days in February 2022 (with intervals of 4 days), and then resampled them to video frames at 12-second intervals for experimental material. Due to the smaller pen width, neighbouring pens are visible in the videos of the current pen. Examples of the collected raw image data of the pig pens are shown in Figure 2a.

2.2. Data Pre-Processing

2.2.1. Camera Calibration

The process of evaluating camera parameters using images containing a specific calibration pattern is called calibration. When these parameters are obtained, the distortion from the image can be removed or the location of the camera can be predicted. To obtain these parameters, the calibration pattern (usually a checkerboard) inside the real-world image should be captured. For the calibration, we customize a checkerboard of (1 × 1 sq. meter) and position it at different angles in front of each of the two lenses to obtain the camera parameters. Since the lens and image plane of the camera are parallel in orientation, we only focus on the correction of radial distortion. Therefore, we omit the camera parameters and directly describe the distortion transformation. The distorted points are denoted as ( x distorted , y distorted ) :
x distorted = x ( 1 + k 1 · r 2 + k 2 · r 4 + k 3 · r 6 ) , y distorted = y ( 1 + k 1 · r 2 + k 2 · r 4 + k 3 · r 6 ) ,
where x and y are normalized image coordinates; k 1 , k 2 , and k 3 are radial distortion coefficients of the lens with r 2 = x 2 + y 2 . In this part, only k 1 and k 2 have been used for calibration. The calibrated video frames are shown in Figure 2b.

2.2.2. Masking and Grey-Scale Conversion

Since this study focuses on the pixel-level correspondence between side-view video frames and front-view video frames in the actual pig pen, masking the content outside the pen in video frames allows the algorithm to pay more attention to the actual transformation. RGB images contain three channels capable of recording rich color information. However, in real farming scenes, the lighting conditions change frequently, which has a direct impact on the generalization of the model [30]. An effective way is to convert RGB images into grey-scale images [31], so that the learning ability of the model can be centralized in the tasks of the interest. Figure 2c illustrates the masked and grey-scaled image pair.

2.2.3. Baseline Method

Projective transformation is considered the baseline method in this work. For each pen, 35 landmark pairs are placed on a randomly chosen frame pair to obtain the parameters for projective transformation. Then these parameters are used to align the remaining image pairs in the same pen. Figure 2d shows the outcome after overlapping a pre-registered side-view frame (magenta) with its corresponding front-view frame (green). Landmark pairs used to perform pre-registration are not employed for training and testing.

2.3. Deformation Models

The deformation model is refining the pre-registration transformation to a full registration of higher quality. We assume that I f : Ω f R n and I m : Ω m R n denote the front-view images (subscript f) and the corresponding pre-registered side-view images (subscript m), respectively, where in our settings n = 2 . The problem of finding a dense non-linear transformation is reduced to an optimization problem, specified by
arg min E ( Φ ) = min Φ M ( I f , I m ( Φ ) ) + μ R ( Φ ) ,
where I m ( Φ ) is I m warped by Φ , the function M measures the similarity between the fixed image I f and the moving image I m that has been transformed by Φ . Function R ( Φ ) represents a constraint imposed on the deformation field Φ , ensuring that the deformation field composed of displacement vectors of all pixels is smooth.
Figure 3 presents an overview of the proposed method. First, an unsupervised convolutional neural network is designed to parameterize the mapping function T θ ( I m , I f ) = Φ using unlabelled training samples. After this, a semi-supervised network is constructed, which has the same convolutional layers as the unsupervised model so that it can adopt the learned weights from the unsupervised network. Besides the similarity metric M between I f and I m ( Φ ) , the loss function of the semi-supervised network also includes an item that refers to the distance errors between transformed landmarks and the ground-truth landmarks. Finally, a small number of annotated samples are fed into the semi-supervised network to fine-tune the pre-trained weights.
This design is motivated by following the concept of transfer learning. In this experiment, the task of registering image pairs with landmarks can be regarded as the target domain; the task of registering image pairs without landmarks can be seen as the source domain. Since landmark annotation is labour-intensive and time-consuming, the cost of learning directly from the target domain from scratch is too high. Fortunately, there are large available unlabelled samples that can be used for source-domain learning, so that the knowledge learned from the source domain can be exploited to assist in learning new knowledge in the target domain.

2.3.1. Unsupervised Deformation Model

In this part, the network T is modelled using a set of learnable parameters θ . The architecture of the network is inspired by a U-Net, which is similarly composed of encoder, decoder, and skip connections. This is motivated as follows.
The architecture of the proposed network is adopted from the U-Net, since the encoder–decoder structure shows state-of-the-art performance on semantic segmentation tasks, which meets the requirement to learn displacement vectors at the pixel level. Moreover, U-Net can be trained end-to-end with a very small number of samples. Finally, the up-sampling layer fuses the features of its corresponding down-sampling layer through the skip connection, which can provide more spatial domain information for image registration.
Figure 4 depicts the architecture of the proposed network; Table 1 and Table 2 show the specifications of the convolution filters used in up-sampling and down-sampling.
The proposed network combines the moving (side view) and the fixed images (front view) as a single input with a size of 928 × 1408 × 2 pixels. The kernel size of all convolutional layers is 3 × 3, and Leaky ReLU is used as the activation function. In the encoding, the convolutional kernels capture the hierarchical features of the images at different resolution scales. The 2 × 2 pooling layers halve the size of the feature maps calculated by the previous convolutional layers until the smallest layer is reached. The feature maps also stream to the decoder and combine with the deconvolution output, which is critical for the estimation of the deformation field. The size of the receptive field of the last convolutional layer determines the maximum moving distance for the pixels. In the decoding, the deconvolution layers double the size of the feature maps to match the counterpart from the encoder. A convolutional layer with two 1 × 1 kernels compose the last layer of the decoding. The output Φ is of size 928 × 1408 × 2, which refers to the moving distance in the spatial directions for each pixel. In each iteration, the mapping from I m to the deformed I m ( T θ ( I m , I f ) ) is implemented by a spatial transformer network [32]. This module enables the spatial transformation within the network and requires no modification to the minimizing optimization process on E ( Φ ) in Equation (2). Furthermore, the differentiable characteristic of the module makes it suitable to be used in gradient descent-based networks. Finally, for robustness to intensity variations, the mean-squared error (MSE) is used to measure the similarity between I m and I f . The MSE is specified as
MSE = 1 n i = 1 n I f ( i ) I m ( i ) ( Φ ) 2 ,
where I f ( i ) is the intensity value of individual pixels i (with n the number of pixels) in I f and I m ( i ) ( Φ ) indicates the intensity value of pixels in I m . We adopt from [33] a regularization process, where diffusion regularization is used to regularize the spatial gradients given by
R ( Φ ) = ( x , y ) Ω Φ ( x , y ) 2 .
The optimal solution to the spatial transformation is searched by minimizing the loss function:
L ( I f , I m , Φ ( θ ) ) = MSE ( I f , I m ( Φ ( θ ) ) ) + λ ( x , y ) Ω Φ ( x , y ) 2 ,
where λ is the regularization parameter.

2.3.2. Semi-Supervised Deformation Model

Besides image pairs, there is a set of corresponding landmark points in the input in the semi-supervised network. The configurations of encoder, decoder, and skip connections in the unsupervised network are adopted here with some adaptations. Specifically, in the semi-supervised framework, there is an additional spatial transformation layer that is expected to transform the landmark points ( x l p , y l p ) on I m using the intermediate deformation field:
( x l p , m , y l p , m ) = f s t l ( x l p , m , y l p , m ) .
Here, ( x l p , m , y l p , m ) are the coordinates of the landmarks on I m and f s t l is the spatial transformation function related to θ that transform the landmarks to the supposed position. Since we aim at minimizing the positioning errors of the landmarks, an extra loss term is added to the loss function of Equation (5). To this end, after obtaining the transformed landmarks in each iteration, the mean value of the Euclidean distances between transformed and ground-truth landmarks is calculated and add it to the loss function. Finally, the learning process minimizes the modified extended loss function:
L ( I f , I m , Φ ( θ ) ) + μ 1 k j = 1 k [ ( x l p , f , y l p , f ) ( x l p , m , y l p , m ) ] 2 ,
where μ is a hyper-parameter and k is the number of landmark pairs.

2.4. Evaluation Method

To validate the proposed methods, six landmarks are manually selected on each moving and fixed image in the test set and are used as ground-truth data to measure the overall misalignment for registration. The registration accuracy can be estimated by the target registration error (TRE), which indicates the average squared Euclidean distance between the fixed, chosen landmarks and their matching correspondences in the deformed images (moving landmarks). The TRE is specified as follows:
TRE = 1 k i = 1 k ( x l p , f , y l p , f ) ( x l p , m , y l p , m ) 2 .
Furthermore, we have annotated two types of landmarks in the images. The first landmark type is located on foreground objects in the images (e.g., the yellow points in Figure 5). The second landmark type is located on background objects in the scene (e.g., the red points in Figure 5). Both types are important because aligning foreground objects while maintaining the relative positions between foreground and background objects is a prerequisite for many object tracking applications.

2.5. Implementation Details

The code is implemented using Tensorflow 2.6.0. The network T is a U-Net-based encoder–decoder network with residual connections [34]. The instance normalization layer is used in all skip connections [35] and Kaiming [36] initialization is applied for all networks. The experiments are conducted on a single GeForce RTX 2080Ti. An Adam Optimizer [37] is used on a minibatch of size 4 with the learning rate l r = 1 × 10 4 . The model is trained for 200 epochs and linear learning-rate decay is activated after 100 epochs. Separate networks are trained with different λ and μ regularization values until convergence occurs and we report results on the test set, which is separated and different from the training set.

2.6. Experiments

The proposed method focuses on computing a registration field across pens and over time. The dataset contains 6291 image pairs, including 2087 image pairs in each of the A1, A2, and A3 pens and 30 image pairs of the A4 pen. Each image pair consist of two 24-bit RGB video frames captured by high-resolution sensors. Prior to the global alignment phase, images are calibrated, masked, and grey-scaled to the resulting input frames with a size of 928 × 1408 pixels. The initial image registration is based on the projective transformation, with 35 manually placed landmarks, also considered as the baseline for the experiments. However, some non-linear misalignment in the dataset also occurs, which cannot be handled by the initial (rigid) registration. For each of the A1, A2, and A3 pens, we randomly split the data into 2057, 20, and 10 image pairs for training, fine-tuning, and testing, respectively. In order to verify the generalization ability of the models on registering image pairs from a new pen, the image pairs of A4 are only used for testing. In 3 fine-tuning sets and 4 testing sets, 6 landmark pairs are placed in the image pairs: 3 landmark pairs are on the animals, while the other 3 are located in the background.
When training the unsupervised network, 6171 non-annotated training image pairs from A1, A2, and A3 are supplied to the network. Each input pair consists of a fixed image and a moving image. Whereas the semi-supervised network is only trained using annotated samples of the fine-tuning sets, the input of the semi-supervised network has an additional layer for landmark coordinates in the moving images. In the fine-tuning session, weights obtained after unsupervised learning, which are able to extract general features across pig pens and distributions, are adopted and inserted into the semi-supervised framework. The TRE definition in Equation (8) of the test image pairs is used to evaluate the performance of the models on the seen and unseen pens. The registration accuracy is improved by regularizing the optimization process using the actual average squared misalignment distance represented by the TRE.

3. Results

3.1. Quantitative Evaluation

We have compared the average registration error in terms of TRE for different registration networks with the baseline method of only global projective transformation in Table 3. Specifically, in addition to the mean-squared error (MSE), the other two commonly used loss terms are also involved, i.e., normalized cross-correlation (NCC) and the sum of squared differences (SSD). These commonly used loss terms NCC and SSD are inserted into the training optimization by maximizing and minimizing, respectively. We also compare the involved networks with feature-based and intensity-based approaches at the bottom of the Table 3. All methods are evaluated on the same test sets. The first baseline method considered is speeded-up robot features (SURF) [38]. It is invariant to changes in scale, rotation, and illumination and has proven to be much faster than the scale invariant feature transform (SIFT) [39]. Registration is based on matching features represented by these descriptors. First, SURF descriptors are extracted from the calibrated, resized, and masked RGB image pairs. Then, these features are matched and estimated with the 2D geometric transformation for matching point pairs, using an affine transformation. On this basis, we also apply a non-rigid transformation, specifying a pyramid vector [100, 50, 25] for iterative computation. A Gaussian smoothing filter is used to harmonize the accumulated field at each iteration. Deformation fields are calculated through a non-parametric diffeomorphic image registration algorithm [40] and re-sampled using linear interpolation. The second method is the mono-modal intensity-based approach. The method uses affine transformation for global alignment. Gradient magnitude tolerance, relaxation factor, and maximum iterations are set to 1 × 10 4 , 200, and 0.5, respectively. The mean-squared error is used in the loss function to control the minimization of the total error to optimize the similarity of the images. For the baseline methods, the post-processing method for nonlinear transformation is the same as used in the previous feature-based method.
As can be observed from Table 3, all networks achieve better results than the baseline methods, except for the semi-supervised network. Two values are reported, one for the measured error on the background (left) and the second based on foreground annotations (right). The SURF-based method performs comparable to the mono-modal intensity-based method on average TRE, and both methods are 3.02 and 2.53 pixels lower in average TRE than the projective transformation. They also have comparable TRE values on foreground and background landmarks. The unsupervised model trained with MSE outperforms the networks trained with NCC and SSD. On one hand, it has a lower average TRE value (33.31 pixels). On the other hand, it has a lower error on both the task of aligning foreground animals and aligning background. The table also shows that a fine-tuned network using MSE as the loss term can reduce the average TRE from 40.12 pixels to 22.95 pixels, yielding the most significant improvement over other frameworks, while the TRE metrics of the foreground and background registration are lower than that of the unsupervised network. However, the semi-supervised network leads to degraded forms of registration. The semi-supervised networks achieve the most significant registration error compared with the unsupervised networks and the fine-tuned networks when the same loss term is used.
The distributions of the TRE score for each method are visualized as box plots. For comparison purposes, the foreground and background registration performances of each technique on three different test sets (seen pens, unseen pens, and all pens) are shown in Figure 6. As can be observed, the SURF feature-based and mono-modal intensity-based methods achieve comparable TRE measures to the baseline on seen-pen and unseen-pen test sets. Both perform slightly better than the baseline on the test sets in all pens. Models based on unsupervised learning and fine-tuning strategies correct more errors in registration on the seen-pen test sets than on the unseen-pen test set. The fine-tuned training network gives the best registration accuracy on each test set. It performs better in terms of low TRE when aligning the background than when aligning the foreground.
Since the set size for testing is less than 2000, the Shapiro–Wilk test results show that only the TRE values calculated from the image pairs registered by the unsupervised and the fine-tuned network on the test set of unseen pens conform to the normal distribution (p > 0.05). We use the Wilcoxon sign rank test and T-test to evaluate whether these methods are significantly improved compared to the baseline method. SURF feature-based and mono-modal intensity-based methods show no significant improvement over the projective method on the test sets (p > 0.05). In contrast, both the unsupervised and the fine-tuned methods have significant differences over the projective method on the test sets (p < 0.05). Furthermore, the TRE values of the fine-tuned model are significantly lower than the numbers of the unsupervised model on the three test sets (p > 0.05).

3.2. Qualitative Evaluation

Figure 7a shows that the proposed registration networks successfully align image pairs from real pig pens. In the first row of Figure 7a, the unsupervised method aligns a pair of images. As seen from the visualization of the deformation field, minor deformation is applied to the ground area of the current pen (see the black centre in the deformation field). At the same time, most deformation is applied to the body of the pigs and visual field boundaries. Furthermore, in the second row in Figure 7a, the fine-tuned network is used to deform the same moving image. The deformation field calculated by the fine-tuned network is similar to the field of the first row. Although it is similar, it will result in a better registration alignment. To visualize the full alignment within this image pair, we overlay (with semi-transparency) the whole scene of the fixed image on top of both moving images before and after the deformable registrations. Grey regions in the composite image show where the two images have the same intensities. Magenta and green regions show where the intensities are originating from moving and fixed images, respectively. The overlay of fixed and registered images shows the most grey-scale values, highlighting the higher quality of both images.
Figure 7b shows the transformation results using deformable fields for landmarks. Two landmark pairs are visualized on a fixed image and its corresponding registered image. The landmark pair that relates to the background is denoted as 1, the landmark pair that is a key point of the pigs is marked as 2. Landmarks on the fixed image are filled with green while landmarks on the registered image are yellow. In the first row in Figure 7b, we use the deformation field calculated by the unsupervised network to transform the landmarks, while in the second row in Figure 7b, the same image pair and the same landmark pairs are used to show the alignment results of the fine-tuned network. As can be observed, the transformation from the fine-tuned network brings the corresponding landmarks closer.
A final check is obtained on the visual quality of the registration. It is important that the skin or other details of the animals are preserved after registration. Therefore, an example is shown here. As can be observed from Figure 8, the pigtail is not visible in the fixed image, while the moving image captures it. After registration, the pigtail is preserved in the registered image. Spatial alignment of image content while maintaining unique appearance features provides a good starting point for subsequent tasks such as object detection, tracking, and behaviour recognition.

3.3. Execution Time Analysis

Table 4 presents the execution time results using an i7- 7700K CPU core and a GeForce RTX-2080Ti GPU. The elapsed execution time is measured for computations following the projective alignment (pre-registering step), which is shared by all presented methods. To our knowledge, SURF-based and mono-modal intensity-based registrations have no complete implementation for GPU-based execution. The SURF method has a registration time of 8 s on the CPU, while the mono-modal intensity-based approach requires about 28 s on the CPU. The presented learning-based methods are more than 10 times faster in execution on a GPU than on a CPU for the registration task. When registering a new image pair on GPU, the time consumption is limited to 0.5 s.

4. Conclusions and Discussion

This work proposes a semi-supervised, fine-tuned learning algorithm that can be used to register double-camera videos for animal surveillance in real farming. Considering the complexity and the time investment for annotation, the proposed method only requires a few image pairs with landmarks to obtain an accurate deformation field. Fine-tuning is obtained as follows. First, an unsupervised network is trained with unlabelled data. Then the network weights are adopted to the semi-supervised network, after which the semi-supervised network is fine-tuned with training data using annotated images. The fine-tuned network has demonstrated an average registration error (TRE value) of 22.95 pixels, which is 17.17, 10.36, 22.72, 14.02, and 14.51 pixels lower than the baseline, unsupervised, semi-supervised, SURF feature-based, and mono-modal intensity-based methods, respectively. The above results show that the knowledge learned in the unsupervised registration process can be well transferred to the semi-supervised learning framework, and no negative transfer occurs.
One of the most important tasks in animal surveillance is to track moving objects independently, using videos from the double-camera system. In this dataset, pigs are about 240 ± 37 pixels long and 125 ± 41 pixels wide. Since the average TRE is about 22 pixels in the proposed method, the deformation field is sufficient to move the appearance pixels of the same object from two viewpoints into the same or adjacent receptive field in the tracking network.
The statistical result shows that the unsupervised and fine-tuned networks can significantly reduce the TRE compared to the baseline on all test sets. Furthermore, a statistically significant improvement has also been detected in the fine-tuned network over the unsupervised network on the unseen test set using the T-test. On one hand, the slightly higher foreground TRE value is due to the lack of depth information. On the other hand, the foreground misalignment error is already more significant than the background errors after pre-registration.
The proposed method can register the targets recorded using different camera positions into the same coordinate system. In other words, by overlapping the registered image pairs together, image information from two perspectives can be presented in the target space region, which is helpful for multi-object tracking, re-identification, and behaviour-recognition tasks. In addition, a more promising point is that it provides a new idea for alleviating the occlusion problem in single-camera systems, because even if an object is completely occluded in a certain view, after registration and fusion, there will be appearance information available from another perspective at the position where it should appear, so that the model can perceive the presence of the same object.
The end-to-end strategy of the network enables the estimation of a deformation field for transforming the entire image, with a displacement vector for each pixel. However, it should be noted that the images registered by the proposed method retain their unique appearance features recorded from the side-view camera. Such features are very friendly to downstream tasks such as re-identification because the proposed method can integrate unique appearance features of objects in the same local space, which is very beneficial for distinguishing different objects for analysis at the individual level.
The semi-supervised model, which learns from scratch, does not show comparable performance in the experiment, because it is based on a limited number of training samples. In the near future, data augmentation techniques should be adopted to expand the variety of the training sample without adding extra manual effort.
The expensive challenge of registering a new image pair has been converted into an optimization problem, aggregated over a set of samples. Following the experimental setup, the model takes 2 days to compute on a single GeForce RTX-2080Ti GPU to obtain the reported parameters. After this, a new image pair can be registered within 0.5 s. The current solution is processing on a frame-by-frame basis. In the future, we may explore whether a common deformation field can be found to register two videos of the same pen and provide online processing.
In the current implementation, the input image pairs need to be pre-aligned to obtain a dense deformation field with good registration performance. A future research topic is constructing a whole learnable registration framework, which involves pre-registration as part of global optimization.
The original U-Net simply concatenates the features of the down-sampling layer directly into the up-sampling layer of the same depth. We can first use the attention module to process the feature maps of the down-sampling layer and the feature maps of the upper layer of the corresponding up-sampling layer and then concatenate them with the current up-sampled feature map. Thereby, the attention of the model will focus on the object region (foreground).
Although the proposed method has been verified to be effective on real data, the models do not emphasize constraints on maintaining geometric properties. Offering regularization to preserve geometric properties is one of the priorities of future work.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/foods12010084/s1.

Author Contributions

Conceptualization, Y.S., T.T. and S.H.; methodology, Y.S., T.T. and S.H.; software, S.H. and J.E.B.; validation, S.H., P.B. and J.E.B.; formal analysis, S.H. and P.B.; investigation, S.H. and Q.G.; resources, P.H.N.d.W. and Y.S.; data curation, S.H. and Q.G.; writing—original draft preparation, S.H.; writing—review and editing, P.H.N.d.W. and S.H.; visualization, S.H.; supervision, P.H.N.d.W. and Y.S.; project administration, P.H.N.d.W. and Y.S.; funding acquisition, P.H.N.d.W., Y.S. and S.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the project IMAGEN (P18-19 project 1) of the research programme Perspectief, which is financed by the Dutch Research Council (NWO). This work was funded by the program of China Scholarship Council (No. 202106930006).

Institutional Review Board Statement

The statement from the Animal Welfare Officer is in the supplementary file.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This work was supported in part by the Dutch NWO project IMAGEN (P18-19 Project 1) of the research programme Perspectief. This work was supported in part by the program of China Scholarship Council (No. 202106930006). The Volmer facility in Germany was offered by Topigs Norsvin in Helvoirt, the Netherlands. The author would like to thank all the participants in the project.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Qiao, Y.; Kong, H.; Clark, C.; Lomax, S.; Su, D.; Eiffert, S.; Sukkarieh, S. Intelligent perception-based cattle lameness detection and behaviour recognition: A review. Animals 2021, 11, 3033. [Google Scholar] [CrossRef] [PubMed]
  2. Shirke, A.; Saifuddin, A.; Luthra, A.; Li, J.; Williams, T.; Hu, X.; Kotnana, A.; Kocabalkanli, O.; Ahuja, N.; Green-Miller, A.; et al. Tracking Grow-Finish Pigs Across Large Pens Using Multiple Cameras. arXiv 2021, arXiv:2111.10971. [Google Scholar]
  3. Tilman, D.; Balzer, C.; Hill, J.; Befort, B.L. Global food demand and the sustainable intensification of agriculture. Proc. Natl. Acad. Sci. USA 2011, 108, 20260–20264. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Collins, L.; Smith, L. Smart agri-systems for the pig industry. Animal 2022, 8, 100518. [Google Scholar] [CrossRef] [PubMed]
  5. Li, G.; Huang, Y.; Chen, Z.; Chesser Jr, G.D.; Purswell, J.L.; Linhoss, J.; Zhao, Y. Practices and applications of convolutional neural network-based computer vision systems in animal farming: A review. Sensors 2021, 21, 1492. [Google Scholar] [CrossRef]
  6. Chen, C.; Zhu, W.; Steibel, J.; Siegford, J.; Han, J.; Norton, T. Classification of drinking and drinker-playing in pigs by a video-based deep learning method. Biosyst. Eng. 2020, 196, 1–14. [Google Scholar] [CrossRef]
  7. Wang, S.; Jiang, H.; Qiao, Y.; Jiang, S.; Lin, H.; Sun, Q. The Research Progress of Vision-Based Artificial Intelligence in Smart Pig Farming. Sensors 2022, 22, 6541. [Google Scholar] [CrossRef]
  8. Symeonaki, E.; Arvanitis, K.G.; Piromalis, D.; Tseles, D.; Balafoutis, A.T. Ontology-Based IoT Middleware Approach for Smart Livestock Farming toward Agriculture 4.0: A Case Study for Controlling Thermal Environment in a Pig Facility. Agronomy 2022, 12, 750. [Google Scholar] [CrossRef]
  9. Benjamin, M.; Yik, S. Precision livestock farming in swine welfare: A review for swine practitioners. Animals 2019, 9, 133. [Google Scholar] [CrossRef] [Green Version]
  10. Cowton, J.; Kyriazakis, I.; Bacardit, J. Automated individual pig localisation, tracking and behaviour metric extraction using deep learning. IEEE Access 2019, 7, 108049–108060. [Google Scholar] [CrossRef]
  11. Wu, H.; Zhang, X.; Story, B.; Rajan, D. Accurate vehicle detection using multi-camera data fusion and machine learning. In Proceedings of the ICASSP 2019–2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 3767–3771. [Google Scholar]
  12. He, W.; Tao, W.; Sun, K.; Xu, L.; Fu, Q.; Zhao, H. Multi-camera object tracking via deep metric learning. In Proceedings of the Sixth International Conference on Optical and Photonic Engineering (icOPEN 2018), SPIE, Shanghai, China, 8–11 May 2018; Volume 10827, pp. 331–339. [Google Scholar]
  13. Ambinder, E.P. A history of the shift toward full computerization of medicine. J. Oncol. Pract. 2005, 1, 54. [Google Scholar] [CrossRef] [PubMed]
  14. Simonovsky, M.; Gutiérrez-Becker, B.; Mateus, D.; Navab, N.; Komodakis, N. A deep metric for multimodal registration. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention: 19th International Conference, Athens, Greece, 17–21 October 2016; pp. 10–18. [Google Scholar]
  15. Cheng, X.; Zhang, L.; Zheng, Y. Deep similarity learning for multimodal medical images. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2018, 6, 248–252. [Google Scholar] [CrossRef]
  16. Blendowski, M.; Heinrich, M.P. Combining MRF-based deformable registration and deep binary 3D-CNN descriptors for large lung motion estimation in COPD patients. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 43–52. [Google Scholar] [CrossRef] [PubMed]
  17. Haskins, G.; Kruecker, J.; Kruger, U.; Xu, S.; Pinto, P.A.; Wood, B.J.; Yan, P. Learning deep similarity metric for 3D MR–TRUS image registration. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 417–425. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Lee, J.G.; Jun, S.; Cho, Y.W.; Lee, H.; Kim, G.B.; Seo, J.B.; Kim, N. Deep learning in medical imaging: General overview. Korean J. Radiol. 2017, 18, 570–584. [Google Scholar] [CrossRef] [Green Version]
  19. Chee, E.; Wu, Z. Airnet: Self-supervised affine registration for 3d medical images using neural networks. arXiv 2018, arXiv:1810.02583. [Google Scholar]
  20. Rohé, M.M.; Datar, M.; Heimann, T.; Sermesant, M.; Pennec, X. SVF-Net: Learning deformable image registration using shape matching. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, 20th International Conference, Quebec City, QC, Canada, 11–13 September 2017; pp. 266–274. [Google Scholar]
  21. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, 18th International Conference, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  22. Lorenzi, M. Deformation-Based Morphometry of the Brain for the Development of Surrogate Markers in Alzheimer’s Disease. Ph.D. Thesis, Université Nice Sophia Antipolis, Nice, France, 2012. [Google Scholar]
  23. De Vos, B.D.; Berendsen, F.F.; Viergever, M.A.; Sokooti, H.; Staring, M.; Išgum, I. A deep learning framework for unsupervised affine and deformable image registration. Med Image Anal. 2019, 52, 128–143. [Google Scholar] [CrossRef] [Green Version]
  24. Klein, S.; Staring, M.; Murphy, K.; Viergever, M.A.; Pluim, J.P. Elastix: A toolbox for intensity-based medical image registration. IEEE Trans. Med Imaging 2009, 29, 196–205. [Google Scholar] [CrossRef]
  25. Hering, A.; Kuckertz, S.; Heldmann, S.; Heinrich, M.P. Enhancing label-driven deep deformable image registration with local distance metrics for state-of-the-art cardiac motion tracking. In Bildverarbeitung für die Medizin 2019, Algorithms-Systems-Applications, Proceedings of the Workshop from March 17th to 19th, 2019 in Lübeck; Springer: Berlin/Heidelberg, Germany, 2019; pp. 309–314. [Google Scholar]
  26. Rühaak, J.; Heldmann, S.; Kipshagen, T.; Fischer, B. Highly accurate fast lung CT registration. In Proceedings of the Medical Imaging 2013: Image Processing, SPIE, Lake Buena Vista, FL, USA, 9–14 February 2013; Volume 8669, pp. 255–263. [Google Scholar]
  27. Nazib, A.; Fookes, C.; Perrin, D. A comparative analysis of registration tools: Traditional vs deep learning approach on high resolution tissue cleared data. arXiv 2018, arXiv:1810.08315. [Google Scholar]
  28. Crum, W.R.; Hill, D.L.; Hawkes, D.J. Information theoretic similarity measures in non-rigid registration. In Proceedings of the Biennial International Conference on Information Processing in Medical Imaging, 18th International Conference, IPMI 2003, Ambleside, UK, 20–25 July 2003; pp. 378–387. [Google Scholar]
  29. Rohlfing, T. Image similarity and tissue overlaps as surrogates for image registration accuracy: Widely used but unreliable. IEEE Trans. Med Imaging 2011, 31, 153–163. [Google Scholar] [CrossRef] [Green Version]
  30. Yukun, S.; Pengju, H.; Yujie, W.; Ziqi, C.; Yang, L.; Baisheng, D.; Runze, L.; Yonggen, Z. Automatic monitoring system for individual dairy cows based on a deep learning framework that provides identification via body parts and estimation of body condition score. J. Dairy Sci. 2019, 102, 10140–10151. [Google Scholar] [CrossRef] [PubMed]
  31. Alameer, A.; Kyriazakis, I.; Bacardit, J. Automated recognition of postures and drinking behaviour for the detection of compromised health in pigs. Sci. Rep. 2020, 10, 1–15. [Google Scholar]
  32. Jaderberg, M.; Simonyan, K.; Zisserman, A. Spatial transformer networks. Adv. Neural Inf. Process. Syst. 2015, 28, 2017–2026. [Google Scholar]
  33. Balakrishnan, G.; Zhao, A.; Sabuncu, M.R.; Guttag, J.; Dalca, A.V. An unsupervised learning model for deformable medical image registration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 9252–9260. [Google Scholar]
  34. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. In Proceedings of the European Conference on Computer Vision, 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; pp. 694–711. [Google Scholar]
  35. Ulyanov, D.; Vedaldi, A.; Lempitsky, V. Instance normalization: The missing ingredient for fast stylization. arXiv 2016, arXiv:1607.08022. [Google Scholar]
  36. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE INTERNATIONAL conference on Computer Vision, Boston, MA, USA, 7–12 June 2015; pp. 1026–1034. [Google Scholar]
  37. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  38. Bay, H.; Ess, A.; Tuytelaars, T.; Van Gool, L. Speeded-up robust features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  39. Lowe, D.G. Object recognition from local scale-invariant features. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; Volume 2, pp. 1150–1157. [Google Scholar]
  40. Vercauteren, T.; Pennec, X.; Perchant, A.; Ayache, N. Diffeomorphic demons: Efficient non-parametric image registration. NeuroImage 2009, 45, S61–S72. [Google Scholar] [CrossRef]
Figure 1. Pig pen layout and lens positions.
Figure 1. Pig pen layout and lens positions.
Foods 12 00084 g001
Figure 2. Image pairs at various stages of pre-processing.
Figure 2. Image pairs at various stages of pre-processing.
Foods 12 00084 g002
Figure 3. The proposed method consists of two stages. First, train the unsupervised network with image pairs without landmarks to obtain the weights across pig pens and over time; second, import these weights into the semi-supervised network and fine-tune with samples labelled with landmarks.
Figure 3. The proposed method consists of two stages. First, train the unsupervised network with image pairs without landmarks to obtain the weights across pig pens and over time; second, import these weights into the semi-supervised network and fine-tune with samples labelled with landmarks.
Foods 12 00084 g003
Figure 4. The architecture of the proposed network.
Figure 4. The architecture of the proposed network.
Foods 12 00084 g004
Figure 5. Examples of two types of landmarks and their correspondences. Red dots indicate background objects and yellow dots indicate the foreground.
Figure 5. Examples of two types of landmarks and their correspondences. Red dots indicate background objects and yellow dots indicate the foreground.
Foods 12 00084 g005
Figure 6. Boxplots of TRE scores of each image registration method for aligning the background and the foreground objects. For each method a cluster of three results are grouped: seen pens, unseen pen, and all pens.
Figure 6. Boxplots of TRE scores of each image registration method for aligning the background and the foreground objects. For each method a cluster of three results are grouped: seen pens, unseen pen, and all pens.
Foods 12 00084 g006
Figure 7. Illustrations of non-linear dense registration effects and errors of un-supervised and fine-tuned networks in surveillance of a pen with pigs.
Figure 7. Illustrations of non-linear dense registration effects and errors of un-supervised and fine-tuned networks in surveillance of a pen with pigs.
Foods 12 00084 g007
Figure 8. The proposed method retains the appearance features specific to moving images while aligning moving images to fixed images.
Figure 8. The proposed method retains the appearance features specific to moving images while aligning moving images to fixed images.
Foods 12 00084 g008
Table 1. Specifications of the down-sampling convolution filters.
Table 1. Specifications of the down-sampling convolution filters.
Down-Sampling Layers12345
Kernel number3264128256512
Kernel size 3 × 3 3 × 3 3 × 3 3 × 3 3 × 3
Activation functionReLUReLUReLUReLUReLU
PaddingSame paddingSame paddingSame paddingSame paddingSame padding
Stride ( s h , s v )(1, 1)(1, 1)(1, 1)(1, 1)(1, 1)
Pooling(2, 2)(2, 2)(2, 2)(2, 2)(2, 2)
Table 2. Specifications of the up-sampling convolution filters.
Table 2. Specifications of the up-sampling convolution filters.
Up-Sampling Layers12345
Up-sampling (deconvolution) 2 × 2 2 × 2 2 × 2 2 × 2 -
Kernel number (deconvolution)2561286432-
Kernel size (deconvolution) 2 × 2 2 × 2 2 × 2 2 × 2 -
Activation function (deconvolution)ReLUReLUReLUReLU-
ConcatenateYesYesYesYes-
Kernel number (convolution)25612864322
Kernel size (convolution) 3 × 3 3 × 3 3 × 3 3 × 3 1 × 1
Activation function (convolution)ReLUReLUReLUReLUReLU
Padding (convolution)Same paddingSame paddingSame paddingSame paddingNo padding
Stride ( s h , s v ) (convolution)(1, 1)(1, 1)(1, 1)(1, 1)(1, 1)
Table 3. Measured TRE values for the proposed registration networks and baseline registration methods (NCC: normalized cross-correlation; SSD: sum of squared differences; left metric number: background; right metric number: foreground).
Table 3. Measured TRE values for the proposed registration networks and baseline registration methods (NCC: normalized cross-correlation; SSD: sum of squared differences; left metric number: background; right metric number: foreground).
MethodsMSENCCSSDOptimal TREAvg. TRE
Projective transformation (proposed)---36.01/44.2240.12
Unsupervised learning network30.92/35.7032.24/36.6633.48/38.9630.92/35.7033.31
Semi-supervised learning network48.46/47.8939.13/52.2241.06/57.2939.13/52.2245.67
Fine-tuned learning network20.55/25.3625.77/28.6722.08/28.3020.55/25.3622.95
Feature-based registration (SURF)---38.09/35.8636.98
Mono-modal intensity-based registration35.11/39.81--35.11/39.8137.46
Table 4. Average execution times (and std. devs.) on CPU and GPU for the proposed registration networks and baseline methods.
Table 4. Average execution times (and std. devs.) on CPU and GPU for the proposed registration networks and baseline methods.
MethodsCPUGPU
Unsupervised learning network1.52 (2.26)0.12 (0.13)
Semi-supervised learning network1.48 (4.18)0.10 (0.06)
Fine-tuned learning network1.51 (2.77)0.13 (0.17)
SURF-based registration8.20 (1.36)-
Mono-modal intensity-based registration28.51 (19.65)-
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huo, S.; Sun, Y.; Guo, Q.; Tan, T.; Bolhuis, J.E.; Bijma, P.; de With, P.H.N. Double-Camera Fusion System for Animal-Position Awareness in Farming Pens. Foods 2023, 12, 84. https://doi.org/10.3390/foods12010084

AMA Style

Huo S, Sun Y, Guo Q, Tan T, Bolhuis JE, Bijma P, de With PHN. Double-Camera Fusion System for Animal-Position Awareness in Farming Pens. Foods. 2023; 12(1):84. https://doi.org/10.3390/foods12010084

Chicago/Turabian Style

Huo, Shoujun, Yue Sun, Qinghua Guo, Tao Tan, J. Elizabeth Bolhuis, Piter Bijma, and Peter H. N. de With. 2023. "Double-Camera Fusion System for Animal-Position Awareness in Farming Pens" Foods 12, no. 1: 84. https://doi.org/10.3390/foods12010084

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop