Next Article in Journal
Exploring Breed-Specific Milk Coagulation in Spanish Dairy Sheep: A Canonical Correlation Approach
Next Article in Special Issue
Deconvolution Enhancement Keypoint Network for Efficient Fish Fry Counting
Previous Article in Journal
Mitochondrial Variation of Bottlenose Dolphins (Tursiops truncatus) from the Canary Islands Suggests a Key Population for Conservation with High Connectivity within the North-East Atlantic Ocean
Previous Article in Special Issue
Automatic Recognition and Quantification Feeding Behaviors of Nursery Pigs Using Improved YOLOV5 and Feeding Functional Area Proposals
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning Rich Feature Representation and State Change Monitoring for Accurate Animal Target Tracking

1
School of Computer Science and Technology, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
2
College of Artificial Intelligence and Big Data, Chongqing College of Electronic Engineering, Chongqing 401331, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Animals 2024, 14(6), 902; https://doi.org/10.3390/ani14060902
Submission received: 29 December 2023 / Revised: 7 February 2024 / Accepted: 12 March 2024 / Published: 14 March 2024

Abstract

:

Simple Summary

Animal movement trajectories are effective indicators of key information such as social behavior, food acquisition, reproduction, migration, and survival strategies in animal behavior analysis. However, manual observation is still relied upon in many analysis scenarios, which is inefficient and error-prone. This paper introduces a computer vision-based method for tracking animal trajectories, which enables monitoring and accurate acquisition of individual target animal movement trajectories over extended periods, overcoming the limitations of manual observation. The experiments demonstrate that the method is efficient and accurate in tracking animals in complex scenes, providing essential basic data for animal behavior analysis and having a wide range of potential applications.

Abstract

Animal tracking is crucial for understanding migration, habitat selection, and behavior patterns. However, challenges in video data acquisition and the unpredictability of animal movements have hindered progress in this field. To address these challenges, we present a novel animal tracking method based on correlation filters. Our approach integrates hand-crafted features, deep features, and temporal context information to learn a rich feature representation of the target animal, enabling effective monitoring and updating of its state. Specifically, we extract hand-crafted histogram of oriented gradient features and deep features from different layers of the animal, creating tailored fusion features that encapsulate both appearance and motion characteristics. By analyzing the response map, we select optimal fusion features based on the oscillation degree. When the target animal’s state changes significantly, we adaptively update the target model using temporal context information and robust feature data from the current frame. This updated model is then used for re-tracking, leading to improved results compared to recent mainstream algorithms, as demonstrated in extensive experiments conducted on our self-constructed animal datasets. By addressing specific challenges in animal tracking, our method offers a promising approach for more effective and accurate animal behavior research.

1. Introduction

Animal tracking [1] is a pivotal technology extensively employed in ecology, zoology, and environmental science, and it can help provide insights into wildlife behavior, migration patterns, territory utilization, and interactions within ecosystems. Moreover, it extends into various domains such as resource management, wildlife conservation, and the prediction of disease spread, and these applications provide crucial additional information about the behavior of plants and animals and the health of ecosystems. This technology has progressed significantly over the past few decades, covering many species from small birds to large mammals. Currently, animal tracking technology comprises a spectrum of methods, incorporating radio tracking, satellite tracking, GPS positioning, and approaches based on computer vision. The ongoing advancement of these technologies empowers researchers to systematically and accurately track the movement trajectories of animals, thereby acquiring data pertinent to ecology and physiology.
In recent years, intelligent camera systems have propelled the increasing prominence of computer vision-based animal tracking technology, which collects animal video data by deploying smart cameras and then processes the video data using computer vision methods to track specific animal targets; this process provides foundational data for subsequent analyses of animal behavior. This approach eliminates the need for capturing animals, avoiding the installation of sensors on the animals to collect data. It minimizes interference with the animals to the greatest extent, allowing for data collection in a non-intrusive manner. Moreover, it can collect data for all animals within the camera’s field of view, reducing the cost associated with data acquisition. However, animal tracking technology still faces a number of challenges due to the influences inherent in acquiring video data in unfavorable environmental conditions.
In this paper, we propose a novel computer vision-based animal tracking method (RFCFT) that achieves robust animal trajectory tracking through a correlation filter framework with response map analysis and adaptive feature updating; we utilize the following strategies to achieve high-quality tracking of animals: since animals undergo continuous morphological and scale changes during locomotion, and there are similarities between their appearance and the field environment, a single feature is difficult to represent the animal model robustly, and we constructed two fusion features by simultaneously extracting different layers of deep features and hand-crafted Histogram of Oriented Gradients (HOG) feature in different combinations, which were used to adapt to the complex appearance changes during animal locomotion. In the tracking stage, the two fusion features are used to track the target separately. Subsequently, we analyze the response map of the tracking results and select the tracking outcome with the optimal fusion feature based on the oscillation level of the response map as the predicted target position for the current frame. We analyze the tracking results for each frame and update the model if the results are deemed unreliable. Considering that the motion process of the target varies continuously in the time series, we combine temporal context information with the robust features of the current frame to obtain features for the updated model. Finally, we employ the updated model for re-tracking to yield the ultimate result.
To summarize, the main contributions of this paper are as follows.
We propose an adaptive feature fusion mechanism to construct an optimal animal representation model that blends the semantic information and discriminative advantages of hand-crafted features and multiple layers of deep features to improve the representation of features in complex environments and enhance tracking robustness.
We reformulate the reliability assessment methods. By analyzing the response map, we actively employ two critical indicators within the response map, namely oscillation amplitude and peak variation rate, to comprehensively evaluate the reliability of the tracking results for the current frame, and to promptly monitor the changing state of the target animal.
We have devised a re-tracking model that integrates temporal context information of the target, along with the target’s robust feature and shallow features in the current frame. This model is deployed for re-tracking when the tracking results are deemed unreliable.
Due to the lack of specialized animal tracking datasets, we collected specific animal data from popular datasets. We also obtained partial animal videos from the internet and annotated them ourselves, creating dedicated animal tracking datasets. We further validated our proposed method on this animal dataset, and experimental results confirmed the effectiveness of our tracker.

2. Related Works

Traditional visual object tracking methods can be classified into generative and discriminative approaches. However, with the growing interest and development of machine learning techniques in recent years, a variety of machine learning methods are applied in visual object tracking. Examples comprise support vector machines (SVM) [2], correlation filter [3], convolutional neural networks (CNN) [4], recurrent neural networks (RNN) [5], and Siamese neural networks [6]. Hence, the principal research approaches in the visual object tracking field in recent years can be generally divided into two main categories: those based on the correlation filter approach and those using the deep learning approach.
The concept of a correlation filter originated in the field of signal processing. Correlation is an intuitive means to describe the similarity between two variables in signal processing. Leveraging this principle, a filter template is trained using a correlation filter in the context of target tracking. This template is then utilized for executing correlation-based calculations with the target region. Consequently, the current frame’s object position is identified as that corresponding to the maximum output response. Visual object tracking based on correlation filter has yielded significant performance and established itself as one of the most prominent frameworks in recent years.
In 2010, Bolme introduced the MOSSE [3], which paved the way for incorporating correlation filter principles into the visual object tracking domain and exhibited outstanding results. To avoid overfitting problems, CSK [7] introduces regularization terms and kernel methods to convert the linear regression problem into a kernel ridge regression problem [8], which improves the solution of the regression problem and increases the speed of computation. KCF [9] uses HOG [10] features to characterize the target, improving the target representation’s robustness. To estimate the target scale more flexibly, DSST [11] proposes a three-dimensional filter consisting of a one-dimensional scale filter and a two-dimensional position filter to achieve adaptive estimation of the target scale. A cyclic sampling technique is commonly used in correlation filter methods, where training samples are generated by cyclically shifting the original target. However, this practice often leads to boundary effects [12]. To address these boundary effects, SRDCF [13] implements a spatial regularization on the filter that significantly expands the search region, thereby preserving a greater amount of true target information. BACF [14] increases the number of true negative samples by expanding the search to crop small samples, AutoTrack [15] employs local and global information from the response maps to achieve adaptive spatio-temporal regularization. STTCF [16] incorporates a weight matrix into traditional spatial regularization and introduces two long and short-term regularization terms to enhance the performance of the tracker further.
Correlation filter tracking methods combined with deep learning have also been widely developed to improve tracking quality further. HCF [17] proposes to use a pre-trained convolutional neural network framework as a feature extractor, replacing traditional hand-crafted features with deep models that have high-level semantic features. C-COT [18] uses VGG [19] neural network to extract the target features, which effectively improves the robustness of the target features. TCNN [20] utilizes multiple CNN models and builds them into a tree structure for target tracking to enhance the reliability of the model. SINT [6] first proposed a Siamese network-based tracking algorithm, through offline training, to learn a matching function to be utilized for locating the target in subsequent frames. SiamFC [21] adopts the tracking method of offline training and online fine-tuning, which maintains tracking accuracy and improves tracking efficiency.
Benefiting from these works, we adopt correlation filtering as the matching mechanism, but use a different approach to feature selection than traditional correlation filtering tracking algorithms, taking advantage of the different characteristics of hand-crafted features and deep features to complement each other’s strengths, in order to ensure that we can provide a robust representation of a specific individual animal and improve the robustness of tracking. Also different from classical deep learning methods, we extract three deep features from different layers of the VGG network with the fully connected layer removed, and add temporal context information as a supplement when necessary to cope with the complex and variable pose changes of the animal and improve the tracking accuracy.

3. Overview

The algorithm is divided into five parts: (1) use of the labeled data in the first frame of the video sequence to train the correlation filter; (2) input of the target image into the VGG19 network without the fully connected layer, extraction of the deep features of the conv3-4 layer, conv4-4 layer, and conv5-4 layer, and extraction of the hand-crafted feature HOG; (3) use of the combination of full deep features and the combination of deep features and HOG feature to construct two fusion features, then selection of the best fusion features by measuring the average peak-to-correlation energy and the maximum peak of the response map; (4) the optimal fusion feature and the current frame image are input into the correlation filter for correlation calculation, and the predicted position is finally calculated by using the fast Fourier transform; and (5) for unreliable tracking results, updating of the model, building of a new model, and evaluation of the confidence of the updated tracking results to obtain the final tracking results. An overview of the algorithm is shown in Figure 1.

4. Method

4.1. DCF Framework

The tracking algorithm is based on the DCF framework. DCF is divided into three stages: training, model update, and detection. In the training stage, the filter is trained by a nonlinear regression equation:
min w i w x i y i 2 2 + λ w 2 2
where x i is the data matrix and λ is regularization parameter λ 0 . Using kernel methods to transform a nonlinear space into a linear space and introducing coefficient α to replace filter coefficients w to represent filter solutions within the dual vector space:
f x = w T x = i α i k x , x i
where k x , x i denotes the kernel function (commonly used kernel functions include Gaussian kernel and polynomial kernel), and f x denotes the response map. The filter can be obtained by the combined solution of (1) and (2):
α ^ = y ^ k x ^ x + λ
where  ^  denotes Fourier transform form in the frequency domain.
In the detection stage, the response map can be obtained using
R z = F 1 k ^ x z α ^
where z denotes the samples in the search region, ⊙ is the element-wise product, k ^ x y is the result of the kernel correlation operation between the search image block z and the target template x, and F 1 · is inverse Fourier transform.
In the model update stage, using linear interpolation to update template x and the filter coefficients α , calculated by
X ^ i = η X ^ i + 1 η X ^ i 1
α ^ i = η α ^ i + 1 η α ^ i 1
where η denote learning rate, X ^ i is the Fourier transform of the cyclic shift sample in i-th frame, α ^ i is the Fourier transform of the filter coefficient in i-th frame.

4.2. Feature Model

The continuous evolution of animals results in their colors and forms becoming increasingly similar to their surrounding environment, as shown in Figure 2. Moreover, the complex video capture environment significantly impedes the tracking task. As a result, single features are difficult to model robustly for tracked animals. Consequently, we used a multi-feature fusion approach to construct a feature model of the animal. HOG [10] features exhibit strong illumination invariance and robust resistance to geometric deformations, and HOG partially addresses the significant pose variations in animals and the interferences introduced by the complex field conditions in video images; as the layers of the convolutional neural network become deeper, the deeper features acquire richer semantics, thus enhancing the robustness of the features; nevertheless, this progression comes at the cost of gradually diminishing resolution, impacting the precision of localization. First, we extract the hand-crafted feature HOG, denoted as F H O G , then we extract deep features of diverse depths from layers conv3-4, conv4-4, and conv5-4 of the pre-trained VGG19 convolutional neural network model. These are denoted as F C O N V 3 4 , F C O N V 4 4 , and F C O N V 5 4 , respectively. We utilize various feature fusion methods to create two fusion features, consequently enhancing the representational capacity in complex tracking scenarios. The fusion feature is calculated as follows.
E A = ω 1 χ s + ω 2 χ m + ω 3 χ d
E B = ω 4 χ h + ω 5 χ m + ω 6 χ d
F C O N V 3 4 , F C O N V 4 4 and F C O N V 5 4 are the shallow, medium and deep features extracted from the convolutional neural network, denoted by χ s , χ m , χ d , respectively, F H O G is the hand-crafted feature HOG, denoted by χ h , and ω i represents the weight parameter.

4.3. Reliability Assessment

During tracking, we input the constructed fusion features into a location filter to calculate correlation, thereby obtaining the response map for the current frame. The response map is a Gaussian response map centered on the target, with the peak point on the response map indicating the predicted position of the target in the current frame. Except for the peak, the values of points at other positions in the response map are relatively similar but substantially deviate from the peak value, and a greater peak value corresponds to a more reliable point position. Simultaneously, the overall oscillation level of the response map can also reflect the reliability of the current result, a smaller oscillation level indicates a more reliable tracking result. We utilize both the maximum response peak value F m a x and the average peak-to-correlation energy (APCE) as integrated discriminative metrics to assess the quality of the tracking results. The calculation method is as follows.
A P C E = F m a x F m i n 2 m e a n w , h F w , h F m i n 2
where F m a x represents the peak value of the response map, F m i n represents the valley value of the response map, and F w , h represents the corresponding response value at position ( w , h ) . In tracking, we independently compute the F m a x and A P C E values for the two fusion features. When the F m a x and A P C E values calculated for the current frame meet the criteria outlined in Equations (10) and (11), this implies a high degree of reliability in the tracking results for both fusion features. Consequently, the tracking results for the current frame are considered trustworthy, and we select the best outcome from the two fusion features as the final result.
F m a x i - t h > α 1 × M e a n F m a x
A P C E i - t h > α 2 × M e a n A P C E
where F max i - t h represents the peak value of the i-th fusion feature, A P C E i - t h represents the A P C E value of the i-th fusion feature, α 1 and α 2 are adjustment factors, and M e a n F x denotes the mean value of the peaks in the response maps obtained from different fusion features, and M e a n A P C E denotes the average A P C E value of the response maps obtained from different fusion features; herein, the variables α 1 and α 2 represent empirical thresholds, with specific values assigned: α 1 = 0.8 and α 2 = 0.8. If the values of F m a x and A P C E do not meet the criteria defined in the formula above, indicating that the tracking results for the current frame lack reliability, the results for the current frame are discarded, and the model update module is initiated.

4.4. Model Update

Unreliable tracking results during the tracking process can adversely affect the tracking quality in subsequent frames, and the continuous accumulation of errors in each frame may ultimately lead to tracking failure. Thus, we propose a model update strategy that directly removes unreliable tracking results. Subsequently, employ a new tracking model to retrack the target in the current frame, with the dependable results obtained through this re-tracking process considered as the final target position for the current frame. In light of the continuity of changes in the target over the time series during tracking, we introduce a method to incorporate temporal context information of the target during the re-tracking phase. Integrating the temporal contextual information of the target animal with the robust features of the current frame, along with the shallow features possessing a specific resolution and semantic information in the current frame, collectively constructs the re-tracking feature model for the target animal. The re-tracking feature model is as follows:
E u p d a t e = ω 7 χ r - t i m e + ω 8 χ r - c u r r e n t + ω 9 χ s
where χ r - t i m e is the optimal tracking feature extracted in the case of robust tracking in a short time interval, χ r - c u r r e n r is the optimal fusion feature extracted from the current frame, χ s is the shallow deep feature of the conv3-3 layer of the current frame, and ω i is the weight parameter. An overview of the model update is shown in Figure 3.

4.5. Algorithm Flow

Combined with the above description of the key steps of our algorithm, the overall process of the algorithm is shown in Algorithm 1.
Algorithm 1 Proposed Tracking Algorithm
Input: Initial target position P t 1 = ( x t 1 , y t 1 , w t 1 , h t 1 ) , the correlation filter w
Output: Estimated object position
Repeat
  1:
Crop out the searching window in frame t centered at P t = ( x t , y t , w t , h t ) and extract HOG features and deep features;
  2:
Construct two fusion features;
  3:
For each sample compute response map R z using ( 4 ) ;
  4:
Estimate the new position ( x t , y t ) on response map;
  5:
Estimate the new scale of sample and extract feature;
  6:
Compute F m a x and A P C E using (9);
  7:
if Not satisfying formula ( 10 ) or ( 11 )  then
  8:
   Update the model using (12);
  9:
end if
   10:
Estimate new position ( x t , y t ) and scale of target;
   11:
Get the final position ( x t , y t ) of the current frame;
Until End of video sequence

5. Experiments

5.1. Experimental Environment

Our trackers are implemented in MATLAB 2022a on a computer with a 13th Gen Intel(R) Core(TM) i5-13490F 2.50 GHz CPU and 16GB RAM and NVIDIA GeForce 4070 GPU. The MatConvNet toolbox is used to extract the deep features from VGG-19. We use the LaSOT benchmark official tools to test the algorithm and the comparison algorithm in the same environment and analyze the performance of the algorithm using the One-Pass Evolution (OPE) standard; at the same time, seven popular tracking algorithms in recent years are used as a comparison: STRCF [22], DAHCF [23], DSARCF [24], BSTCF [25], AutoTrack [15], ARCF [26], SiamFC [21].

5.2. Datasets

Currently, most public datasets in object tracking primarily focus on tracking generic targets, with a conspicuous absence of specialized datasets for animal tracking. This situation presents a substantial challenge to the progress of animal tracking. To ascertain the robustness of our method, we established a dataset specifically designed for animal tracking. We extracted animal data from existing generic object tracking datasets, including OTB50 [27], OTB100 [28], and LaSOT [29], and supplemented our collection with animal videos obtained from the internet and conducted manual annotations to create our animal datasets. The datasets encompass twenty-nine animal categories, including bears, birds, tigers, elephants, cats, deer, and others. It includes 54 video sequences comprising more than ninety thousand frames. To ensure the effectiveness of the algorithm in real-world scenarios, the datasets we use are captured in authentic environments. The datasets include a broad spectrum of animal video data, spanning various video scenes such as wildlife habitats, domestic settings, public environments, and other scenarios. This contributes to validating the algorithm’s effectiveness in tracking animals across various scenarios. Figure 4 presents a selection of data from the datasets.
To comprehensively evaluate the performance of tracker, as with the LaSOT datasets, we label each sequence with 14 attributes, including out-of-view (OV), partial occlusion (POC), deformation (DEF), motion blur (MB), aspect ratio change (ARC), full occlusion (FOC), fast motion (FM), background clutter (BC), scale variation (SV), rotation (ROT), low resolution (LR), viewpoint change (VC), illumination variation (IV), and camera motion (CM). The attributions are defined in Table 1.

5.3. Evaluation Metrics

To objectively evaluate the effectiveness of the algorithm, we utilize precision and success score provided by the OTB tool as evaluation metrics. Precision is defined as the percentage of video frames in which the distance between the estimated center point of the target bounding box generated by the tracking algorithm and the ground-truth center point is less than a specified threshold. In our paper, the threshold is set at 20 pixels. The success score is obtained by calculating the overlap score (OS). If the OS for a specific frame exceeds the designated threshold, we consider that frame as successfully tracked. The overall success score is the percentage of frames identified as successful out of all frames. In our paper, the threshold is set at 0.5, given a tracked bounding box r t and the ground-truth bounding extent r 0 of a target object, the overlap score is defined as
S = r t t 0 r t t 0
where ∩ and ∪ represent the intersection and union operators, respectively, and · denotes the number of pixels in a region.

5.4. Experimental Process

In the experiment, all data are derived from real-world scenarios. The experimental procedure is illustrated in Figure 5, and during the operational process, the system simply requires the sequential input of video data frame by frame. In the initial frame, we manually delineated the individuals to be tracked. Subsequently, the system autonomously tracks each frame, extracts position data, and thereby obtains the motion trajectory. Due to the limitations of target detection, which can only identify the category of the target without accurately distinguishing specific individuals within a category, such as a particular deer in a herd; manual selection of the target is necessary when tracking specific individuals. After the initial frame, the system will automatically continue tracking the chosen target.

5.5. Ablation Study

To verify the impact of adaptive fusion feature selection and feature update on tracking performance used in this paper, we performed ablation experiments on the datasets. The experimental results are presented in Table 2, where RFCFT-NU represents the algorithm without feature update mechanism, and RFCFT-NF represents the algorithm without adaptive fusion feature selection. It can be observed that, after removing the feature update mechanism, the success rate decreased by 0.5%, and precision decreased by 1.4%. Similarly, without adaptive fusion feature selection, there was a 1.6% decrease in success rate and a 3.2% decrease in precision.

5.6. Quantitative Analysis

For performance evaluation, we compare the proposed RFCFT method with seven other state-of-the-art trackers on our constructed animal datasets. As shown in Figure 6, the proposed method RFCFT achieves the best precision and success scores 76.5% and 56.2%, respectively. However, due to the introduction of animal state monitoring and updating mechanisms in our algorithm compared to traditional algorithms, the complexity of our algorithm has correspondingly increased. Table 3 presents the precision scores, success scores, and fps data for all compared algorithms experimented on the animal dataset, where bold font denotes the optimal values, and underscored values denote the second-best results. Ultimately, our algorithm achieves a speed of approximately 1.915 fps, which still requires further improvement. In terms of algorithm precision, the DAHCF algorithm ranks second with a precision of 74.1%. Compared to DAHCF, our algorithm has improved precision by 2.4% and excels in both real-time performance and success rate. In terms of success rate, the algorithm ranking second is SiamFC, with a success rate of 55.5%. Compared to SiamFC, our algorithm has improved the success rate by 0.7%; despite the lower complexity of the SiamFC algorithm, its accuracy is also 5.1% lower than our algorithm. In overall effectiveness, through comparisons with seven contrasting algorithms, our algorithm achieves the optimal results. Overall, while there is room for improvement in the complexity of our algorithm, it performs exceptionally well in key metrics such as precision and success rate. Through comparison with other contrastive algorithms, our algorithm demonstrates superiority in overall performance.
To further validate the capability of our method in handling complex scenarios, we conducted a detailed analysis of the tracking results for different challenging video sequences. Figure 7 shows the tracking precision in various challenging environments, and Table 4 shows the specific values of the precision scores of our tracker and each of the seven other state-of-the-art trackers based on different attributes on the datasets, where the bolded data indicates the optimal data and the underlined data indicates the sub-optimal results. From the results, we can conclude that our method RFCFT has obtained the best results on 10 out of 14 attributes including aspect ratio change (79.4%), background clutter (69.3%), camera motion (76.8%), deformation (81.8%), full occlusion (70.4%), low resolution (82.9%), out-of-view (79.1%), rotation (77.4%), scale variation (74.4%), motion blur (72.2%), and obtained sub-optimal in Illumination Variation. Figure 8 presents the success rate of the tracking process, and we can also see that our method RFCFT has obtained the best results on 3 out of 14 attributes and sub-optimal results on 8 attributes. To conclude, we can demonstrate that the proposed method RFCFT is more effective in distinguishing animals from backgrounds and similar objects and achieves excellent performance improvement in most of the attributes of animal tracking.

5.7. Qualitative Evaluation

Figure 9 shows some tracking results of our tracker and the other seven trackers on different challenging sequences. As we can see, the tracked zebra in the Zebra2 video sequence suffers from partial occlusion, deformation, rotation, scale variation, and aspect ratio change, initiating from the 345th frame; all other trackers experienced instances of drifting or tracking failures, only our tracker and DAHCF demonstrated the capability to consistently and steadfastly track the target zebra, but the fitting accuracy of our tracker is superior. The tracked chameleon in the chameleon-12 video sequence suffers deformation, background clutter, scale variation, and aspect ratio change; starting from frame 1858, most of the rest of the trackers were confused by the mirrored chameleon in the mirror and experienced tracking failures, and only our tracker was able to track the target stably. In the video sequence squirrel16, the tracked squirrel is notably small, and its color closely resembles the background; additionally, the squirrel appears to fast motion during its trajectory; ultimately, only our tracker and tracker DAHCF successfully tracked the target, most other trackers were confused by the squirrel’s tail, leading to tracking failures. The target squirrel in video sequence squirrel12 suffers from deformation, motion blur, camera motion, rotation, scale variation, out-of-view, low resolution, aspect ratio change and so on; among trackers considered for the analysis, our tracker and SiamFC were able to track the target in most of the frames, all other trackers showed tracking drift, especially DSARCF, which failed to track from frame 320.

6. Conclusions

In this paper, we propose an animal tracking algorithm based on correlation filters with response map analysis and adaptive fusion feature update (RFCFT) for tracking specific individual animals’ trajectories. The key contribution is to design a novel adaptive feature fusion mechanism that integrates multi-level deep features and hand-crafted features for effective and efficient animal tracking; by analyzing the response map, we assess the reliability of the tracking results. When the tracking results are unreliable, we introduce temporal information and current frame target information to re-model the target and perform re-tracking to obtain the optimal predicted position; the proposed method automatically enhances the discrimination and suppresses misleading information, enabling online tracking adaptation. In addition, we constructed a specialized animal tracking dataset that enables us to evaluate the effectiveness of animal tracking algorithms objectively. The extensive experimental results on animal datasets demonstrate the effectiveness and robustness of our tracker in comparison with the other seven state-of-the-art trackers.
Furthermore, despite significant findings in achieving a comprehensive balance between tracking accuracy and robustness for individual animals, we candidly acknowledge certain limitations in this study. For instance, the increased complexity resulting from the additional inclusion of animal state change detection and updating modules has elevated the algorithm’s complexity, leading to some loss in tracking speed. Nevertheless, we maintain the belief that these preliminary results provide a valuable foundation for further research.
In future research, we aim to further explore and improve animal trajectory tracking algorithms to address increasingly complex and challenging scenarios. Firstly, we will strive to enhance the accuracy and robustness of the algorithms, particularly in complex environments characterized by large numbers of animals and chaotic motion. Secondly, we plan to investigate methods for better handling occlusions, dynamic backgrounds, and other challenges during the tracking process to improve tracking performance. Additionally, we will explore the utilization of cutting-edge technologies such as deep learning to further enhance algorithm performance, reduce algorithm complexity, and achieve faster end-to-end trajectory tracking. Moreover, we will consider applying the algorithms to a wider range of animal species and real-world application scenarios. Lastly, we aim to strengthen collaboration with fields such as biology and ecology to better understand animal behavior and provide more effective technical support for areas such as animal conservation and ecological environment monitoring.

Author Contributions

Conceptualization, K.Y. and J.F.; methodology, K.Y. and J.F.; software, K.Y. and S.D.; validation, K.Y. and S.D.; writing original draft preparation, K.Y. and J.F.; writing—review and editing, S.D.; visualization, K.Y., J.F. and S.D.; investigation, K.Y., J.F. and S.D.; funding acquisition, J.F., K.Y. All authors have read and agreed to the published version of the manuscript.

Funding

The work was supported by the National Natural Science Foundation of China (Grant number 41971365), Chongqing College of Electronic Engineering Project (XJWT202107).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data will be available from the authors upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Panadeiro, V.; Rodriguez, A.; Henry, J.; Wlodkowic, D.; Andersson, M. A review of 28 free animal-tracking software applications: Current features and limitations. Lab Anim. 2021, 50, 246–254. [Google Scholar] [CrossRef] [PubMed]
  2. Ning, J.; Yang, J.; Jiang, S.; Zhang, L.; Yang, M.H. Object tracking via dual linear structured svm and explicit feature map. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4266–4274. [Google Scholar]
  3. Bolme, D.S.; Beveridge, J.R.; Draper, B.A.; Lui, Y.M. Visual object tracking using adaptive correlation filters. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 2544–2550. [Google Scholar]
  4. Nam, H.; Han, B. Learning multi-domain convolutional neural networks for visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4293–4302. [Google Scholar]
  5. Fan, H.; Ling, H. Sanet: Structure-aware network for visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 42–49. [Google Scholar]
  6. Tao, R.; Gavves, E.; Smeulders, A.W. Siamese instance search for tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1420–1429. [Google Scholar]
  7. Henriques, J.F.; Caseiro, R.; Martins, P.; Batista, J. Exploiting the circulant structure of tracking-by-detection with kernels. In Computer Vision–ECCV 2012, Proceedings of the 12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; Proceedings, Part IV 12; Springer: Berlin/Heidelberg, Germany, 2012; pp. 702–715. [Google Scholar]
  8. Vovk, V. Kernel Ridge Regression, In Empirical Inference: Festschrift in Honor of Vladimir N. Vapnik; Springer: Berlin/Heidelberg, Germany, 2013; pp. 105–116. [Google Scholar]
  9. Henriques, J.F.; Caseiro, R.; Martins, P.; Batista, J. High-speed tracking with kernelized correlation filters. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 37, 583–596. [Google Scholar] [CrossRef] [PubMed]
  10. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; IEEE: Piscataway, NJ, USA, 2005; Volume 1, pp. 886–893. [Google Scholar]
  11. Danelljan, M.; Häger, G.; Khan, F.; Felsberg, M. Accurate scale estimation for robust visual tracking. In Proceedings of the British Machine Vision Conference, Nottingham, UK, 1–5 September 2014; Bmva Press: Nottingham, UK, 2014. [Google Scholar]
  12. Liu, S.; Liu, D.; Srivastava, G.; Połap, D.; Woźniak, M. Overview and methods of correlation filter algorithms in object tracking. Complex Intell. Syst. 2021, 7, 1895–1917. [Google Scholar] [CrossRef]
  13. Danelljan, M.; Hager, G.; Shahbaz Khan, F.; Felsberg, M. Learning spatially regularized correlation filters for visual tracking. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December; pp. 4310–4318.
  14. Kiani Galoogahi, H.; Fagg, A.; Lucey, S. Learning background-aware correlation filters for visual tracking. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1135–1143. [Google Scholar]
  15. Li, Y.; Fu, C.; Ding, F.; Huang, Z.; Lu, G. Autotrack: Towards high-performance visual tracking for uav with automatic spatio-temporal regularization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11923–11932. [Google Scholar]
  16. Zhao, J.; Wei, F.; Chen, N.; Zhou, Z. Spatial and long–short temporal attention correlation filters for visual tracking. IET Image Process. 2022, 16, 3011–3024. [Google Scholar] [CrossRef]
  17. Ma, C.; Huang, J.B.; Yang, X.; Yang, M.H. Hierarchical convolutional features for visual tracking. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 3074–3082. [Google Scholar]
  18. Danelljan, M.; Robinson, A.; Shahbaz Khan, F.; Felsberg, M. Beyond correlation filters: Learning continuous convolution operators for visual tracking. In Computer Vision–ECCV 2016, Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part V 14; Springer: Cham, Switzerland, 2016; pp. 472–488. [Google Scholar]
  19. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  20. Nam, H.; Baek, M.; Han, B. Modeling and propagating cnns in a tree structure for visual tracking. arXiv 2016, arXiv:1608.07242. [Google Scholar]
  21. Bertinetto, L.; Valmadre, J.; Henriques, J.F.; Vedaldi, A.; Torr, P.H.S. Fully-Convolutional Siamese Networks for Object Tracking. In Proceedings of the Computer Vision–ECCV 2016 Workshops, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  22. Li, F.; Tian, C.; Zuo, W.; Zhang, L.; Yang, M.H. Learning spatial-temporal regularized correlation filters for visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4904–4913. [Google Scholar]
  23. Zhang, J.; Liu, Y.; Liu, H.; Wang, J.; Zhang, Y. Distractor-aware visual tracking using hierarchical correlation filters adaptive selection. Appl. Intell. 2022, 52, 6129–6147. [Google Scholar] [CrossRef]
  24. Feng, W.; Han, R.; Guo, Q.; Zhu, J.; Wang, S. Dynamic saliency-aware regularization for correlation filter-based object tracking. IEEE Trans. Image Process. 2019, 28, 3232–3245. [Google Scholar] [CrossRef] [PubMed]
  25. Zhang, J.; He, Y.; Feng, W.; Wang, J.; Xiong, N.N. Learning background-aware and spatial-temporal regularized correlation filters for visual tracking. Appl. Intell. 2023, 53, 7697–7712. [Google Scholar] [CrossRef]
  26. Huang, Z.; Fu, C.; Li, Y.; Lin, F.; Lu, P. Learning aberrance repressed correlation filters for real-time uav tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2891–2900. [Google Scholar]
  27. Wu, Y.; Lim, J.; Yang, M. Online object tracking: A benchmark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 2411–2418. [Google Scholar]
  28. Wu, Y.; Lim, J.; Yang, M. Object tracking benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1834–1848. [Google Scholar] [CrossRef] [PubMed]
  29. Fan, H.; Lin, L.; Yang, F.; Chu, P.; Deng, G.; Yu, S.; Bai, H.; Xu, Y.; Liao, C.; Ling, H. Lasot: A high-quality benchmark for large-scale single object tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5374–5383. [Google Scholar]
Figure 1. Overview of algorithm.
Figure 1. Overview of algorithm.
Animals 14 00902 g001
Figure 2. Some wildlife exhibits a significant resemblance in appearance to its surrounding environment.
Figure 2. Some wildlife exhibits a significant resemblance in appearance to its surrounding environment.
Animals 14 00902 g002
Figure 3. Illustration of model update.
Figure 3. Illustration of model update.
Animals 14 00902 g003
Figure 4. Dataset overviews.
Figure 4. Dataset overviews.
Animals 14 00902 g004
Figure 5. System operation procedure.
Figure 5. System operation procedure.
Animals 14 00902 g005
Figure 6. Comparisons with the state-of-the-art tracking algorithms on animal datasets.
Figure 6. Comparisons with the state-of-the-art tracking algorithms on animal datasets.
Animals 14 00902 g006
Figure 7. Precision plots evaluated trackers on the animal datasets in terms of 14 challenging attributes.
Figure 7. Precision plots evaluated trackers on the animal datasets in terms of 14 challenging attributes.
Animals 14 00902 g007
Figure 8. Success plots evaluated trackers on the animal datasets in terms of 14 challenging attributes.
Figure 8. Success plots evaluated trackers on the animal datasets in terms of 14 challenging attributes.
Animals 14 00902 g008
Figure 9. The qualitative results of the algorithm on certain video sequences.
Figure 9. The qualitative results of the algorithm on certain video sequences.
Animals 14 00902 g009
Table 1. Descriptions of 14 different attributes in animal datasets.
Table 1. Descriptions of 14 different attributes in animal datasets.
AttributeDefinitionAttributeDefinition
CMAbrupt motion of the cameraVCViewpoint affects target appearance significantly
ROTThe target rotates in the imageSVThe ratio of bounding box is outside the rage
DEFThe target is deformable during trackingBCThe background has the similar appearance as the target
FOCThe target is fully occluded in the sequenceMBThe target region is blurred due to target or camera motion
IVThe illumination in the target region changesARCThe ratio of bounding box aspect ratio is outside the rage
OVThe target completely leaves the video frameLRThe target box is smaller than 1000 pixels in at least one frame
POCThe target is partially occluded in the sequenceFMThe motion of the target is larger than the size of its bounding box
Table 2. Results of ablation experiments.
Table 2. Results of ablation experiments.
 PrecisionSuccess
RFCFT0.7650.562
RFCFT-NU0.7510.557
RFCFT-NF0.7330.546
The bolded data represents the optimal result for this metric.
Table 3. Tracking performance on the animal datasets.
Table 3. Tracking performance on the animal datasets.
 RFCFTDAHCFBSTCFSTRCFARCFAutoTrackDSARCFSiamFC
precision0.7650.741 0.6670.5870.5120.4660.4060.714
Success0.5620.5470.5390.4760.4030.3930.3760.555
Fps1.9150.00417.30927.65522.72060.8018.89673.910
The bolded data indicates the optimal data and the underlined data indicates the sub-optimal results.
Table 4. Performance evaluations of different attributes on animal datasets.
Table 4. Performance evaluations of different attributes on animal datasets.
 RFCFTDAHCFBSTCFSTRCFARCFAutoTrackDSARCFSiamFC
ARC0.7940.7170.60.5590.4440.4240.370.698
BC0.6930.6390.6180.5230.3750.4140.4060.563
CM0.7680.7240.6050.5790.4390.3680.4180.635
DEF0.8180.7630.660.5860.5290.4640.3930.710
FM0.6860.6890.7840.7520.4720.4580.6850.732
FOC0.7040.5790.6720.5590.3160.3040.4230.633
LR0.8290.7010.7420.6530.5700.5370.5450.808
OV0.7910.6710.5210.3560.2900.2410.4300.703
POC0.7260.7360.6460.5290.3900.3960.3580.728
ROT0.7740.7650.6870.5290.5290.4790.3490.753
SV0.7440.7140.6530.5710.4900.4510.3970.711
IV0.8530.8310.7950.8090.6930.580.4790.880
MB0.7220.7170.5590.5630.420.3850.4060.617
VC0.4410.4260.5110.4570.4420.3590.2640.552
The bolded data indicates the optimal data and the underlined data indicates the sub-optimal results.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yin, K.; Feng, J.; Dong, S. Learning Rich Feature Representation and State Change Monitoring for Accurate Animal Target Tracking. Animals 2024, 14, 902. https://doi.org/10.3390/ani14060902

AMA Style

Yin K, Feng J, Dong S. Learning Rich Feature Representation and State Change Monitoring for Accurate Animal Target Tracking. Animals. 2024; 14(6):902. https://doi.org/10.3390/ani14060902

Chicago/Turabian Style

Yin, Kuan, Jiangfan Feng, and Shaokang Dong. 2024. "Learning Rich Feature Representation and State Change Monitoring for Accurate Animal Target Tracking" Animals 14, no. 6: 902. https://doi.org/10.3390/ani14060902

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop