Next Article in Journal
Characterization of the Fat Channel for Intra-Body Communication at R-Band Frequencies
Next Article in Special Issue
High-Fidelity Inhomogeneous Ground Clutter Simulation of Airborne Phased Array PD Radar Aided by Digital Elevation Model and Digital Land Classification Data
Previous Article in Journal
Reflectance-Based Vegetation Index Assessment of Four Plant Species Exposed to Lithium Chloride
Previous Article in Special Issue
A High Precision Quality Inspection System for Steel Bars Based on Machine Vision
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unmanned Aerial Vehicle Object Tracking by Correlation Filter with Adaptive Appearance Model

School of Computer Science and Engineering, Northwestern Polytechnical University, Xi’an 710129, Shaanxi, China
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(9), 2751; https://doi.org/10.3390/s18092751
Submission received: 28 June 2018 / Revised: 13 August 2018 / Accepted: 13 August 2018 / Published: 21 August 2018
(This article belongs to the Special Issue Sensors Signal Processing and Visual Computing)

Abstract

:
With the increasing availability of low-cost, commercially available unmanned aerial vehicles (UAVs), visual tracking using UAVs has become more and more important due to its many new applications, including automatic navigation, obstacle avoidance, traffic monitoring, search and rescue, etc. However, real-world aerial tracking poses many challenges due to platform motion and image instability, such as aspect ratio change, viewpoint change, fast motion, scale variation and so on. In this paper, an efficient object tracking method for UAV videos is proposed to tackle these challenges. We construct the fused features to capture the gradient information and color characteristics simultaneously. Furthermore, cellular automata is introduced to update the appearance template of target accurately and sparsely. In particular, a high confidence model updating strategy is developed according to the stability function. Systematic comparative evaluations performed on the popular UAV123 dataset show the efficiency of the proposed approach.

1. Introduction

Recent years have witnessed significant development in the field of computer vision. An enormous amount of research effort has gone into vision-based tasks, such as object tracking [1,2,3], recognition [4,5] and saliency detection [6]. As an important field of computer vision, visual tracking [7,8,9,10,11] plays an active role in a wide range of applications, in which tracking using UAVs is a very promising one. Since the camera can follow the target based on visual feedback and meanwhile change its orientation and position to improve the tracking performance, aerial tracking [12] is widely applied to a diverse set of objects, which cannot be physically or persistently tracked from the ground, such as humans, animals, cars, boats, etc. Apart from those related to surveillance, a large number of new applications based on aerial tracking have been applied including infrastructure inspection [13], person following [14] and aircraft avoidance [15]. However, compared with static tracking systems, aerial tracking requires the ability of analyzing a dynamic scene and handling new challenges posed on the UAV videos.
UAV tracking usually requires long-term tracking, since the camera can follow the target in contrast to the static surveillance scenario. To alleviate the model drifting and handle occlusion problem, the TLD tracker [1] combined the traditional tracking algorithm with the detection algorithm to make the system more reliable through an online learning mechanism. The tracking algorithm was based on the optical flow method, and the detection algorithm generated numerous candidate regions where each region must be accepted by three filters to become the detection result. The detection algorithm re-initialized the tracker when the tracking algorithm failed. However, it is difficult for TLD to meet the requirement of a large number of frames per second without resorting to parallel computation on a standard PC in dealing with real-time processing tasks. From this viewpoint, correlation filters [16,17,18,19,20,21,22] show their strengths both in speed and accuracy, where tracking problem is converted from time domain to frequency domain with fast Fourier transform (FFT). In so doing, convolution can be substituted with multiplication in an effort to achieve fast learning and target detection. Similar to TLD, the redetection procedure was carried out based on an online random fern classifier in the long-term correlation tracking (LCT) method [16], and the training samples were collected by a k-nearest neighbor (KNN) classifier. Zhu et al. applied the CUR theory to collaborative correlation tracking (CCT) method [17]. The CUR matrix approximation not only computed the low rank property of object representation, but also projected the representation matrix of historical objects to achieve a robust object representation.
By recognizing the success of deep convolutional neural networks (CNNs) on a wide range of visual-recognition tasks, tracking methods based on CNNs have also been developed [23,24,25,26]. Empirical studies using large object-tracking benchmark datasets have shown that such CNN-based trackers perform favorably against methods based on the use of handcrafted features. Notwithstanding, the underlying approach is computationally expensive and often cannot meet the speed requirements for real-time applications. In [24], Held et al. proposed Generic Object Tracking Using Regression Networks (GOTURN) [24] for offline training of a simple feed-forward network that can track generic objects at more than 100 fps with GPU. Yet, if only a CPU is available, the tracker runs at 2.7 fps [24]. A different piece of work showing favorable results for real-time visual tracking is the biologically inspired tracker (BIT) [27], which can extract low-level biologically inspired features and imitate an advanced learning mechanism. Note that BIT operates in real-time since fast Gabor approximation and fast Fourier transform are used for online learning and detection. However, all these methods cannot cope well with challenges presented by UAV videos, which typically involve low resolution, scale variation, aspect ratio change and occlusion. To address the issues we propose a robust tracking approach, which offers two unique advantages:
(1)
In order to handle the scale variation and aspect ratio change, a novel propagation method based on cellular automata (CA) is introduced to update the appearance template of target adaptively and sparsely, which benefits from the consistency among similar image pixels.
(2)
A new adaptive template update scheme is proposed to further alleviate the problem of model drift which is caused by occlusion or distracter. The effectiveness of this approach is demonstrated through extensive comparisons against other methods.
The rest of this paper is organized as follows: Section 2 discusses relevant previous work on correlation filter, CA and feature selection in correlation tracking. Under the general framework of correlation filter, Section 3 describes our approach. Section 4 presents an evaluation of the proposed approach and a comparative study with state-of-the-art techniques. Section 5 discusses the tracking speed of different methods and assesses the actual effect of CA in the proposed method. Finally, Section 6 concludes this work and points out interesting further research.

2. Related Work

2.1. Correlation Filter Trackers

Because of their impressive high-speed, correlation filters have attracted a great deal of interests in object tracking. For instance, Bolme et al. [18] have proposed the minimum output sum of squared errors (MOSSE) filter, which works by trying to find the maximum cross correlation response between the model and candidate patch. Henriques et al. [19] exploited the circulate structure and Fourier transformation in a kernel space (CSK), offering excellent performance on a range of computer vision problems. A vector correlation filter (VCF) was proposed by Boddeti et al. [20] to minimize localization errors while improving the tracking speed. Danelljan et al. [28] exploited the color attributes of an object and introduced the color name (CN) features into CSK to perform object tracking. Combining techniques of kernel trick and cycle shift [19], kernelized correlation filter (KCF) [29] entails more adaptive performance for diverse scenarios using histogram of oriented gradient (HOG) features. The DSST tracker [30] learns adaptive multi-scale correlation filters by the use of multi-channel HOG features to handle the scale change of target objects. To learn a model that is inherently robust to both color changes and deformations, Staple [31] combines two image patch representations that are sensitive to challenging factors. SRDCF [32] utilizes a spatial regularization component in the learning process to penalize correlation filter coefficients as a function of their spatial location. Recently, to drastically reduce the number of parameter in the model, Danelljan et al. [33] proposed a factorized convolution operator. Meanwhile a compact generative model of the training sample distribution significantly reduces the memory and time complexity, while providing better diversity of samples.
Whilst many methods exist, as outlined above, these methods do not address the critical issue of aerial tracking. When scale variation and aspect ratio change exist, the traditional correlation trackers only enlarge or narrow the bounding box in equal proportion, which will introduce a great number of background information and finally result in model drifting. In addition, dense updating scheme will also reduce the aerial tracking efficiency when serious occlusion exists.

2.2. CA

CA is a dynamic system with simple structure but is of complex self-organizing behavior, as proposed in [34]. Consisting of a lattice of cells with discrete states, the model evolves in discrete time steps according to the definite rules. The current state of the cell and the states of its nearest neighbors make joint efforts to its next state. CA has been applied to simulate the process of various complicated dynamic systems [35].
Specifically, a CA always operates on a lattice of sites p P Z n . A CA is a triplet A = ( S , N , δ ) , where S is a non-empty state set, N is the neighborhood system, and δ : S N S is the local transition rule. According to the states of the neighborhood cells at previous time step t , this function defines the rule of calculating the state of cell at t + 1 time step.
Von Neumann (Equation (1)) and Moore (Equation (2)) offer two commonly used neighborhood systems. The neighborhood structures associated with these systems are shown in Figure 1:
N ( p ) = { q Z n : p q 1 : = i = 1 n | p i q i | = 1 }
N ( p ) = { q Z n : p q : = max i = 1 , n ¯ | p i q i | = 1 }
Given a certain cell, the neighboring cells above and below, and those on the right and left together with the cell itself is called Von Neumann neighborhood of this cell. The radius of the neighborhood definition is 1, as only the next layer is considered.
In addition to the four cells of Von Neumann neighborhood, Moore neighborhood also includes the four next nearest cells along the diagonal. In this case, the radius is equal to 1 also. The cell state S p in our case is actually a triplet ( l p , F p , C p ) , where the label l p denotes the current cell, F p means the ‘strength’ of current cell and C p is cell feature vector.

2.3. Feature Selection in Correlation Tracking

Features play an important role in computer vision, in which gradient and color features are the most widely exploited. In particular, HOG features are the most commonly employed to catch texture and gradient information in object tracking [30,31,32,33], while color measurements can vary significantly over an image sequence due to variations in illuminant, shadows, shading, camera and object geometry. Henriques et al. [19] utilized color attributes or features to obtain excellent results for visual tracking problems. Recent work [36] has verified that there exists a strong complementarity between gradient and color features. On this basis, Danelljan et al. [28] introduced CN features and HOG features together to construct a correlation filter, in an attempt to capture color characteristics and abundant gradient information, considerably improving the tracking performance.

3. Proposed Methods

We aim to develop a robust tracking algorithm that is adaptive to significant appearance change without being prone to drifting. For this, the fused features to be extracted are represented as a multi-dimensional vector of input features (which themselves are each encoded as a one-dimensional vector of multiple real values). Further, CA is introduced to sparsely update the aspect ratio of the bounding box, which makes our method less susceptible to the noise from background. In particular, the adaptive model updating strategy is also put forward in order to achieve better performance. The proposed tracking framework is illustrated in Figure 2. Our algorithm can be divided into three modules: object location, model updating and sparse template updating. In object location, fused features are extracted first, followed by deriving the response map that is calculated by the proposed correlation filter. Target location is estimated by searching for the location of the maximum value within the response map. Afterward, if the conditions are satisfied, CA will be introduced to obtain a new appearance template for re-initialization (while discarding the previous template and training a correlation filter with the current one). In addition, model updating also plays an important role in our method, with the filter being updated only when the indicator T is bigger than a given threshold.

3.1. Correlation Tracking through Fused Features

The UAV videos are affected by platform motion and jitter, so the color and shape of the target are rapidly changing. Compared with generic object tracking, tracking challenges are amplified in aerial scenarios including abrupt camera motion, distance-induced low resolution, significant changes in scale and aspect ratio, fast moving objects, as well as, partial and full occlusion. Having taken notice of these issues caused by such conditions and related implications on extraction of object details using a single feature, a method fusing CN and HOG features is employed in this work to achieve robust performance in aerial tracking. Furthermore, the aforementioned features are concatenated directly to form a vector as a fused feature descriptor. In this paper, we utilize fused feature vector representation which better fits with the correlation tracking framework. Denote x d as the fused feature vector of a cardinality d R D . We consider y d as the desired correlation output corresponding to a given sample x d . A correlation filter w with the same dimensionality of x d is then learned by solving the following minimization problem:
w * = argmin w x d y d 2 + λ w 2 2
where   λ is a regularization parameter. Note that the minimization problem in Equation (3) is akin to training the vector correlation filters in [20], and can be solved within each individual feature channel using FFT. Let the capital letters be the corresponding Fourier transformed signals. The learned filter in the frequency domain on the d th   ( d { 1 , , D } ) channel can be written as:
W d = Y ¯ X d i = 1 D X ¯ i X i + λ
where Y , X , W denote the discrete Fourier transforms (DFT) of y , x , w , respectively; Y ¯ represents the complex conjugation of Y ; and Y ¯ X d is a point-wise product. Given an image patch in the next frame, the fused feature vector is denoted by Z R D . The correlation response map is computed by:
r = F 1 ( d = 1 D W d Z ¯ d )
where the operator F 1 denotes the inverse FFT. Then the target location can be estimated by searching for the position of the maximum value of the correlation response map r:
( x , y ) = arg max a , b ( R ( a , b ) )

3.2. Adaptive Appearance Template Updating Based on CA

Typically, the appearance of an object can be divided into shape and scale. The conventional algorithms usually only update the scales, which result in some disadvantages. For example, if the aspect ratio of template cannot adjust to the change of target, only simple scale expanding or narrowing will draw into plenty of noises on UAV videos, when aspect ratio of object changes frequently. To deal with this problem, apart from learning a separate 1-dimensional correlation filter to estimate the target scale straightly [30], we propose our adaptive appearance template updating scheme based on CA. Here, the single-layer CA is introduced in our tracker to regularly adjust the aspect ratio of the bounding box. An unlabeled digital image may be then considered as a particular configuration state S I of a cellular automaton, where cellular space I is defined by the array set of image, and initial states S i for i I are set to: l i = 0 , F i = 0 , c i = R G B i , where R G B i is the three dimensional vector of pixel’s color in RGB space.
Then, a few number of “object seeds” are selected around the center of target while pixels on the image boundaries are all served as “background seeds” in the current frame. The states of object seeds and background seeds are set to 1 and −1, respectively. It is intuitive to accept that neighbors with more similar color features have a greater influence on the next state of the cell. The similarity of any pair of pixels is measured by a defined distance in RGB color space. So we construct the impact factor matrix F i = [ f i , j ] N x N by defining the impact factor f i , j of pixel i to j as:
f i , j = 1 | | c i , c j | | max ( c ) i f j N B ( i ) f i , j = 0 i f j N B ( i )
where | | c i , c j | | denotes the Euclidean distance in RGB color space between the pixel i and j , N B ( i ) is the set of neighbors of cell i . The form of f insures its value within the range [ 0 , 1 ] , which is mathematically tractable.
Then, a novel propagation mechanism dependent on CA is proposed to exploit the intrinsic relevance of similar regions through interactions with neighbors:
c i ( t + 1 ) = max ( c j ( t ) · F i j ( t ) , c i ( t ) )
where c i ( t ) , c i ( t + 1 ) denote the value of pixel i in current frame t and the next frame t + 1 , respectively. What is more, if c i ( t + 1 ) is different from c i ( t ) , the value of l i ( t + 1 ) will be set the same as l j ( t ) .
To obtain a more precise template, we first identify the center point of the target in the previous frame. A particular patch within its neighborhood is intercepted from the original image manually, which is of a size five times as large as the previous target. Then we repeat the above steps for every pixel in the sampled patch until the states of pixels have no further variation in the current frame. After that, we bag the pixels which states are equal to 1 into a connected domain and take its minimum bounding box. If the area of minimum bounding-box are within the interval [a, b], the connected domain will be considered as the new template to reinitialize our tracker.
Therefore, as indicated in Figure 3, the aspect ratio of bounding box can change with the appearance variation adaptively without introducing too much noise, which greatly reduces the risk of model drift. Note that updating with moderate frequency generally leads to an improved tracking result. The appearance of the target usually only changes slightly and the traditional updating strategy in correlation filter is able to handle it well. Under the circumstances, re-initialization densely requires a significant amount of computing resources and may lead to temporal information loss. It is difficult to make a significant improvement over the tracking accuracy merely by the use of excessive dense re-initialization. Of course, when the appearance of the target changes radically, the resulting model draft can cause tracking failure. In this situation, an overly sparse re-initialization cannot introduce new templates in time. From the empirical observations, we choose to reinitialize our correlation filter every 60 frames, trading off between computational efficiency and tracking effectiveness.

3.3. Model Updating via High Confidence

No matter the tracking result is accurate or not, the traditional correlation trackers update their models at each frame. In fact, unsupervised updating will lead to model drifting and finally cause a deterministic failure when the occlusion exists severely. To obtain a robust and efficient approximation, we adopt a stability function T [37] to measure the stability of response map R . First, T is defined as:
T = | R max R min | 2 m e a n ( a , b ( R a . b R min ) 2 )
where R max , R min and R a , b denote the maximum, minimum and the a th row b th column elements of R, respectively.
Figure 4 shows two original images and their response map in different situation. From Figure 4, it is obvious that the more stable the response map is, the better the location accuracy is. Only when these two criteria R max and T of the current frame are greater than their respective historical average values with certain ratios β1, β2, the tracking result in the current frame is considered to be of high-confidence. Under the circumstances, we will update the numerator A d and the denominator B d of the correlation filter W d in Equation (4) separately, using a moving average:
A t d = ( 1 η ) A t 1 d + η Y X ¯ t d
B t d = ( 1 η ) B t 1 d + η i = 1 D X t i X ¯ t i
W t d = A t d B t d + λ
where t is the frame index and η denotes the learning rate.
When the target is in severely occlusion or totally missing in the current frame, the peak value R max and stability function T may be relative small, our tracker will not update the model in this frame. In this instance, most of interference noise from background are prevented from our model and therefore achieves a robust aerial tracking.

4. Experiment and Results

In order to present an objective evaluation about the performance of the proposed approach, the UAV123 dataset [38] is selected to show full results of all chosen trackers. UAV123 provides an evaluation of trackers on more than 100 new fully annotated HD videos captured from a professional grade UAV. This benchmark both complements current benchmarks establishing the aerial component of tracking and provides a more comprehensive sampling of tracking nuisances that are ubiquitous in low-altitude UAV videos. Apart from aspect ratio change (ARC) and fast motion (FM), these video sequences are also affected by several adverse conditions such as background clutter (BC), camera motion (CM), full occlusion (FOC), illumination variation (IV), low resolution (LR), out of view (OV), partial occlusion (POC), similar object (SOB), scale variation (SV), viewpoint change (VC). Thus, the experiments carried out covered all typically challenges typically involved in real-world aerial tracking problems.
The trackers are running on these challenging sequences to test their general ability and also their special scenarios handling. We compare our proposed tracker with nine state-of-art trackers, including ORVT [12], GOTURN [24], BIT [27], DSST [30], fDSST [39], KCFDP [40], SAMF [41], OCT_KCF [42] and CNT [43]. Among these trackers, ORVT is an onboard robust visual algorithm for aerial tracking using a reliable global-local object model.
The proposed tracker is implemented in Matlab2014a on a PC equipped with an Intel i5-7500 processor (four cores, 3.4 GHz clock speed, without hyper-threading technology). In addition, a 16 GB RAM (RAM clock: 2400 MHZ) is utilized without using any sophisticated program optimization. The interval [a, b] is set twice as much as the width and height of the target.

4.1. Quantitative Evaluation

We follow the standard evaluation metrics for the tracking algorithms in two aspects: the precision rate and success rate [44]. The precision rate shows the percentage of successfully tracked frames on which the center location error (CLE) of a tracker is within a given threshold (e.g., 20 pixels), and CLE is defined as the average Euclidean distance between the center locations of the targets and the manually labeled ground truths. A tracking result in a frame is considered successful if | r d r t | | r d r t | > θ for a threshold θ ( 0 , 1 ] , where r d and r t denote the areas of the bounding boxes of the tracking and the ground truth, respectively, and represent the intersection and union of two regions, respectively, and | | denotes the number of pixels in the region. Thus, the success rate is defined as the percentage of frames where the overlap rates are greater than a threshold θ . Normally, the threshold θ is set to 0.5.
We present the results under one-pass evaluation (OPE) using the average precision and success rate over all sequences. OPE is the most common evaluation method which runs trackers on each sequence for once. It initializes the trackers with the ground truth object state in the first frame and report the average precision or success rate of all the results. Figure 5 shows overall quantitative evaluation on precision and success plots with OPE. Note that our approach provides a gain of 4.0% in success rate as compared to the aerial tracking method ORVT, by achieving a score of 41.8%.

4.2. Attribute-Based Comparison

We also perform an attribute-based comparison with other methods on the UAV123 dataset. Figure 6 and Figure 7 show the success plots and precision plots of twelve respective attributes on the precision and success rates, respectively. The overall results with different attributes are summarized in Table 1 and Table 2, which show the averaged rates of the success plots and those of the precision plots, respectively. As can be seen from these results, our tracker always performs reliably and can achieve optimal or at least close to optimal solution in most cases. Specifically, for the amplified challenging factors in aerial tracking, including CM, LR, SV, ARC, FM, FOC and POC, our tracker achieves satisfactory results, benefitting from the robustness of fused features as well as the efficiency of appearance template and model updating strategy. For videos with fast moving objects, camera motion and low resolution, the fused features have more strong abilities to capture the information from object and therefor gain better results compared with the classic single-feature trackers. In addition, when aspect ratio of object changes significantly, our adaptive appearance template updating strategy can adjust the template to the appearance of object. Moreover, the high confidence model updating method prevents the noise from background as much as possible when serious occlusion exists in aerial video. However, our tracker may not perform well when dealing with background clutter and illumination variation. It is probably because these challenges have created serious problems for CA, which result in a partial or inexact template and finally lead to mistaken tracking.

4.3. Qualitative Evaluation

For qualitative evaluation, we select four representative sequences from UAV123 dataset, on which we compared our tracker against five state-of-the-art methods to validate the ability of the proposed approach. Sequences are shown in Figure 8 (from top to down are sequences car2, car9, car14, person16, respectively). Because the targets in the dataset are rather small, we present the tracking results and their partial enlargements to show the comparison of the tracking results of each algorithm more clearly.
As shown in Figure 8a, the aspect ratio of the target has changed significantly, due to car movement. Only our tracker can adaptively adjust to this variation, while other algorithms still try to track the target with the original aspect ratio. As our tracker introduces less noise from the background, model drifting risk is reduced. Regarding the sequence car9, similar objects and scale validation exist over a long period. Apart from that, the target has been severely occluded by a road sign. In such a situation, only our method can track the target stably and outperform on coverage, benefitting from the high confidence model updating strategy. The tracking results on the sequence person16 are shown in Figure 8c. In this sequence, the target suffers from partial or full occlusion and has quite similar color with the background. Under such interference no other method can track the target expect for ours. Owing to the employed stability function, our method is able to prevent low confidence model updating and track the target even after such complex background disturbances. Figure 8d illustrates that only our tracker and the KCFDP tracker can adapt well to the aspect ratio and target scale changes on the sequence with low resolution. Our tracker achieves better scale and position accuracy in comparison with the KCFDP algorithm. However, after violently shaking of the camera, the target is out of view for a long time. At this moment, all the trackers in the experiment drift away.
Qualitative analysis shows that our method can effectively address the problems in aerial tracking, especially the SV, OCC and ARP. These robust results are attributed to the model updating via high confidence as well as the adaptive appearance template updating scheme. Moreover, the fused features also make a great contribution to improve the tracking results on the UAV videos with low resolution.

5. Discussion

5.1. Speed Performance

For practical applications of aerial tracking, the computational efficiency of trackers also needs to be taken into account. Table 3 lists the running speed of each tracker on nine sequences of the UAV123 dataset, and the average speeds over all of the sequences are shown in the last row.
As we can see, the fDSST tracker achieves the highest running speed which is almost 99 fps and the biologically inspired BIT tracker performs well in terms of running efficiency, too. However, CNN-based CNT and GOTURN trackers show low running efficiencies on all of the nine test sequences, which may not meet the standard of real-time running. It is also worthwhile to note that our tracker can meet the real-time requirements, while gaining the outstanding results on both success rate and precision rate. This owes much to the robustness of fused feature and the efficiency of adaptive appearance template updating strategy. Under this basis, we are trying to find an optimization method to speed up our tracker. Meanwhile, our code will be run on a more appropriate running platform to test its portability, preparing for the real-world application.

5.2. Effect of Adaptive Appearance Template Updating

As mentioned in Section 3.2, when the aspect ratio of object is changing, classic correlation trackers are only able to simply expand or narrow the scale, which will draw into plenty of noises from background in aerial tracking. To deal with this problem, we employ an adaptive appearance template updating strategy in our trackers. We analyze the impact of introducing CA to dig out the interrelationship between pixels of object and adjust the template to the appearance change of target on the UAV videos. Figure 9 shows the tracking performance in success plots and precision plots. From this figure we can see that the performance of our tracker improves considerably as compared to the corresponding version of it without employing the adaptive appearance updating strategy. Our results suggest that noises can be significantly reduced with adaptive appearance template updating strategy, while preserving the satisfying tracking performance.

6. Conclusions

In this paper, we propose a novel method to achieve robust aerial tracking. Our approach is based on learning separate discriminative correlation filters for translation and scale estimation. Furthermore, the fused features consisting of CN and HOG features are utilized to improve the ability of our tracker to capture a wealth of information. In addition, we design a series of strategies to adaptively update the appearance template of our tracker based on CA. This allows the template to adapt the aspect ratio change of target and bring in less disturb from background. Notably, a stability function is introduced to update the model in a more reliable way. Finally, we conduct extensive experiments on the UAV123 dataset. The results clearly demonstrate that our approach achieves the state-of-the-art tracking accuracy. Future work includes investigating more powerful fused features to combine intensity and color information. Another research direction is to exploit efficient deep neural network models (e.g., the light-weight neural network), in an effort to achieve more robust aerial tracking for real time applications.

Author Contributions

All the authors made significant contributions to this work. X.X. and Y.L. devised the approach and prepared experimental data; all three analyzed experimental results; Q.S. provided advice for the preparation and revision of the work; X.X. performed the experiments.

Funding

This research was funded by National Key Research and Development Program grant number [2016YFB0502502], Foundation Project for Advanced Research Field grant number [614023804016HK03002].

Acknowledgments

The authors would like to thank the editors and the anonymous referees for their constructive comments which have been very helpful in revising this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kalal, Z.; Matas, J.; Mikolajczyk, K. Tracking-Learning-Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 1409–1422. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Hare, S.; Saffari, A.; Torr, P.H.S. Struck: Structured Output Tracking with Kernels. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 263–270. [Google Scholar]
  3. Lu, H.; Jia, X.; Yang, M.H. Visual tracking via adaptive structural local sparse appearance model. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012; pp. 1822–1829. [Google Scholar]
  4. Hosang, J.; Dollar, P.; Dollar, P. What Makes for Effective Detection Proposals? IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 814–830. [Google Scholar] [CrossRef] [PubMed]
  5. Huang, G.; Liu, Z.; Weinberger, K.Q.; van der Marten, L. Densely connected convolutional networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar]
  6. Borji, A.; Cheng, M.; Hou, Q. Salient Object Detection: A Survey. arXiv, 2014; arXiv:1411.5878. [Google Scholar]
  7. Islam, M.M.; Hu, G.; Liu, Q. Online Model Updating and Dynamic Learning Rate-Based Robust Object Tracking. Sensors 2018, 18, 2046. [Google Scholar] [CrossRef] [PubMed]
  8. Li, F.; Zhang, S.; Qiao, X. Scene-Aware Adaptive Updating for Visual Tracking via Correlation Filters. Sensors 2017, 17, 2626. [Google Scholar] [CrossRef] [PubMed]
  9. Blake, A.; Isard, M. Active Contours: The Application of Techniques from Graphics, Vision, Control Theory and Statistics to Visual Tracking of Shapes in Motion; Springer Science Business Media: Berlin, Germany, 2012. [Google Scholar]
  10. Battiato, S.; Farinella, G.M.; Furnari, A.; Puglisi, G.; Snijders, A.; Spiekstra, J. An integrated system for vehicle tracking and classification. Expert Syst. Appl. 2015, 42, 7263–7275. [Google Scholar] [CrossRef]
  11. Andriluka, M.; Roth, S.; Schiele, B. People-tracking-by-detection and people-detection-by-tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  12. Fu, C.; Duan, R.; Kircali, D. Onboard Robust Visual Tracking for UAVs Using a Reliable Global-Local Object Model. Sensors 2016, 16, 1406. [Google Scholar] [CrossRef] [PubMed]
  13. Fu, C.; Suarez-Fernandez, R.; Olivares-Mendez, M.; Campoy, P. Real-time adaptive multi-classifier multi-resolution visual tracking framework for unmanned aerial vehicles. In Proceedings of the 2nd Workshop on Research, Development and Education on Unmanned Aerial Systems (RED-UAS), Compiegne, France, 20–22 November 2013; pp. 99–106. [Google Scholar]
  14. Lim, H.; Sinha, S.N. Monocular localization of a moving person onboard a Quadrotor MAV. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA, 26–30 May 2015; pp. 2182–2189. [Google Scholar]
  15. Fu, C.; Carrio, A.; Olivares-Mendez, M.; Suarez-Fernandez, R.; Campoy, P. Robust real-time vision-based aircraft tracking from Unmanned Aerial Vehicles. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014; pp. 5441–5446. [Google Scholar]
  16. Ma, C.; Yang, X.; Zhang, C.; Yang, M.H. Long-term correlation tracking. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5388–5396. [Google Scholar]
  17. Zhu, G.; Wang, J.; Wu, Y.; Lu, H. Collaborative Correlation Tracking. In Proceedings of the British Machine Vision Conference, Swansea, UK, 7–10 September 2015; pp. 184.1–184.12. [Google Scholar]
  18. Bolme, D.S.; Beveridge, J.R.; Draper, B.A.; Lui, Y.M. Visual object tracking using adaptive correlation filters. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010; pp. 2244–2250. [Google Scholar]
  19. Henriques, F.; Caseiro, R.; Martins, P.; Batista, J. Exploiting the Circulant Structure of Tracking-by-Detection with Kernels. In Proceedings of the European Conference on Computer Vision (ECCV), Florence, Italy, 7–13 October 2012; pp. 702–715. [Google Scholar]
  20. Boddeti, V.N.; Kanade, T.; Kumar, B.V. Correlation filters for object alignment. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013; pp. 2291–2298. [Google Scholar]
  21. Zhang, B.; Luan, S.; Chen, C.; Han, J.; Wang, W.; Perina, A.; Ling, S. Latent constrained correlation filter. IEEE Trans. Image Process. 2018, 27, 1038–1048. [Google Scholar] [CrossRef] [PubMed]
  22. Yang, L.; Chen, C.; Wang, H.; Zhang, B.; Han, J. Adaptive Multi-class Correlation Filters. In Advances in Multimedia Information Processing—PCM; Springer International Publishing: Cham, Switzerland, 2016. [Google Scholar]
  23. Nam, H.; Han, B. Learning multi-domain convolutional neural networks for visual tracking. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4293–4302. [Google Scholar]
  24. Held, D.; Thrun, S.; Savarese, S. Learning to track at 100 fps with deep regression networks. In Proceedings of the 2016 European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 749–765. [Google Scholar]
  25. Zhu, Z.; Wu, W.; Zou, W.; Yan, J. End-to-end flow correlation tracking with spatial-temporal attention. illumination. arXiv, 2017; arXiv:1711.01124v1. [Google Scholar]
  26. Bertinetto, L.; Valmadre, J.; Henriques, J.F. Fully-convolutional siamese networks for object tracking. In Proceedings of the 2016 European Conference on Computer Vision, Amsterdam, The Netherlands, 8–10 October 2016; pp. 850–865. [Google Scholar]
  27. Cai, B.; Xu, X.; Xing, X.; Jia, K.; Miao, J.; Tao, D. BIT: Biologically Inspired Tracker. IEEE Trans. Image Process. 2016, 25, 1327–1339. [Google Scholar] [CrossRef] [PubMed]
  28. Danelljan, M.; Khan, F.S.; Felsberg, M.; van de Weijer, J. Adaptive Color Attributes for Real-Time Visual Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014; pp. 1090–1097. [Google Scholar]
  29. Henriques, F.; Caseiro, R.; Martins, P.; Batista, J. High-Speed Tracking with Kernelized Correlation Filters. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 583–596. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Danelljan, M.; Hager, G.; Khan, F.S.; Felsberg, M. Accurate scale estimation for robust visual tracking. In Proceedings of the British Machine Vision Conference (BMVC), Nottingham, UK, 1–5 September 2014; pp. 65.1–65.11. [Google Scholar]
  31. Bertinetto, L.; Valmadre, J.; Golodetz, S.; Miksik, O.; Torr, P.H.S. Staple: Complementary Learners for Real-Time Tracking. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1401–1409. [Google Scholar]
  32. Danelljan, M.; Hager, G.; Khan, F.S.; Felsberg, M. Learning Spatially Regularized Correlation Filters for Visual Tracking. In Proceedings of the 2016 IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2016; pp. 4310–4318. [Google Scholar]
  33. Danelljan, M.; Bhat, G.; Khan, F.S.; Felsberg, M. ECO: Efficient Convolution Operators for Tracking. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6931–6939. [Google Scholar]
  34. Von Neumann, J. The general and logical theory of automata. In Cerebral Mechanisms in Behavior; Wiley: Oxford, UK, 1951; pp. 1–41. [Google Scholar]
  35. Qin, Y.; Lu, H.; Xu, Y.; Wang, H. Saliency detection via Cellular Automata. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 110–119. [Google Scholar]
  36. Khan, R.; Weijer, J.V.D.; Khan, F.S.; Muselet, D.; Ducottet, C.; Barat, C. Discriminative Color Descriptors. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013; pp. 2866–2873. [Google Scholar]
  37. Wang, M.; Liu, Y.; Huang, Z. Large Margin Object Tracking with Circulant Feature Maps. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4800–4808. [Google Scholar]
  38. Mueller, M.; Smith, N.; Ghanem, B. A Benchmark and Simulator for UAV Tracking. In Proceedings of the 2016 European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; pp. 445–461. [Google Scholar]
  39. Danelljan, M.; Häger, G.; Khan, F.S.; Felsberg, M. Discriminative Scale Space Tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1561–1575. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Huang, D.; Luo, L.; Wen, M.; Chen, Z. Enable Scale and Aspect Ratio Adaptability in Visual Tracking with Detection Proposals. In Proceedings of the 2015 British Machine Vision Conference, Swansea, UK, 7–10 September 2015; pp. 185.1–185.12. [Google Scholar]
  41. Li, Y.; Zhu, J. A Scale Adaptive Kernel Correlation Filter Tracker with Feature Integration. In Proceedings of the 2014 European Conference on Computer Vision, Zurich, Switzerland, 6–7 September 2014; pp. 254–265. [Google Scholar]
  42. Baochang, Z.; Zhigang, L.; Xianbin, C.; Qixiang, Y.; Chen, C.; Linlin, S.; Alessandro, P.; Rongrong, J. Output constraint transfer for kernelized correlation filter in tracking. IEEE Trans. Syst. Man Cybern. Syst. 2017, 47, 693–703. [Google Scholar]
  43. Zhang, K.; Liu, Q.; Wu, Y. Robust visual tracking via convolutional networks without training. IEEE Trans. Image Process. 2016, 25, 1779–1792. [Google Scholar] [CrossRef] [PubMed]
  44. Wu, Y.; Lim, J.; Yang, M.H. Online object tracking: A benchmark. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 2411–2418. [Google Scholar]
Figure 1. Neighborhood structures.
Figure 1. Neighborhood structures.
Sensors 18 02751 g001
Figure 2. Flowchart of the proposed tracking algorithm.
Figure 2. Flowchart of the proposed tracking algorithm.
Sensors 18 02751 g002
Figure 3. Tracking results of the adaptive appearance template updating scheme. (a) Tracking result with the original template; (b) Mask of new template obtained by CA; (c) Tracking result with the new template.
Figure 3. Tracking results of the adaptive appearance template updating scheme. (a) Tracking result with the original template; (b) Mask of new template obtained by CA; (c) Tracking result with the new template.
Sensors 18 02751 g003
Figure 4. Images and their responses of in different situation. (a) Original image without occlusion; (b) Original image with serious occlusion; (c) Response map of (a); (d) Response map of (b).
Figure 4. Images and their responses of in different situation. (a) Original image without occlusion; (b) Original image with serious occlusion; (c) Response map of (a); (d) Response map of (b).
Sensors 18 02751 g004
Figure 5. Success (left) and precision (right) plots of proposed tracker compared with state-of-the-art approaches on UAV123 dataset.
Figure 5. Success (left) and precision (right) plots of proposed tracker compared with state-of-the-art approaches on UAV123 dataset.
Sensors 18 02751 g005
Figure 6. Success plots of our tracker compared with state-of-the-art approaches on UAV123 dataset.
Figure 6. Success plots of our tracker compared with state-of-the-art approaches on UAV123 dataset.
Sensors 18 02751 g006aSensors 18 02751 g006b
Figure 7. Precision plots of our tracker compared with state-of-the-art approaches on UAV123 dataset.
Figure 7. Precision plots of our tracker compared with state-of-the-art approaches on UAV123 dataset.
Sensors 18 02751 g007
Figure 8. Tracking results of different methods on four representative sequences.
Figure 8. Tracking results of different methods on four representative sequences.
Sensors 18 02751 g008
Figure 9. The success and precision plots of our tracker compared with which without adaptive appearance template updating strategy on the UAV123 dataset.
Figure 9. The success and precision plots of our tracker compared with which without adaptive appearance template updating strategy on the UAV123 dataset.
Sensors 18 02751 g009
Table 1. Overall precision rates on different attributes, where the entries in red denote the best results and the ones in green indicate the second best.
Table 1. Overall precision rates on different attributes, where the entries in red denote the best results and the ones in green indicate the second best.
OursBITfDSSTKCFDPSAMFDSSTORVTCNTGOTURNOCT_KCF
SV0.5230.4740.5040.4190.5040.5050.4800.4290.4150.503
ARC0.4880.4490.4710.3880.4620.4560.4370.3650.3710.461
LR0.4730.3830.4400.3720.4130.4520.4120.4050.3550.457
FM0.3670.3250.3520.3250.3770.3570.3250.2620.2790.350
FOC0.4040.3930.3690.3760.3940.3880.3860.3470.3200.374
POC0.5110.4520.4580.4220.4720.4770.4520.3860.3700.450
OV0.4730.4180.4320.3790.4380.4430.3970.3530.3800.401
BC0.4040.4490.4210.3850.4350.4640.4550.4110.4140.409
IV0.4170.4520.4660.3100.4330.4820.4400.3630.3750.415
VC0.4610.4580.4870.4030.4530.4690.4420.4070.4090.474
CM0.5150.4910.4860.4480.5180.4930.4830.3890.4200.509
SOB0.6090.5680.5600.5370.5790.5650.5590.4990.4690.562
Overall0.5570.5190.5390.4550.5470.5470.5190.4750.4600.538
Table 2. Overall success rates on different attributes, where the entries in red denote the best results and the ones in green indicate the second best.
Table 2. Overall success rates on different attributes, where the entries in red denote the best results and the ones in green indicate the second best.
OursBITfDSSTKCFDPSAMFDSSTORVTCNTGOTURNOCT_KCF
SV0.3860.3060.3730.2950.3590.3120.3440.3150.2690.322
ARC0.3610.2930.3390.2620.3250.2860.3160.2620.2430.300
LR0.2680.1840.2580.1970.2240.2280.2230.2280.1730.223
FM0.2650.2080.2440.2150.2500.1860.2350.1740.1680.223
FOC0.2210.2000.1940.2020.2110.2000.2110.1710.1580.195
POC0.3650.2980.3290.2840.3260.3060.3270.2660.2430.296
OV0.3690.2810.3290.2680.3120.2890.3110.2740.2660.281
BC0.2650.2910.2770.2490.2820.3140.3090.2540.2710.255
IV0.3110.3010.3470.2090.3090.3070.3310.2450.2510.284
VC0.3530.3120.3630.2880.3260.3040.3320.2970.2750.312
CM0.3950.3360.3690.3270.3790.3320.3630.3000.2890.350
SOB0.4450.3640.4070.3780.4080.3620.4000.3380.2940.358
Overall0.4180.3490.4050.3280.3960.3560.3790.3510.3110.358
Table 3. Running speed (frame per second) of each tracker on sequences from UAV123 dataset, where the entries in red denote the best results and the ones in green indicate the second best.
Table 3. Running speed (frame per second) of each tracker on sequences from UAV123 dataset, where the entries in red denote the best results and the ones in green indicate the second best.
Target SizeOursBITfDSSTKCFDPSAMFDSSTORVTCNTGOTURNOCT_KCF
boat627 × 1633121134306124320.916.10116
car169 × 8924946243792230.758.7416
car239 × 2131105165176105261.000.36172
car999 × 1697123118109430.730.4264
car1443 × 6814548936538170.760.6521
person250 × 11110427215519290.821.3924
person633 × 9511497722633271.380.7120
person1633 × 71155810129545140.751240
person2217 × 47249715846692280.739.46196

Share and Cite

MDPI and ACS Style

Xue, X.; Li, Y.; Shen, Q. Unmanned Aerial Vehicle Object Tracking by Correlation Filter with Adaptive Appearance Model. Sensors 2018, 18, 2751. https://doi.org/10.3390/s18092751

AMA Style

Xue X, Li Y, Shen Q. Unmanned Aerial Vehicle Object Tracking by Correlation Filter with Adaptive Appearance Model. Sensors. 2018; 18(9):2751. https://doi.org/10.3390/s18092751

Chicago/Turabian Style

Xue, Xizhe, Ying Li, and Qiang Shen. 2018. "Unmanned Aerial Vehicle Object Tracking by Correlation Filter with Adaptive Appearance Model" Sensors 18, no. 9: 2751. https://doi.org/10.3390/s18092751

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop