Next Article in Journal
The Influence of Land Use on the Grassland Fire Occurrence in the Northeastern Inner Mongolia Autonomous Region, China
Next Article in Special Issue
Conditional Random Field (CRF)-Boosting: Constructing a Robust Online Hybrid Boosting Multiple Object Tracker Facilitated by CRF Learning
Previous Article in Journal
A Topology Control Strategy with Reliability Assurance for Satellite Cluster Networks in Earth Observation
Previous Article in Special Issue
A Real-Time High Performance Computation Architecture for Multiple Moving Target Tracking Based on Wide-Area Motion Imagery via Cloud and Graphic Processing Units
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Effective Visual Tracking Using Multi-Block and Scale Space Based on Kernelized Correlation Filters

Graduate School of Advanced Imaging Science, Multimedia, and Film, Chung-Ang University, Seoul 06974, Korea
*
Author to whom correspondence should be addressed.
Sensors 2017, 17(3), 433; https://doi.org/10.3390/s17030433
Submission received: 9 December 2016 / Revised: 13 February 2017 / Accepted: 17 February 2017 / Published: 23 February 2017
(This article belongs to the Special Issue Video Analysis and Tracking Using State-of-the-Art Sensors)

Abstract

:
Accurate scale estimation and occlusion handling is a challenging problem in visual tracking. Recently, correlation filter-based trackers have shown impressive results in terms of accuracy, robustness, and speed. However, the model is not robust to scale variation and occlusion. In this paper, we address the problems associated with scale variation and occlusion by employing a scale space filter and multi-block scheme based on a kernelized correlation filter (KCF) tracker. Furthermore, we develop a more robust algorithm using an appearance update model that approximates the change of state of occlusion and deformation. In particular, an adaptive update scheme is presented to make each process robust. The experimental results demonstrate that the proposed method outperformed 29 state-of-the-art trackers on 100 challenging sequences. Specifically, the results obtained with the proposed scheme were improved by 8% and 18% compared to those of the KCF tracker for 49 occlusion and 64 scale variation sequences, respectively. Therefore, the proposed tracker can be a robust and useful tool for object tracking when occlusion and scale variation are involved.

1. Introduction

Visual tracking is a core field of computer vision with many applications such as human computer interaction, surveillance, robotics, driverless vehicles, motion analysis and various intelligent systems. Over the past few decades, visual tracking algorithms with improved performance have been proposed, but they have not provided the desired results in situations involving illumination variation, scale variation, background clutter, and occlusion.
The current tracking algorithms mostly use either the generative method [1,2,3,4,5,6,7,8] or the discriminative method [9,10,11,12,13,14]. The correlation filter-based tracker which is discriminative method has been proven to have high efficiency. Tracking a target object more accurately necessitates estimation of the extent to which the object changes scale. The correlation filter-based tracker [15,16,17,18,19,20,21,22,23] uses a fixed template size, and it cannot take into account the change in scale. Usually, an exhaustive search method that uses a pyramid structure is used for scale estimation; however, it involves complex computation. In order to isolate the problem, this paper uses the scale space filter [15] for efficiently estimating the object scale. A part-based method [24,25,26,27,28,29] has been actively researched to solve problems related to changes in the appearance of the target object such as partial occlusion and deformation. This method segments a target object into multiple parts by using a pre-designated approach and is thus robust in nature. When partial occlusion occurs, apart from the occluded area, there is an area in which the targeted object continues to remain visible. Estimation of the position of the target object in the next frame according to its position in the previous frame makes it possible to acquire trustworthy results. The kernelized correlation filters (KCF) tracker [16] uses the correlation filter. Recently, Zhang et al. proposed a circulant sparse tracker (CST) [8] that combined circulant matrix and sparse representation. Danelljin, who had proposed DSST [15], developed a spatially regularized discriminative correlation filter (SRDCF) tracker [21] which reported the outstanding performance at the cost of heavy computations. Ruan et al. presented the fusion features [22] considering color information and discriminative descriptors with 44 dimensional HOG features. The sum of template and pixel-wise learners (STAPLE) [23] is a novel tracker employing a new color histogram model, and it showed the good performance among the recently proposed color feature-based approaches. However, this model is not physically robust to occlusion. In particular, deep-learning increasingly becomes important in computer vision, and thus convolutional neural network-based tracker has been highlighted. Zhang et al. proposed a robust visual tracker without training [30] using convolutional network. The aforementioned and recent visual tracking mainly focused on the performance in terms of accuracy at the cost of computational time.
A novel scheme is required to realize efficient and effective performance for visual tracking. The KCF tracker is exceptionally fast, even among other correlation filter-based trackers. Therefore, we apply the multi-block model, which we believe to be more effective based on the KCF tracker for occlusion and scale variation as shown in Figure 1.
The remainder of this paper is organized as follows: Section 2 discusses previous studies related to correlation filter-based trackers and part-based models. Section 3 explains the KCF tracker and presents the proposed algorithm. Section 4 evaluates the performance of the proposed method in challenging sequences and compares it with state-of-the-art methods. Finally, Section 5 concludes the work with some discussion.

2. Related Work

The field of visual tracking has long been a focus area for research; therefore, various approaches and categorizing methods have been proposed. Current trackers can be categorized as generative model trackers or discriminative model trackers. Generative trackers [1,2,3,4,5,6,7,8] typically adopt a model that describes the appearance of the target object. Therefore, when there is a change in appearance in an image sequence, the generative trackers reliably represent the change and find the most similar candidate. There are many different models that are currently used such as histogram and sparse representation [1,2,3,4,5,6,7,8]. Incremental visual tracking (IVT) [1], which is based on a low-dimensional principle component analysis (PCA) subspace, uses an adaptive appearance update model. IVT is robust to illuminant changes and simple pose changes; however, it is very sensitive to partial occlusion and background clutter. In similar environments such as those with occlusion, there are many outliers that affect the performance of IVT. This problem was solved by using the probability continuous outlier model (PCOM) [2] to remove outliers of partial occlusion using graph cut based on IVT. Some of the other generative models include visual tracking by decomposition (VTD) [3], which extends particle filter tracking, the L1 minimization tracker [4] with a sparse representation, fragment-based tracker (Frag) [5] designed to be robust to occlusion using a local patch, multi-task tracker (MTT) [6], low-rank sparse tracker [7], and circulant sparse tracker (CST) [8] which combine circulant matrix and sparse representation. In contrast, discriminative model trackers are mainly concerned with object classification problems. The purpose of these trackers is to obtain the position of the current target object from the previous position and to separate the discriminative background and object [9,10,11,12,13,14]. Some of the discriminative model trackers are ensemble tracking [9], which has an ensemble structure consisting of a combination of several weak classifiers; Online AdaBoosting (OAB) [10], which applies discriminative feature selection and online boosting; online random forests (ORF) [11], which learn random forests online; structured output tracking with kernels STRUCK [12], which uses a support vector machine (SVM), multiple instance learning (MIL) [13], and tracking-learning-detection [14] which executes online learning with detectors and trackers at the same time. Some of the recent trackers include transfer learning with Gaussian processes regression (TGPR) [31] and multi-expert entropy minimization (MEEM) [32]. TGPR statistically analyzes the Gaussian processes regression on the basis of semi-supervised learning. MEEM uses an ensemble learning structure and appearance change based on minimum entropy. All correlation filter-based trackers belong to the discriminative model tracker category. Thus, the proposed approach is the discriminative method because it is based on the type of correlation filter.

2.1. Correlation Filter-Based Tracking

The correlation filter-based tracker is currently the most actively researched tracking algorithm [15,16,17,18,19,20,21,22,23]. According to the convolution theory, correlation is computationally highly efficient because it can be calculated as a simple product of two signals in the frequency domain. Consequently, trackers based on correlation filters have low computation. The minimum output sum of squared error (MOSSE) [17] by Bolme et al. successfully used correlation filters on tracking and showed impressive performance and speed. Henriques et al. presented a more effective method using the correlation filter proposed by the circulant structure with kernels (CSK) tracker [18]. The MOSSE tracker uses the intensity feature of the image and processes several hundred frames per second (FPS) because of the linear correlation filter applied. The CSK tracker uses the same intensity feature as the Gaussian kernel; therefore, the speed is slightly lower than that of MOSSE, but the accuracy is higher. The color name (CN) tracker [19], which is based on the CSK tracker, uses a feature that can express color properties well based on the Color Name [33]. As the dimension increases, the CN tracker proposes an updated model suitable for dimension reduction and high dimension feature through PCA. The scale adaptive with multiple features tracker [20] combines the histogram of gradient (HOG) feature with CN and also considers the change in size of the object by creating a pyramid scale pool. The discriminative scale space tracker (DSST) constructs a correlation filter with a three-dimensional correlation filter and proposes an effective tracking algorithm using a translation filter and a joint scale space filter. The KCF [16], an extended version of the CSK tracker, is the most widely used tracker that is currently employed because it offers high accuracy and speed. Therefore, this study, which is based on the KCF tracker, estimates the scale using the scale space and uses the highly effective multi-block scheme to ensure the tracker is robust to partial occlusion.

2.2. Part-Based Tracking

Various approaches have been used to overcome the problem of occlusion [24,25,26,27,28,29]. The part-based model is particularly robust to occlusion. For example, crowded scenes are characterized by occlusions of individual persons and Shu et al. [24] employed the part-based model with person-specific SVM classifiers to address the partial occlusion of persons. Zhang et al. presented a part-matching tracker [25] that is based on a locality-constrained low-rank sparse learning method among multiple frames. The online weighted MIL (WMIL) tracker is an enhancement of the MIL tracker [26]. WMIL determines the most important sample in the current frame and presents more efficient learning procedures. Others proposed a part-based model based on the correlation filter [24,25,26]. Osman et al. [27] used four parts based on the CSK tracker. Liu et al. [28] proposed a model based on the KCF tracker and particle filter and used Bayesian inference to merge the response map of difference parts. The method proposed by Yao et al. [29] is based on KCF tracker. It combines a response map using a graph and a minimum spanning tree.

3. Proposed Method

In this section, we propose our robust model to address occlusion and scale variation based on the KCF tracker [16], which has both impressive performance and speed. We briefly describe the KCF tracker and scale space filter of the pyramid searching method. Then, based on the size of the estimated scale, we explain our multi-block scheme for the part-based model. Finally, we explain the state-update scheme aims to improve the robustness of the results of each process.

3.1. The KCF Tracker

The KCF tracker [16] ranked high in the Visual Object Tracking challenge 2014 (VOT 2014) and has demonstrated impressive performance and speed as a correlation filter-based tracker. The goal of a correlation filter is to learn the filter h that minimizes the error from a given regression target. Therefore, the KCF tracker involves finding the optimal filter that solves the ridge regression problem in the spatial domain:
min h i = 1 n ( h T x i y i ) 2 + λ h 2 2
where y is the desired regression target, f ( x ) = h T x is the filter result that minimizes the squared error between samples x i and their regression targets y i , and λ is the regularization parameter in SVM to avoid overfitting. The closed-form solution of linear regression is h = ( X H X + λ I ) 1 X H y [34]. Since the correlation filter is performed in the frequency domain, the hermitian transpose X H is expressed instead of X T to handle the complex number. The non-linear regression was solved by using the kernel trick [35] because the dual space was problematic. Then, the kernelized version of the ridge regression solution is given by [29]:
α = ( K + λ I ) 1 y
where α is the represented vector [35] of filter h at dual space, K is a kernel matrix and I is the identity matrix. The n × n kernel matrix K can be written with elements K = κ ( x i , x j ) and expressed as K = C ( k x x ) owing to its circulant structure, as was demonstrated by Henriques et al. [16]. The kernel matrix can be diagonalized by DFT, and it can obtain the final kernel ridge regression solution as follows:
α = F 1 ( F ( y ) F ( k x x ) + λ )
where k x x is the kernel correlation of x . F and F 1 are Fourier and its inverse transform, respectively. We can also obtain the kernel correlation solution by using the circulant structure [16]:
k x x ' = exp ( 1 σ 2 ( x 2 + x ' 2 2 F 1 ( c F ( x c ) F ( x c * ) ) ) )
Radial Basis Function kernel is employed among the Mercer kernels and the HOG feature [36] is used. Owing to the linearity of DFT, a multi-channel correlation filter can be used for calculation by simply summing over them in the Fourier domain [16]. The regression function f ( z ) is calculated as follows:
R Z = f ( z ) = F 1 ( F ( k x z ) F ( α ) ) Z * = arg max Z R z
where k x z is the kernel correlation from Equation (4) between input sample x and appearance updated patch z. is an element-wise product operator. Then, the new frame can be estimated by finding the maximum value of the response map. For more details, readers are advised to refer to [15].

3.2. Scale Estimation Strategy via Scale Space Filter

The scale estimation method using DSST [15] is efficient from a view point of computation. In a new frame, the target translation is estimated by the translation filter. Subsequent to that, we estimate the accurate scale of the target size. In this study, the translation filter is replaced by global tracking in the proposed method, which is a multi-block process. Then, we estimate the scale using the scale space as follows:
{ r n = β n τ 2 S f = a r ( P × R ) , n = 1 , 2 , ... , τ
where τ is the number of the scale space, P is the width of patch, R is the height of patch, and a is the scale step. We extract the image patch of size a r ( P × R ) centered around the target corresponding to τ ; this is the scale function S f . The extracted scale space image is vectorized to one dimension. Then, we calculate the scale correlation between S f and the updated scale function. The scale correlation is defined as follows:
f s ( z ) = F ( l = 1 d N ˜ t 1 l X t l D t 1 + λ ) 1
where X t is the d-dimensional input sample of the current t frame and f s ( z ) is the scale correlation output. The accurate patch size is calculated by finding the maximum value of the scale correlation response. l { 1 , ... , d } is the feature channel. The numerator N t 1 l and denominator D t 1 are the terms introduced by the proposed updating process, which is a suitable multi-channel feature from DSST. The reader is advised to refer to [15] for further details.

3.3. Multi-Block Scheme for Partial Occlusion

In visual tracking, occlusion is frequently observed. The part-based model is robust against occlusion and deformation; however, it has relatively high complexity. Therefore, there is a trade-off between the performance and speed that has to be optimized for maximum efficiency and accuracy. As the complexity of the algorithm increases, its real-time applicability is hindered and becomes limited. Therefore, a combination of the high speed KCF tracker and the proposed simple multi-block scheme can be utilized for improved efficiency. The conventional part-based method combines the response maps from each part [27,28,29]. However, in case occlusion occurs, the conventional approach can average the error, and it does not know which block is reliable.
A global block is first used to cover the entire original target object. This global block is then divided into two parts, i.e., it becomes a multi-block, as shown in Figure 2. The splitting direction is simply determined by the ratio of the height and width of the target object. If the height is greater than the width, the sub-block is divided into upper and lower blocks. Otherwise, it is separated into left and right blocks. As shown in Figure 3, each set of sub-blocks overlaps:
S * = arg max S { 1 , 2 , 3 } ( max ( R 1 ) , max ( R 2 ) , max ( R 3 ) )
In this work, three response maps are generated from multi-blocks, and we need to select the proper block using Equation (8). If the response map has a lower peak value, the region may experience change of state such as occlusion and deformation. For robust tracking in the case of partial occlusion, we select the maximum response value among the three response maps as a new tracking point. Then, R s * is the newly selected block. If the selected block is one of sub-blocks, the next tracking region is shifted in correspondence to the previous center coordinates such that the original target object is covered.

3.4. Adaptive Update Model Using PSR

The appearance of an object changes in accordance with many different factors such as deformation and illumination. Moreover, the appearance update has a huge influence on the efficiency of tracking. In addition, it is necessary to update the correlation filter and to modify its learning rate adaptively according to the change in object shape appearance. The KCF tracker and many other correlation filter-based trackers use a simple interpolation-based update model, as:
{ z ˜ t = ( 1 ω ) z ˜ t 1 + ω z t α ˜ t = ( 1 ω ) α ˜ t 1 + ω α t
where ω is the learning rate, which has a fixed value of 0.02 in the conventional KCF tracker. It is affected more by the previous state than the present state, and thus, it is relatively sturdy against sudden changes. However, having a fixed value implies that updates do not occur actively according to the object appearance and correlation filter of the sequence. When anomalies such as occlusion or deformation occur, there is a high risk of not being able to manage such circumstances. Therefore, this paper uses the ratio of the predefined peak-to-sidelobe ratio (PSR) [17] of the desired output and the PSR from the proposed method as the adaptive rate in order to address these problems. The adaptive update model reflects the status of the target object when deformation, illumination change, or occlusion occurs. The PSR of the desired output is the optimal result, and thus, the ratio can be trusted entirely. In general, the PSR range of a KCF tracker is in between 3.0 and 15.0. Higher values produce a stronger peak and can return more accurate tracking results. However, when occlusion or other anomalies occur, the PSR value drops and the peak, which is presumed to be the positions of the object, can be difficult to presume as being the actual position. The learning rate proposed by utilizing the PSR can be expressed as:
{ ρ i = R i max μ i R i s t d γ i = ρ i ρ 0 × c , i = 1 , 2 , 3
The side lobe required for the calculation of the PSR was used as the overall size of the response map. In Equation (10), ρ is the PSR result for each block i, ρ 0 is the PSR result of the desired output, and c is the scaling factor. We obtain a new learning rate by calculating the ratio of these PSR results. Therefore, the appearance and correlation filter update are rewritten, respectively, as:
{ z ˜ t i = ( 1 γ i ) z ˜ t 1 i + γ i z t i α ˜ t i = ( 1 γ i ) α ˜ t 1 i + γ i α t i , i = 1 , 2 , 3
where γ determines the extent to which the current state of the object is reflected. In a normal translation, γ has a similar value; however, it has a low value when occlusion and deformation occur. This implies that the current state of the target object is reflected to a lesser extent than the previous state. We update the numerator N t 1 l and denominator D t 1 of the scale filter with a new sample X t as:
{ N t l = ( 1 γ s ) N t 1 l + γ s Y X ˜ t l D t = ( 1 γ s ) D t 1 + γ s k = 1 d X t k X ˜ t k
In this paper, the updating scale filter is based on Equation (11). The learning rate of the scale filter is determined by the selected block γ s . Figure 4 shows the adaptive learning rate to the state of the changing object.

4. Experiments

The two experiments were conducted to evaluate the precision and success rate of our proposed tracker, the proposed algorithm compared with the state-of-art trackers with challenging sequences in terms of quantitative and qualitative measures.

4.1. Experimental Setup

Each of the algorithms was implemented in MATLAB to evaluate their performance. The computer hardware comprised a Core i5 CPU with 16 GB RAM. We evaluated our proposed method on a commonly used Visual Tracker Benchmark 100 dataset [37], which has several attributes (almost 59,000 frames), such as illumination variation, deformation, scale variation, and occlusion. These attributes affect the performance of the tracking algorithm.

4.2. Features and Parameters

FHOG [36] feature was used for image representation and its implementation methodology was provided by [38]. The HOG cell size is 4 × 4 and the number of orientation bins is nine. To mitigate the boundary effect, the extracted features are multiplied by a cosine window. The basic parameters are used in a manner identical to the KCF tracker. The search range is 2.5 times the target object, and the initial learning rate ω is 0.02 that is adaptively changed at every frame. The σ used in Gaussian kernel is assigned to 0.5. The scale pool S is 33, the step size is set to 1.02, and the scaling factor c for learning rate is 0.01.

4.3. Evaluation Methodology

We apply One-Pass Evaluation (OPE), which is a traditional evaluation method used from the Object Tracker Benchmark (OTB), from the first frame to the last frame of the sequence. Two criteria, namely the distance precision and success rate, are employed for quantitative evaluations [37]:
Precision: the center location error (CLE) is a widely used measure for evaluating tracking performance. CLE calculates the distance between the center coordinate of the bounding box and the ground-truth. The precision is defined by the percentage of the CLE result belonging to a specific range, and the numeric value 20 is assigned to the basic threshold in practice.
Success Rate: As another measure, an overlap score from Pascal VOC overlap ratio (VOR) [39], which is defined as: o = | r t g t | / | r t g t | , is used. We calculate the overlapped area as the extent to which the tracking output bounding box r t and ground-truth bounding box g t overlap, where | | indicates the area. Compared to simple precision, which involves determining the difference from the ground truth, this method is more accurate because it finds and evaluates the overlap area. In the test we used a threshold of 0.5 to calculate the success rate and the area under the curve (AUC).

4.4. Results

We use two criteria, the distance precision and success rate, as quantitative evaluations metrics [38].

4.4.1. Quantitative Evaluation

The proposed algorithm is compared with the following correlation filter-based trackers and OTB trackers. Correlation filter-based trackers include CSK [18], CN [19], DSST [15], KCF [16], and SKCF that is the same as the KCF tracker except for applying only the scale space. The results we obtained by testing the precision, CLE, and VOR score on 100 sequences of OTB are presented in Table 1. The proposed method provided the improved results compared to other algorithms. We observed 4% improvement on the VOR score compared to DSST and a 10% increase compared to KCF. Figure 5 shows the graphical results from both the correlation filter-based and OTB trackers. As for OTB trackers, we tested the ASLA [40], BSBT [41], CPF [42], CT [43], CXT [44], DFT [45], FRAG [5], IVT [1], KMS [46], LOT [47], MIL [13], MS [48], OAB [9], PD [48], RS [48], SCM [49], STRUCK [12], TM [48], VTD [3], and VTS [50]. Including PCOM [2] where partial occlusion was used as the target, we compared our proposed method with a total of 29 trackers, and as can be seen in Figure 5 and Figure 6, Table 1 and Table 2, the proposed method showed the most promising results. In terms of speed, CSK, which only used intensity features, was the fastest followed by KCF and CN. We discovered that the proposed method was more time consuming due to its need for additional scale estimation and the multi-block method. However, since the proposed method is based on the correlation filter, it continues to be faster than all of the other latest trackers.

4.4.2. Qualitative Evaluation

The factors of occlusion, scale variation and above these illumination variation, deformation, and fast motion, affect the performance of the tracking algorithms. Scale variation implies a change in the target size. In Figure 7, the images Singer1, Dog1, and Human4, are typical sequences with the scale variation attribute. However, in Figure 8, the Walking2 and Human6 sequences have scale variation and partial occlusion at the same time. Thus, each of the tracking attributes exists in a complex manner. Among the attributes, occlusion occurs frequently in tracking. Heavy occlusion implies that the object is covered in its entirety; therefore, it is difficult to control with tracking. On the other hand, partial occlusion occurs when regions of the object remain visible, and therefore, in this case tracking remains possible. In Figure 8, the target in the video FaccOcc1 is partially occluded. In the Walking2 sequence, the target is covered by a walking man, but approximately one-third of the target object remains visible. Regions such as this that remain partially visible throughout a sequence of images are considered reliable regions and are selected by the proposed multi-block model. Thus, the tracking result for the Walking2 sequence was successful, whereas in the Struck and VTD sequences, the tracking algorithm loses the woman at times during which she is occluded by the man, but approximately one-third of the woman remains visible. Human3, Human4, and Human6 are outdoor sequences. These outdoor images are frequently affected by partial occlusion, scale variation, and fast motion. In Figure 7 and Figure 8, the results show that the tracking procedure of the proposed method is more successful than any other method. Figure 9 presents a comparison of the most successful state-of-the-art trackers. Each sequence includes plural attributes. This resulted in degraded performance, even though the method is robust against occlusion. The proposed algorithm is able to overcome occlusion and scale variation, and outperforms other trackers.
Figure 10 shows the probability of selection of each block or sub-block. In the David3 and Walking sequences, sub-block 2 has a very small likelihood of being selected, because in the sequence of images showing these people walking, the lower bodies continue moving, which implies there are several instances in which deformation occurs. On the other hand, if only the upper body experiences movement, sub-block 1 is not selected, as is the case with the Singer1 sequence. The SUV sequence has frequent occlusion from side to side. Therefore, all blocks are selected.
We conducted the experiment using center location error (CLE) to prove the performance of the proposed method. The Graph in Figure 11 shows that the proposed method has a low CLE in sequences containing the attributes of scale variation, occlusion, or deformation.

5. Conclusions

This paper proposed simple multi-block-based scale space for kernelized correlation filters (MSKCF) capable of efficiently overcoming occlusion and scale variation in visual tracking. We achieved robust partial occlusion and scale variation by employing a multi-block method and scale space. The overall robustness of the system is improved by using an adaptive learning rate for appearance and scale updates with the use of occlusion detection through the distribution of the response map. The experimental results showed that the proposed method outperforms the other trackers in terms of precision and VOR score on average for all OTB 100 sequences. In particular, the proposed scheme achieved an improvement of 8% and 18% in the results compared to the KCF tracker for 49 occlusion and 64 scale variation sequences, respectively.

Acknowledgments

This research was supported by the Chung-Ang University Research Scholarship Grants in 2015 and the Commercialization Promotion Agency for R&D Outcomes (COMPA) (No. 2016K000202).

Author Contributions

Soowoong Jeong and Guisik Kim designed the main algorithm and the experiments under the supervision of Sangkeun Lee. Soowoong Jeong and Guisik Kim wrote the paper and the experimental results were analyzed by Soowoong Jeong. Soowoong Jeong edited the final document. All authors participated in discussions on the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ross, D.A.; Lim, J.; Lin, R.; Yang, M. Incremental learning for robust visual tracking. IJCV 2008, 77, 125–141. [Google Scholar] [CrossRef]
  2. Wang, D.; Lu, H. Visual Tracking via Probability Continuous Outlier Model. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014.
  3. Kwon, J.; Lee, K.M. Visual tracking decomposition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010.
  4. Mei, X.; Ling, H. Robust Visual Tracking using L1 Minimization. In Proceedings of the IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 1436–1443.
  5. Adam, A.; Rivlin, E.; Shimshoni, I. Robust Fragments-Based Tracking Using the Integral Histogram. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Washington, DC, USA, 17–22 June 2006.
  6. Zhang, T.; Ghanem, B.; Liu, S.; Ahuja, N. Robust Visual Tracking via Multi-task Sparse Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012.
  7. Zhang, T.; Ghanem, B.; Liu, S.; Ahuja, N. Low-rank sparse learning for robust visual tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Florence, Italy, 7–13 October 2012.
  8. Zhang, T.; Bibi, A.; Ghanem, B. In defense of sparse tracking: Circulant sparse tracker. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 3880–3888.
  9. Avidan, S. Ensemble Tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 261–271. [Google Scholar] [CrossRef] [PubMed]
  10. Grabner, H.; Grabner, M.; Bischof, H. Real-Time tracking via on-line boosting. In Proceedings of the British Machine Vision Conference, Edinburgh, UK, 4–7 September 2006.
  11. Saffari, A.; Leistner, C.; Santner, J.; Godec, M.; Bischof, H. On-line random forests. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Miami, FL, USA, 20–25 June 2009.
  12. Hare, S.; Saffari, A.; Torr, P.H.S. Struck: Structured Output Tracking with Kernels. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011.
  13. Babenko, B.; Yang, M.-H.; Belongie, S. Visual Tracking with Online Multiple Instance Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Miami, FL, USA, 20–25 June 2009.
  14. Kalal, Z.; Matas, J.; Mikolajczyk, K. Tracking-Learning-Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 1409–1422. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Danelljan, M.; Gustav, H.; Khan, S.F.; Felsberg, M. Accurate scale estimation for robust visual tracking. In Proceedings of the British Machine Vision Conference, Nottingham, UK, 1–5 September 2014.
  16. Henriques, F.; Caseiro, R.; Martins, P.; Batista, J. High-Speed Tracking with Kernelized Correlation Filters. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 583–596. [Google Scholar] [CrossRef] [PubMed]
  17. Bolme, D.S.; Beveridge, J.R.; Draper, B.A.; Lui, Y.M. Visual object tracking using adaptive correlation filters. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010.
  18. Henriques, F.; Caseiro, R.; Martins, P.; Batista, J. Exploiting the Circulant Structure of Tracking-by-Detection with Kernels. In Proceedings of the European Conference on Computer Vision (ECCV), Florence, Italy, 7–13 October 2012.
  19. Danelljan, M.; Khan, F.S.; Felsberg, M.; van de Weijer, J. Adaptive Color Attributes for Real-Time Visual Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014.
  20. Li, Y.; Zhu, J. A scale adaptive kernel correlation filter tracker with feature integration. In Lecture Notes in Computer Science; Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics; Springer: Berlin, Germany, 2015; Volume 8926, pp. 254–265. [Google Scholar]
  21. Danelljan, M.; Gustav, H.; Khan, F.S.; Felsberg, M. Learning Spatially Regularized Correlation Filters for Visual Tracking. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 58–66.
  22. Ruan, Y.; Wei, Z. Real-Time Visual Tracking through Fusion Features. Sensors 2016, 16, 949. [Google Scholar] [CrossRef] [PubMed]
  23. Bertinetto, L.; Valmadre, J.; Golodetz, S.; Miksik, O.; Torr, P. Staple: Complementary Learners for Real-Time Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 3880–3888.
  24. Shu, G.; Dehghan, A.; Oreifej, O.; Hand, E.; Shah, M. Part-based multiple-person tracking with partial occlusion handling. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012.
  25. Zhang, T.; Jia, K.; Xu, C.; Ma, Y.; Ahuja, N. Partial occlusion handling for visual tracking via robust part matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014.
  26. Zhang, K.; Song, H. Real-time visual tracking via online weighted multiple instance learning. Pattern Recognit. 2013, 46, 397–411. [Google Scholar] [CrossRef]
  27. Akın, O.; Mikolajczyk, K. Online Learning and Detection with Part-based Circulant Structure. In Proceedings of the IEEE International Conference on Pattern Recognition (ICPR), Washington, DC, USA, 24–28 August 2014.
  28. Liu, T.; Wnag, G.; Yang, Q. Real-time part-based visual tracking via adaptive correlation filters. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015.
  29. Yao, R.; Xia, S.; Shen, F.; Zhou, Y.; Niu, Q. Exploiting Spatial Structure from Parts for Adaptive Kernelized Correlation Filter Tracker. IEEE Signal Process. Lett. 2016, 23, 658–662. [Google Scholar] [CrossRef]
  30. Zhang, K.; Liu, Q.; Wu, Y.; Yang, M.-H. Robust Visual Tracking via Convolutional Networks without Training. IEEE Trans. Image Process. 2016, 25, 1779–1792. [Google Scholar] [CrossRef] [PubMed]
  31. Gao, J.; Ling, H.; Hu, W.; Xing, J. Transfer learning based visual tracking with gaussian processes regression. In Lecture Notes in Computer Science; Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics; Springer: Berlin, Germany, 2015; Volume 8926, pp. 188–203. [Google Scholar]
  32. Zhang, J.; Ma, S.; Sclaroff, S. MEEM: Robust tracking via multiple experts using entropy minimization. In Lecture Notes in Computer Science; Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics; Springer: Berlin, Germany, 2015; Volume 8926, pp. 188–203. [Google Scholar]
  33. Van De Weijer, J.; Schmid, C.; Verbeek, J.; Larlus, D. Learning color names for real-world applications. IEEE Trans. Image Process. 2009, 18, 1512–1523. [Google Scholar] [CrossRef] [PubMed]
  34. Rifkin, R.; Yeo, G.; Poggio, T. Regularized least-squares classification. Nato Sci. Ser. Sub Series III Comput. Syst. Sci. 2003, 190, 131–154. [Google Scholar]
  35. Scholkopf, B.; Smola, A.J. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond; The MIT Press: Cambridge, MA, USA, 2002. [Google Scholar]
  36. Felzenszwalb, P.F.; Girshick, R.B.; McAllester, D.; Ramanan, D. Object detection with discriminatively trained part-based models. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1627–1645. [Google Scholar] [CrossRef] [PubMed]
  37. Wu, Y.; Lim, J.; Yang, M.H. Online Object Tracking: A Benchmark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013.
  38. Piotr’s Toolbox. Available online: https://pdollar.github.io/toolbox/ (accessed on 3 November 2016).
  39. Everingham, M.; Gool, V.L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. IJCV 2010, 88, 303–338. [Google Scholar] [CrossRef]
  40. Jia, X.; Lu, H.; Yang, M.-H. Visual Tracking via Adaptive Structural Local Sparse Appearance Model. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012.
  41. Severin, S.; Grabner, H.; Gool, L.V. Beyond semi-supervised tracking: Tracking should be as simple as detection, but not simpler than recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Miami, FL, USA, 20–25 June 2009.
  42. Pérez, P.; Hue, C.; Vermaak, J.; Gangnet, M. Color-based probabilistic tracking. In Proceedings of the European Conference on Computer Vision (ECCV), London, UK, 28–31 May 2002.
  43. Zhang, K.; Zhang, L.; Yang, M.H. Fast compressive tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 2002–2015. [Google Scholar] [CrossRef] [PubMed]
  44. Dinh, T.B.; Vo, N.; Medioni, G. Context tracker: Exploring supporters and distracters in unconstrained environments. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 20–25 June 2011.
  45. Laura, S.-L.; Erik, L.-M. Distribution fields for tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012.
  46. Comaniciu, D.; Visvanathan, R.; Peter, M. Kernel-based object tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 564–577. [Google Scholar] [CrossRef]
  47. Shaul, O.; Aharon, B.H.; Dan, L.; Shai, A. Locally orderless tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012.
  48. Collins, R.T.; Liu, Y.; Leordeanu, M. Online selection of discriminative tracking features. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1631–1643. [Google Scholar] [CrossRef] [PubMed]
  49. Zhong, W.; Lu, H.; Yang, M.H. Robust object tracking via sparsity-based collaborative model. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012.
  50. Kwon, J.; Lee, K.M. Tracking by sampling trackers. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011.
Figure 1. Tracking results with state-of-the-art trackers. Top two rows are occlusion sequences and bottom two rows are scale variation sequences. These screen shots were acquired to illustrate situations of occlusions and scale variations.
Figure 1. Tracking results with state-of-the-art trackers. Top two rows are occlusion sequences and bottom two rows are scale variation sequences. These screen shots were acquired to illustrate situations of occlusions and scale variations.
Sensors 17 00433 g001
Figure 2. Proposed multi-block model. The red-dashed rectangle represents the global block that covers the target object. The blue and green dot-dashed rectangles represent sub-blocks that are divided by the height and width ratio of the target object.
Figure 2. Proposed multi-block model. The red-dashed rectangle represents the global block that covers the target object. The blue and green dot-dashed rectangles represent sub-blocks that are divided by the height and width ratio of the target object.
Sensors 17 00433 g002
Figure 3. Procedure of the proposed method. First, we perform the scale estimation from the global tracking results. Then, we divide the selected region into two blocks using the proposed multi-block scheme, and apply the feature extract function ϕ ( ) to each block. Subsequently, we calculate the correlation filter responses.
Figure 3. Procedure of the proposed method. First, we perform the scale estimation from the global tracking results. Then, we divide the selected region into two blocks using the proposed multi-block scheme, and apply the feature extract function ϕ ( ) to each block. Subsequently, we calculate the correlation filter responses.
Sensors 17 00433 g003
Figure 4. Proposed adaptive update method. The graphs represent the changing learning rate of the occurrence of occlusion and deformation. The method can roughly detect the change of state of the target object.
Figure 4. Proposed adaptive update method. The graphs represent the changing learning rate of the occurrence of occlusion and deformation. The method can roughly detect the change of state of the target object.
Sensors 17 00433 g004
Figure 5. Precision (a) and success (b) plots over all 100 sequences using one pass evaluation for correlation filter-based trackers [14,15,17,18].
Figure 5. Precision (a) and success (b) plots over all 100 sequences using one pass evaluation for correlation filter-based trackers [14,15,17,18].
Sensors 17 00433 g005
Figure 6. Precision (a) and success (b) plots over all 100 sequences using OPE for 29 trackers in [38]. The results of the top five ranks are indicated via the legend in the figure, and the other trackers are indicated using gray colored lines.
Figure 6. Precision (a) and success (b) plots over all 100 sequences using OPE for 29 trackers in [38]. The results of the top five ranks are indicated via the legend in the figure, and the other trackers are indicated using gray colored lines.
Sensors 17 00433 g006
Figure 7. Tracking results under scale variations. (a) Singer1; (b) Surfer; (c) Dog1; and (d) Human4.
Figure 7. Tracking results under scale variations. (a) Singer1; (b) Surfer; (c) Dog1; and (d) Human4.
Sensors 17 00433 g007
Figure 8. Tracking results for partial occlusions for the sequences (a) Walking2; (b) FaceOcc1; (c) Human6; (d) Girl; (e) Subway; and (f) Rubik.
Figure 8. Tracking results for partial occlusions for the sequences (a) Walking2; (b) FaceOcc1; (c) Human6; (d) Girl; (e) Subway; and (f) Rubik.
Sensors 17 00433 g008
Figure 9. Average VOR ranking scores of the five most successful state-of-the-art trackers [15,16,31,32]. The OTB [38] 100 sequences are annotated with attributes such as illumination variation (IV), out-of-plane rotation (OPR), scale variation (SV), occlusion (OCC), deformation (DEF), motion blur (MB), fast motion (FM), in-plane rotation (IPR), out-of-view (OV), background clutter (BC), and low resolution (LR). The number next to each attribute indicates the number of sequences with this attribute.
Figure 9. Average VOR ranking scores of the five most successful state-of-the-art trackers [15,16,31,32]. The OTB [38] 100 sequences are annotated with attributes such as illumination variation (IV), out-of-plane rotation (OPR), scale variation (SV), occlusion (OCC), deformation (DEF), motion blur (MB), fast motion (FM), in-plane rotation (IPR), out-of-view (OV), background clutter (BC), and low resolution (LR). The number next to each attribute indicates the number of sequences with this attribute.
Sensors 17 00433 g009
Figure 10. Distribution of selected blocks according to the image sequence. The y-axis represents the probability of each block being selected.
Figure 10. Distribution of selected blocks according to the image sequence. The y-axis represents the probability of each block being selected.
Sensors 17 00433 g010
Figure 11. Center location error of each frame. The sequences contain challenging attributes such as scale variation, occlusion, and deformation.
Figure 11. Center location error of each frame. The sequences contain challenging attributes such as scale variation, occlusion, and deformation.
Sensors 17 00433 g011
Table 1. Quantitative comparison of the proposed tracker with correlation filter-based trackers over all 100 challenging dataset. The high score indicates the best performance among the algorithms.
Table 1. Quantitative comparison of the proposed tracker with correlation filter-based trackers over all 100 challenging dataset. The high score indicates the best performance among the algorithms.
PrecisionCLEVORVOR (AUC)FPS
CSK [18]51.84304.600.41330.3817455
CN [19]60.0482.480.47810.4220220
DSST [15]69.5048.310.61580.524834
KCF [16]69.6044.730.55210.4782260
SKCF68.2346.110.59700.501072
MSKCF71.1746.300.65000.529052
Table 2. Quantitative comparison of our tracker with OTB trackers over all 100 challenging dataset. The high score indicates the best performance among the algorithms.
Table 2. Quantitative comparison of our tracker with OTB trackers over all 100 challenging dataset. The high score indicates the best performance among the algorithms.
PrecisionCLEVORVOR (AUC)
STRUCK [12]63.8447.070.51890.4618
SCM [45]56.8062.020.43220.3982
VTD [3]51.1961.770.39150.3536
CXT [44]55.2467.420.43260.3887
CSK [18]51.84304.600.41330.3817
OAB [10]47.9570.300.40310.3618
IVT [1]43.1788.110.34190.3076
FRAG [5]42.4480.660.35860.3308
ASLA [40]51.1468.100.38630.3600
MSKCF71.1746.300.65000.5290

Share and Cite

MDPI and ACS Style

Jeong, S.; Kim, G.; Lee, S. Effective Visual Tracking Using Multi-Block and Scale Space Based on Kernelized Correlation Filters. Sensors 2017, 17, 433. https://doi.org/10.3390/s17030433

AMA Style

Jeong S, Kim G, Lee S. Effective Visual Tracking Using Multi-Block and Scale Space Based on Kernelized Correlation Filters. Sensors. 2017; 17(3):433. https://doi.org/10.3390/s17030433

Chicago/Turabian Style

Jeong, Soowoong, Guisik Kim, and Sangkeun Lee. 2017. "Effective Visual Tracking Using Multi-Block and Scale Space Based on Kernelized Correlation Filters" Sensors 17, no. 3: 433. https://doi.org/10.3390/s17030433

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop