Next Article in Journal
Recent Advances in ASIC Development for Enhanced Performance M-Sequence UWB Systems
Previous Article in Journal
Non-Carbon 2D Materials-Based Field-Effect Transistor Biosensors: Recent Advances, Challenges, and Future Perspectives
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Proposal-Based Visual Tracking Using Spatial Cascaded Transformed Region Proposal Network

1
Faculty of Space, Xi’an Institute of Optics and Precision Mechanics of CAS, Xi’an 710119, China
2
School of Astronautics, Northwestern Polytechnical Universty, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(17), 4810; https://doi.org/10.3390/s20174810
Submission received: 20 July 2020 / Revised: 13 August 2020 / Accepted: 14 August 2020 / Published: 26 August 2020
(This article belongs to the Section Sensor Networks)

Abstract

:
Region proposal network (RPN) based trackers employ the classification and regression block to generate the proposals, the proposal that contains the highest similarity score is formulated to be the groundtruth candidate of next frame. However, region proposal network based trackers cannot make the best of the features from different convolutional layers, and the original loss function cannot alleviate the data imbalance issue of the training procedure. We propose the Spatial Cascaded Transformed RPN to combine the RPN and STN (spatial transformer network) together, in order to successfully obtain the proposals of high quality, which can simultaneously improves the robustness. The STN can transfer the spatial transformed features though different stages, which extends the spatial representation capability of such networks handling complex scenarios such as scale variation and affine transformation. We break the restriction though an easy samples penalization loss (shrinkage loss) instead of smooth L1 function. Moreover, we perform the multi-cue proposals re-ranking to guarantee the accuracy of the proposed tracker. We extensively prove the effectiveness of our proposed method on the ablation studies of the tracking datasets, which include OTB-2015 (Object Tracking Benchmark 2015), VOT-2018 (Visual Object Tracking 2018), LaSOT (Large Scale Single Object Tracking), TrackingNet (A Large-Scale Dataset and Benchmark for Object Tracking in the Wild) and UAV123 (UAV Tracking Dataset).

1. Introduction

Visual tracking has drawn constant attention of the researchers and engineers over last decades. Some novel applications are also inspired by the improvement of related research, such as auto-track by drone [1], pose recognition by mobile payment [2], and remote control by space robot [3]. Although the researchers are making much progress persistently, it is still a vital problem to achieve a tracking procedure that simultaneously balances the accuracy, robustness, and tracking speed under complex scenarios, such as occlusion, illumination change, and scale variation, to name a few [4].
Much progress [5] has made by the combined region proposal networks (RPN) and Siamese networks recently [6]. Some of the trackers treat the tracking problems as the generation of the similarity response map, which could distinguish the differences between the target templates and the search candidates. The position candidates where reach the highest similarity score is performed as the new target groundtruth. SiamRPN [7] combines Siamese networks and region proposal networks in order to jointly perform classification and regression for tracking. The DaSiamRPN [8] comes up with the distractor-aware module to distinguish hard negatives from easy ones, which could improve the discriminating power of such model. The other methods formulate the tracking problems as tracking-by-detection problems, which firstly generate the proposals from the search area, then calculate the classification score between proposal candidates and template target. The SiamRPN++ [9] introduces a ResNet-driven Siamese tracker, which makes layer-wise and depth-wise aggregations explicit when modeling network architecture, which can improve the accuracy and reduce the model size at the same time. The DCFNet [10] further combines the Siamese network with region proposal networks, and performs the domain specific updating to achieve a light-weight network in end-to-end learning.
There are still a few vital problems have not been settled during the tracking procedure though the region proposal networks based trackers introduced above have achieved excellent performance both in accuracy and speed. We notice that most trackers still employ the semantic features though the networks which cannot ensure the tracking accuracy which is shown in Figure 1. During target localization, One-stage region proposal network [11] proposes to perform a single regressor depends on pre-defined anchor boxes, while it is still difficult to estimate the changing scale of target [12]. The RPN-based trackers applies the log loss function as classification loss function which cannot effectively reduce the easy samples though training stage. Since the proposals generated by the proposed networks need to be strictly screened, the more effective proposals re-ranking method need to be proposed.
To overcome the restriction, we introduce a simple spatial cascaded strategy to apply the different hierarchical features from convolutional layers. By analyzing the feature transfer function of the spatial transformer networks (SPN), we successfully perform it as feature extraction model. The localization network of STN can obtain the position of target in image frame accurately, which helps us to solve the spatial transformation problems when suffering from heavy scale change and rotation. Benefiting from the shrinkage loss, we could penalize the weights of easy samples to alleviate the data imbalance issue. Considering the redundancy of the proposals, we find that multi-cue such as shape, color, and scale can be applied to refine the high-quality proposals that can not only improve the tracking performance in complex scenarios, but also reduce the computational effort. We name our proposed method as SCTRPN (Spatial Cascaded Transformed Region Proposal Network).
To summarize, the main contributions of this work are threefold and are listed below:
1. We present the spatial cascaded region proposals networks that combine region proposal networks and spatial transformer networks. In this circumstance, the deep and shallow layers’ features can be extracted to represent the appearance and semantic characteristic of a certain target.
2. We propose to employ the shrinkage loss to penalize the weighting coefficient of easy samples during the training procedure of the proposed networks to successfully reduce the samples imbalance issue.
3. We provide the multi-cue proposals re-ranking method for the tracking frameworks, which helps to refine the high quality proposals from the candidates.
The rest of the paper is organized as follows: in Section 2, we illustrate the related research work of our proposed method. In Section 3, we describe the proposed tracking framework in details, including network structure, feature extraction model, loss function design, and a proposals ranking strategy. In Section 4, we perform the numerous experimental results on the tracking dataset. In Section 5, we summarize the conclusions drawn from our proposed method.

2. Related Work

Visual tracking based on deep convolutional networks have shown significant potentials in recent decades. In the following section, we discuss the most relevant work, and refer readers to [13,14] for recent surveys.

2.1. Deep Tracking

At their very beginning, deep neural networks were introduced to deal with the image recognition problem [15]. Inspired by its successes, researchers are paying more attention to the application of CNN frameworks in visual tracking. Wang et al. [16] proposed to employ the fully convolutional network to accomplish a tracking procedure with an improved accuracy of estimated bounding box. Danelljan et al. [17] proposed a continuous convolution operator to combine the discrete features with the deep features, and achieve the efficient integration of deep feature maps by training spatial continuous convolution filters. Danelljan et al. [18] also proposed a factorized convolution operation to obtain an efficient convolution operator (ECO) for visual tracking, in order to prevent the low efficiency caused by the convolutional neural network. Song et al. [19] performed different kinds of adversarial networks to generate variable samples, which helped to identify richer representation for tracking. Fiaz et al. [20] proposed a soft mask feature fusion mechanism, which can be easily integrated into the conventional Siamese tracking framework to enhance the discriminative capability when distinguish the target from the background. Gordon et al. [21] introduced the real-time recurrent regression networks to combine the multiple appearance features and motion information together, then perform the spatial-temporal fusion to accomplish a tracking network that increases the precision of the tracking results. Guo et al. [22] proposed an effective online update mechanism using the dynamic Siamese matching strategy, and the FFT (Fast Fourier Transform) acceleration can ensure the real-time processing.

2.2. Tracking through Region Proposal Networks

Region proposal networks (RPN) draw much attention as an effective approach when faced with target detection problems. They also have increasing potential for visual tracking for its object classification and bounding box regression function. [23].
As far as we known, Li et al. [7] first interpolated the region proposal network to the Siamese network, and obtained the one-stage Siamese-RPN tracker to gain tracking performance. Zhu et al. [8] introduced the utilization of much more negative samples to train a distractor-aware Siamese-RPN tracker. Despite the fact that the distractor-aware tracker has achieved a significant improvement, it requires large extra training data from other computer vision datasets.

2.3. Tracking though Multiple Features Fusion

The multi-layer features strategy had made outstanding progress through the visual tracking procedure. The features from shallow layers contains more spatial information [24]. On the other hand, the deep layer features represent more semantic cues. Benefitting from the multi-features, tracking can perform the balanced spatial accuracy and robustness. In order to learn multiple correlation filters, Ma et al. [25] extracted the hierarchical convolutional features from three different layers of both deep and shallow networks. Wang et al. [16] proposed to employ two regression models, which contain the features from hierarchical convolutional layers to distinguish similar distractors.

2.4. Loss Function Variation for Data Imbalance

The loss function acts an important role in deep convolutional tracker by solving the data imbalance problem [26], though little attention had been paid to this kind of issue [27]. So far, the cost-sensitive loss [28] is proven to be an effective approach when suffering data imbalance. When pre-training the Siamese networks, Bertinetto et al. [29] proposed to balance the loss of positives and negatives in order to improve the discriminative ability of the network. Li et al. [30] used a temporal sampling scheme to balance positive and negative samples to facilitate CNN training.

2.5. Our Approach

In this paper, we introduce the multi-stage spatial cascaded region proposal networks to generate the high-quality proposals. The baseline of our proposed method is one-stage region proposal network, which does not take data imbalance problem into consideration. On the contrary, our method proposes the spatial cascaded frameworks mainly to address the problem and filter the easy samples. We also introduce spatial transformer networks (STN) to perform the feature extraction and transfer procedure, which helps to improve the spatial transformer robustness. The shrinkage loss is also utilized to constrain the weights of easy samples during the training procedure. In addition, the multi-cue proposal re-ranking method is proposed to effectively refine the proposal candidates, improving the capability of maintaining the hard samples, which can distinguish the target from complex scenario. We name our proposed method SCTRPN. Figure 2 shows the flow diagram of our proposed method.

3. Proposed Method

3.1. Spatial Cascaded Region Proposal Networks

3.1.1. One-Stage Region Proposal Network

Before describing SCTRPN, we first represent the one-stage region proposal network [23], including both classification and regression branches. The anchors are be obtained by the network structure, which is shown in Figure 3.
For training the one-stage region proposal network, we first assign to each anchor the binary class label, including the positives and negatives. The positives contain the anchors that has an IoU (intersection-over-union) overlap higher than 0.7 with any groundtruth bounding box. The negatives represent the anchors that has an IoU lower than 0.3 with the groundtruth. The positives and negatives are applied to the training objective.
During the training procedure, we can obtain the classification scores { p i } and the regression offsets { t i } for each anchor by computing L ( { p i } , { t i } ) . We minimize the loss function based on the multi-task loss in Fast R-CNN [31]. Thus, the loss function for one-stage region proposal network is defined as,
L ( { p i } , { t i } ) = 1 N c l s i L c l s ( p i , p i * ) + λ 1 N r e g i p i * L r e g ( t i , t i * ) ,
where i represents the index of the anchor, and p i is the probability of anchor i , which represents the object. When the anchor belongs to the positives, the label p i * is assigned to 1. The anchor is negative if the label is 0. t i represents the 4 coordinates of the predicted bounding box, and t i * is the groundtruth box, which is affiliated with the positive anchor. The classification loss L c l s is log loss between the object and non-object. The regression loss is represented by L r e g ( t i , t i * ) = R ( t i t i * ) , where R represents the smooth L 1 loss function which is defined in [32]. The term p i * L r e g means that the regression loss can be activated when p i * equals to 1, and is disabled when p i * equals to 0. The outputs of the c l s and r e g layers consist of { p i } and { t i } respectively. The one-stage region proposal network is illustrated in detail in [20]. It can be employed to obtain the proposals for a visual tracker. When obtaining the proposals, we can perform the tracking procedure by calculating the maximum similarity score from the proposal candidates and target. However, the proposals that were obtained by the one-stage region proposal network usually cannot meet the requirements of trackers due to its low recall, which makes the tracker drift from the complex scenarios.

3.1.2. The Proposed Networks

The previous RPN-based trackers only employ the high-level semantic features from the last layer, which leads to the class imbalance. The phenomenon may result in unpromising performance when suffering from similar distractors. Faced with these problems, we introduce the multi-stage tracking framework, which is able to combine a set of L ( L N ) RPNs for the proposals generation.
For the l t h ( 1 < l L ) stage R P N l , it receives fused features Φ l ( x ) of the conv- l layer and the high-level layers from feature extraction model (FEM), instead of features φ l ( x ) from a single separate layer [7]. The Φ l ( x ) are obtained as follows,
Φ l ( x ) = F E M ( Φ l 1 ( x ) , φ l ( x ) ) ,
where denotes the F E M ( , ) , as described in Section 3.2. For R P N 1 , Φ 1 ( x ) = φ 1 ( x ) . We can obtain the related classification scores { p i l } and regression offsets { t i l } for the specific anchors in stage l as follows,
{ p i l } = L c l s s h r ( Φ l ( x ) ) { t i l } = L r e g ( Φ l ( x ) ) ,
where L c l s s h r ( Φ l ( x ) ) classification loss function L c l s (shrinkage loss), which is illustrated in Section 3.3 in detail, and L r e g ( Φ l ( x ) ) are achieved by accomplishing the convolutional operations on Φ l ( x ) .
Assuming that A l represents the anchor set in stage l . The negative anchors A l can be filtered out by threshold θ according to the classification scores { p i l } , we screen the anchors whose confidences are larger than the pre-defined threshold. We then achieve the positive anchors into a new set of anchors A l + 1 . The positives are mainly utilized to train the networks. The initialization of the regression branch has a great influence on accurate anchor localization. In our method, we obtain the refined anchors A l + 1 by the the regression results { t i l } . Compared with the one-stage regression [23,25], the cascaded structure improved the accurate localization when transferring between different stages, as illustrated in Figure 4. We can see from Figure 4 that the results achieved by the response map of deep stage are closer to the center of the tracking target.
The loss function R P N l for R P N l is composed of classification loss function L c l s s h r (shrinkage loss) and regression loss function L l o c (smooth L 1 loss), which is shown in Equation (4),
R P N l ( { p i l } , { t i l } ) = i L c l s s h r ( p i l , p i l * ) + ξ i p i l * L l o c ( t i l , t i l * ) ,
where i is the anchor index in A l of stage l , ξ is proposed to balance the classification and regression loss. By default, we set ξ = 1 , and thus, both L c l s s h r and L l o c are roughly equally weighted. We show by experiments that the training results are insensitive to the values of ξ . p i l * represents the groundtruth label of anchor i , and t i l * represents the distance between anchor i and groundtruth. Following [23], t i l * = ( t i ( x ) l * , t i ( y ) l * , t i ( w ) l * , t i ( h ) l * ) is a 4d vector, such that
t i ( x ) l * = ( x * x a l ) / w a l         t i ( y ) l * = ( y * y a l ) / h a l t i ( w ) l * = log ( w * / w a l )       t i ( h ) l * = log ( y * / h a l ) ,
where x , y , w , and h represent the center of the tracking bounding box and its width and height, respectively. x * and x a l are for the groundtruth and anchor of stage l (likewise for y , w and h ). As far as we know, the previous method [7] proposed to utilize fixe anchors in RPN-based tracker. We employ the adjustable anchors in SCTRPN, which can change according to the regressor in the previous stage constantly, and computed as
x a l = x a l + w a l 1 t i ( x ) l 1         y a l = y a l + h a l 1 t i ( y ) l 1 w a l = w a l 1 exp ( t i ( w ) l 1 )       h a l = h a l 1 exp ( t i ( h ) l 1 ) ,
For the anchor in the first stage, x a 1 , y a 1 , w a 1 , and h a 1 are predefined.
The proposed cascaded structure of the RPN module is formed above. We perform the easy negative anchors penalization, to ensure the balanced distribution of training samples gradually. The cascaded structure makes full use of the multi-level features, which make the classifier more discriminative in distinguishing intricate distractors in complex scenarios. Figure 4 also shows the discriminative powers of different RPNs by demonstrating the detection response map at each stage. The red spot represents the localization that achieve the highest response score.
The loss function S C T R P N of SCTRPN consists of the loss functions of all R P N l . We compute the isolated loss function by Equation (4), and S C T R P N is expresses as
S C T R P N = l = 1 L R P N l ,

3.2. Feature Extraction Model(FEM) though Spatial Transformer Network (STN)

Inspired by the affine robustness of the spatial transformer network (STN) proposed in [9], the feature extraction model is built upon a combination with STN. The STN is utilized to calculate the affine transformed parameters in order to make the classification procedure better. In [9], the STN helps to change the posture of object to meet the requirement of accurate classification. In our research, the STN acts as the feature extraction model to transfer the features of different stages to obtain multi-features. In addition, the original function can also make the features extracted more robust to target deformation.
The Overview of STN. The STN [32] consists of three essential parts, including localization network, grid generator and sampler. Given the feature map, we perform the localization network to estimate the translation, rotation, and scale of certain object. The variables for deformation will transfer to the grid generator for updated feature map grid generation, and the sampler can utilize the updated feature map to gain the deformed feature mapping. The deformed feature map can be transferred to next layer, in order to improve the affine robustness of trained networks. To be noticed, the STN is independent and can be inserted to any existing networks. The structure of STN in the existing networks can also be series and parallel. The whole process is differentiable when inserting the STN into the main networks. We can optimize the localization network, in order to gain the minimum classification objective though back-propagation directly.
Feature Extraction Model. We obtain multi-level features from the multi-stages RPN structure, in order to effectively leverage these features, we propose to elect FEM to combine features across layers. During the combination, the high-level semantic features can be fused to improve the discriminability. In detail, the STN layer is used to transfer the feature map and match the feature dimensions at the same time. Different level features are fused by element-wise summation, followed a ReLU layer. We apply the grid generation model of STN to rescale the fused features, so that the FEM can ensure the same groundtruth for anchors in each RPN. In the meantime, we obtain the same resolution for all RPN in the output classification maps and regression maps. Figure 5 shows the feature extraction model for next layer.
In our experiments, we find it very important to limit the rotation degrees produced by FEM. Otherwise it is very easy to rotate the object upside down, which is the hardest to recognize in most cases. We constrain the rotation degree within 10 degrees clockwise and anti-clockwise.

3.3. Learning with Shrinkage Loss

Considering of the classification loss of the SCTRPN, we realize that the surrounding background contains much contextual information in. We can strengthen the discriminative power of classifier by utilizing the related wide background. In the meantime, it also brings large number of easy samples from the background, which cause the data imbalance issue. The easy samples may lead to undesirable results that generate the large loss. The learning process may pay much attention to the invaluable samples, which are far from the tracking target.
The research work in [33] found that the modulating factor can be applied to the loss, which can alleviate the data imbalance issue effectively. We treat the modulating factor as the function of the output possibility, and its function is to constraint the loss from easy samples.
Inspired by the shrinkage estimator [34] and the cost-sensitive weighting strategy [29], we propose the modulating factor, which is represented by l to re-weight the loss. We penalize the easy samples to achieve the hard samples by the optimization process. We compute the modulating function as a Sigmoid-like function by,
f ( l ) = 1 1 + exp ( a ( c l ) ) ,
where a and c are hyper-parameters. The parameters are obtain to control the shrinkage speed and the localization, respectively. We apply the modulating factor to weight the log loss, the proposed shrinkage loss can be expressed as,
L S = l 1 + exp ( a ( c l ) ) ,
The proposed shrinkage loss only penalizes the importance of easy samples (when l < 0.5 ) and keeps the loss of hard samples unchanged (when l > 0.5 ). Instead, we replace the classification loss by the shrinkage loss L c l s s h r ( p i , p i * ) = log ( p i ) p i * , which is employed in Equation (4),
L c l s s h r ( p i , p i * ) = log ( p i ) p i * 1 + exp ( a c ) [ ( p i ) p i * ] a ,
Considering of the implementation details, we set the value of a to be 10, in order to shrink the weight function quickly. We also set the value of c to be 0.2, so that the localization is suitable for the distribution of l . We constraint the value of c ranging from 0 to 1. Extensive comparison with the other losses shows that the proposed shrinkage loss can improve the tracking accuracy and the training speed at the same time.

3.4. Proposals Ranking Strategy

As we know, the high recall proposals can be generated by the spatial cascaded region proposal networks, while it may bring any redundancy to the tracking frameworks. For this reason, we present a multi-cue proposals re-ranking method, to obtain fewer and better proposals with high recall, which can provide the top tracking candidates for evaluation. The re-ranking method is based on multiple cues between proposal candidates and groundtruth, which include shape, color, and scale, which is shown in Figure 2.
Shape. The contours which enclosed by the bounding box ρ can be calculated by the existing method [35]. Furthermore, we can treat the number of contours enclosed by bounding box as shape score. Thus, the shape cue value s i , t between the i -th proposal candidate and target region τ t is illustrated by,
s i , t = e ( | ρ i ρ t | ) ,
where ρ i and ρ t represent the shape score of i -th proposal and the target, respectively.
Color. The color c i , t between the i -th proposal candidate and the target can be computed by the response map. The mean of all the values of the pixels from the region of the response map corresponded to the i -th proposal candidate, is computed as the color value between the i -th proposal candidate and target.
Size. We propose to utilize the size information to filter out the mismatching proposal candidates, which are undersized or oversized. The size value z i , t between the i -th proposal candidate and target is defined as,
z i , t = e ( | ω i ω t | ) e ( | h i h t | ) ,
where ω i / h i and ω t / h t denote the width/height of the i -th object proposal candidate and the target, respectively.
The three cues introduced above are independent, thus the multi-cues a i , t between the i -th proposal candidate and target can be performed as the product of the three core cues by,
a i , t = s i , t c i , t z i , t ,
During the proposals re-ranking process, we propose to rank the proposal candidates though the multi-cues in descending order. Then, we could obtain the high-quality proposals successfully by wiping out the proposal candidates with the low values. The high-quality proposals can maintain high recall, which helps the tracker to achieve better performance.

4. Experimental Results and Analysis

4.1. Training Dataset and Evaluation

Training. We train the SCTRPN by random interval sampling the images from the same sequences. Usually, we generate at most 64 samples from one image. We also perform the end-to-end network training, due to the combined spatial cascaded loss function. When it comes to the ratio of the anchor, the scale of target change smoothly between two consecutive frames in most situations, except for the fast motion sequences. Thus, we assign the ratios of anchors to [ 0.33 , 0.5 , 1 , 2 , 3 ] which is same as [7]. As for the positives and negatives, the positives are represented by the anchors, whose intersection-over-union (IOU) with groundtruth is over the threshold θ p o s . On the contrary, the negatives are defined by the anchors, whose IOU with groundtruth is less than the threshold θ n e g . The settings of shrinkage loss are described in Section 3.3.
Tracking. We evaluate the short-term object tracking on OTB2015 [36], VOT2018 [37], and UAV123 [1], respectively. LaSOT [38] and TrackingNet [39] are two recent largest datasets for single object tracking, and we validate the proposed method on these two datasets, to test its generalization performance.
The tracking procedure is more like the multi-stage detection: we first extract the features from the image in the first frame, using the pre-trained networks. For each stage, we utilize the FEM to combine the features and calculate the classification score and regression offset. Then, we perform coarse refining the anchor though RPN. The remaining anchors are regarded as proposal candidates, from which we utilize the multi-cues proposal re-ranking strategy to filter out the candidates in descending order, to obtain high-quality proposals. The final tracking results are determined by non-maximum-suppression (NMS), which is performed afterwards to get the final tracking bounding box. After the final bounding box is selected, the target size is updated by linear interpolation, to keep the shape changing smoothly. The whole tracking process of SCTRPN is summarized in Algorithm 1.
Algorithm 1 Proposed Tracking Method.
Input: Given sequences { X t } t = 1 T ; Groundtruth boundingbox of first frame X 1 named b 1 ; The trained model SCTRPN;
Output: Tracking results { b t } t = 2 T ;
 Initialize anchors A 1 ;
  For t = 2 to T do
    Extract features { φ l ( x ) } l = 1 L for x from SCTRPN;
    For l = 1 to L do
    If l equals to 1 then
     Φ ( x ) = φ l ( x ) ;
    Else
     Φ l ( x ) = F E M ( Φ l 1 ( x ) , φ l ( x ) ) ;
    End
    Calculate the classification score and regression offset using Equation (3);
    Coarse refining the anchor i from A l using Equation (6);
    Fine re-ranking the proposal candidates using multi-cues re-ranking strategy in Equation (13);
  End
    Select the best proposal as tracking result b s by the selection strategies in [22];
  End

4.2. Implementation Details

Network Architecture. In experiments, the backbone networks adopts the AlexNet [15] by reserving Conv layers to extract the features of images. The networks framework is described in detail in Section 3.1. The networks we combined include AlexNet, spatial transformer networks, and region proposal networks.
Optimization. We implement the whole training and tracking process using MatConvNet Deep Learning Frameworks [40] on a PC with an Intel i7, 16GB RAM and single Nvidia GTX1080Ti with 11GB video memory. The pre-trained parameters are directly come from the existing model on ImageNet [15]. SCTRPN is end-to-end trained with stochastic gradient descent (SGD) by 40 epoches. We employ a warmup learning rate of 0.001 for first 5 epoches to train the RPN braches. For the last 15 epoches, the whole network is end-to-end trained with the learning rate exponentially decayed from 0.001 to 0.00001. We set the stage number L to 3. We also assign the IOU of the positives and negatives to 0.7 and 0.3, respectively. The training loss is illustrated in Equation (7).

4.3. Relablity Ablation Study

Multi-features fusion through FEM. In order to test the validation of the multi-stage feature combination though FEM, we first to do the experiments on VOT2018 dataset, and the results are shown in Table 1. We firstly test the one-stage tracker and obtain the competitive performance with 0.321 in EAO. When we add another stage to the baseline, the EAO has increased to 0.352, and the accuracy and robustness are both increased by 8% and 17%, respectively. After combining all three stages, both accuracy and robustness steadily improve, with gains between 2.1% and 4.9% for VOT2018, compared with two-stage results. When it comes to EAO, the three stages are 12.4% higher than that of a single stage. We also provide the experimental results on different stages without STN, which can be seen in the last three lines of Table 1. The Tracker with STN outperforms those without STN in three main metrics except the tracking speed due to the accurate localization and affine transformed evaluation of the proposed FEM based on STN. The spatial transformer networks in FEM makes the effective progress when the image sequence suffering from deformation changes and affine transformation which refers to the attributes-based comparison in Section 4.4. Considering of the tracking speed, the three-stage tracker can also meet the demand of real-time tracking, though it costs more computational time during the tracking process.
Shrinkage loss. When replacing the proposed shrinkage loss with the original log loss, we evaluate the experimental analysis on the VOT2018 dataset. Compared with the RPN baseline, the proposed loss gains the large margin of 0.361, 5.1% higher than log loss. We also compare our proposed method with online hard negative mining [41], which aims to evaluate the capability of alleviating the data imbalance issue. Both methods penalize the importance of easy samples, due to the attribute of cost-sensitivity. We experimentally set the threshold of mining to 0.01. Our proposed method outperforms the online hard negative mining method. We can infer from the comparative results that easy samples still contribute to the learning process, but they should not dominate the whole gradient. Online hard negative mining was proposed to manually set the threshold, which cannot be appropriate for all the testing sequences.
Multi-cue re-rank. As shown in Figure 6, we evaluate the qualitative proposal generation results obtained by the proposal re-ranking method on some videos of the OTB-2015 dataset. Only the top ten generated proposals are shown in the test images. Under most circumstances, the top score generated proposals can cover the tracking groundtruth. We compare the proposed proposal re-ranking method with several state-of-the-art proposal generation methods in terms of recall. Four methods are chosen to achieve the evaluation, including CADM [42], MSTE [43], EdgeBoxes [44], and SelectiveSearch [45]. Selective Search and CADM are based on image segmentation; MSTE and EdgeBoxes are based on the boundary or the edge feature. The multi-cue re-ranking strategy has the capability of integrating the proposals around the target, and distinguishing the foreground target from the background clutters, which can obtain highly accurate proposal generation. As shown in Table 2, the recall obtained by multi-cues proposal re-ranking method is the highest among those obtained by the competing methods, higher by 34–70% when the top 50 ranked object proposals are used for calculating the recall. This is because the proposed method can combine color, shape, and scale information to accomplish the whole proposal generation process, which performs more robust to motion blur, illumination, deformation, and some other complex scenarios. In this section, we briefly test the proposal generation capability of our proposed networks. The OTB-2015 dataset, as a traditional benchmark, can clearly obtain the visualization of the results. Due to the eleven multiple tracking attributes, which almost contain all the tracking problems, the researchers still employ the OTB-2015 benchmark to accomplish a qualitative and quantitative comparison in recent paper work. We need to solve the out-of-view problems in the benchmark of OTB-2015 during proposals generation.

4.4. Comparison with State-of-the-Art Methods

OTB-2015 Dataset. The OTB-2015 dataset contains 100 sequences that mainly evaluate the accuracy and robustness of the compared tracker via the location error ratio and overlap ratio. The RPN-based tracker formulate the tracking as one-shot detection without online updating during the whole procedure. As the proposed tracker employs the proposals to do the final tracking, it can overcome the fast motion attributes most times. The improved the classification loss and multi-cues proposals re-ranking also help the proposed tracker to be top-performing method. We compare our SCTRPN tracker on the OTB2015 with the state-of-the-art trackers. Figure 7 shows that our SCTRPN tracker produces the leading result in overlap success and ranks second place in precision plots. Compared with other RPN-based trackers, our SCTRPN outperforms DaSiamRPN by improving 3.2% in precision and 5.1% in success. The VITAL tracker, which is performed via adversarial learning representations, obtained the best results in precision, and the network has the capability of consistently amending the edge information though adversarial learning. To compare with the correlation filters method ECO [18], we elect the features fusion strategy for both, while the multi-stage features extracted by feature extraction model are less sensitive to deformation variation and illumination, which results in a more comparable performance. We should also notice that the online updating methods fail to outperform other methods in the precision plot, due to the drift away caused by template updating. We also evaluate the proposed tracking method without the STN model, the scores of the precision and success drop dramatically, according to the details in the figure.
VOT-2018 Dataset. We validate our SCTRPN tracker on the VOT-2018 dataset in comparison with seven other state-of-the-art methods. The VOT-2018 public dataset includes 60 public sequences with different challenging attributes. It is treated as one of the most recent datasets for evaluating online model-free single object trackers. According to protocol, the expected average overlap (EAO), accuracy (A), and robustness (R) and no-reset-based average overlap (AO) are used to compare different trackers. The comparisons between state-of-the-art methods are reported in Table 3, and red, blue represent 1st and 2nd, respectively. Table 3 shows that the proposed SCTRPN tracker achieves the top-ranked performance on EAO and AO criteria. We also rank second place in the criteria of accuracy. As for robustness, the MFT and LADCF rank 1st and 2nd place, respectively. Due to the multi-stage fusion and proposal re-ranking strategy, our proposed SCTRPN method yields substantial gains of 10.9% on accuracy. The STN plays the key role in feature extraction model. Without the STN model, the other RPN-based methods outperform the SCTRPN-No STN, due to the lack of appearance change processing. The SiamRPN also obtains a high accuracy, mainly considering of the Siamese matching structure. Compared with the LADCF method, the proposed method achieves a performance gain of 1.5% on EAO criteria. Due to not matching the template, the robustness still has a gap with the state-of-art Siamese based trackers. We adopt one pass evaluation to test the different trackers. From the results of the AO criteria, we can observe that our proposed method outperforms the SiamRPN and DaSiamRPN method by 3.4% and 8.9% on AO, that is to say, our method achieve comparable performance compared with other RPN-based methods, and also has the capability of long-term tracking.
Accuracy vs. Speed. As shown in Figure 8, we perform the EAO performance on VOT2018 dataset to evaluate the tracking speed with respect to frames-per-second (FPS). The proposed network is not that complex, thus the device we chosen is single NVidia GTX1080Ti with 16GB Memory. We compare the proposed method with the results provided by the VOT2018 official data. We visualize that our SCTRPN achieves the best performance compared with other state-of-the-art methods running at real-time speed (22 FPS). When it comes to the MFT and LADCF, the tracking speed is low, due to the computational complexity and hardware limitation. From Figure 8, we can see that the SiamRPN is the most efficient approach among the compared methods. However, the proposed method achieved the higher EAO score than that of the SiamRPN, and the tracking speed can also meet the demand of real-time processing (22 FPS) in the computer vision application. Compared with the one-stage and two-stage SCTRPN, the fewer stages we employ, the faster the trackers perform. We set the stage number L to 3, considering the balance between effectiveness and efficiency.
LaSOT Dataset. We elect the experiments on LaSOT to further validate the proposed method on a larger and more challenging dataset. We have got 1400 videos in total and 280 videos with large scale and high quality in the LaSOT dataset. Figure 9 reports the overall performances of our SCTRPN tracker on LaSOT testing set. Our SCTRPN method outperforms other state-of-the-art methods by 0.559 and 0.487 on normalized precision and success. Compared with the RPN-based method DaSiamRPN, our SCTRPN tracker increases the normalized distance precision and success by 12.7% and 18.2%. Due to the shrinkage loss, we could penalize the importance of easy samples to alleviate the data imbalance issue, which makes the tracker more appropriate for long-term tracking.
TrackingNet Dataset. The TrackingNet dataset contains a large amount of data in the wild, which can evaluate the trackers’ performance in complex scenarios. There are 511 videos in the dataset. Among these videos, the tracking object is mainly suffering from illumination variation, heavy occlusion and background clutters. The TrackingNet dataset provides three criteria including success (A), precision (P) and normalized precision (Pnorm) for evaluation. Table 4 visualizes the comparison results, the SCTRPN achieves the best results on all three criteria from the second row. Specifically, the SCTRPN achieves 69.7%, 66.4%, and 76.4% on A, P, and Pnorm, respectively. The RPN-based methods outperform the methods of other tracking frameworks. Among the RPN-based methods, DaSiamRPN tracker obtains the second best results of 63.8% and 59.2% on A and P, and the SiamRPN achieves the second best results of 74.1% on Pnorm. The SCTRPN tracker gains the capability of tracking the object in the wild, due to multi-stage feature fusion and proposals re-ranking strategy, thus, the proposed method is more appropriate to handle the attributes of illumination change, background clutters, and also scale variation.
UAV123 Dataset. UAV123 dataset includes 123 sequences with average sequence length of 915 frames, recent researches mainly employ the length of the UAV123 dataset to validate the long-term tracking ability of trackers. We test state-of-the-art methods on UAV123 dataset, and infer from the results that the convolutional features are more robust than conventional handcrafted features, and the networks-based tracking frameworks outperform other compared frameworks, such as correlation filters during long-term tracking. Figure 10 illustrates the precision and success plots of the compared trackers. Specifically, our tracker achieves 0.797 and 0.613 on precision and success ranking the first place, which outperforms other two RPN-based methods DaSiamRPN (0.794, 0.584) and SiamRPN (0.748, 0.527). As for ECO and ECO-HC, the ECO only gains the precision and success by 3.4% and 2.5%, compared with ECO-HC. The convolutional features do little to influence the filters based trackers in long-term tracking, while the multi-stage features fusion strengthen the ability of robust tracking in complex scenarios.
Attribute-based Evaluation. The VOT2018 dataset is per-frame annotated with visual attributes to allow the detailed analysis of per-attribute tracking performance. Six attributes are chosen to accomplish the attribute-based evaluation for their importance to the compared trackers, including deformation variation, heavy occlusion, in-plane rotation, off-plane rotation, illumination change, and background clutter. Figure 11 shows the per-attribute plot for the top-performing trackers on VOT2018 in EAO. The proposed method was consistently ranked among the top trackers on the four attributes. The proposed SCTRPN performs the best in terms of deformation variation, occlusion, in-plane rotation, and illumination change. In summary, our proposed SCTRPN method can handle the extreme tracking situation under complex scenarios.

4.5. Hyper Parameters Selection

Different values of ξ . In Table 5, we briefly test the different values of ξ in Equation (4). By default, we use ξ = 1 which makes the two terms in Equation (4) roughly equally weighted after normalization. Table 5 shows that our result is impacted just marginally (by 2%) when ξ is within a scale of about two orders of magnitude (1 to 100). This demonstrates that the result is insensitive to ξ in a wide range.
The values of a and c in shrinkage loss. The modulating function is with the shape of a Sigmoid-like function, which is shown in Equation (8). Where a and c are hyper-parameters controlling the shrinkage speed and the localization, respectively. Figure 12a shows the shapes of the modulating function with different hyper-parameters. When applying the modulating factor to weight and the square loss, we have the proposed shrinkage loss as Equation (9). In this section, we mainly discuss the values of a and c hyper parameters in our proposed shrinkage loss. As shown in Figure 12b, the shrinkage loss only penalizes the importance of easy samples (when l < 0.5), and keeps the loss of hard samples unchanged (when l > 0.5) when compared to the square loss (L2). The focal loss (L3) penalizes both the easy and hard samples. We set the value of a to be 10 to shrink the weight function quickly and the value of c to be 0.2 to suit for the distribution of l , which ranges from 0 to 1. An extensive comparison with the other losses shows that the proposed shrinkage loss not only improves the tracking accuracy but also accelerates the training speed.

5. Conclusions

In this paper, we proposed a brand new framework for visual tracking which consists of spatial cascaded region proposals networks combining region proposal networks and spatial transformer networks to fully utilize multi-features. We also introduced the employment of the shrinkage loss to penalize the importance of easy samples to effectively alleviate the data imbalance issue. We provided the multi-cue proposals re-ranking method for the tracking frameworks, which can screen high quality proposals. Numerous experimental results demonstrated that the proposed tracker outperforms state-of-the-art trackers, highlighting the significant benefits of our method. However, we still cannot solve the problems of the “Tracking-by-understanding” mechanism. For future research, we will focus on the application of the combination between different networks accomplishing the action and pose detection during visual tracking. The enhancement of the deep learning network can not only improve the tracking performance, but also make the high-level computer vision applications based on tracking methods come true.

Author Contributions

Conceptualization, X.Z. and X.F.; methodology, X.Z. and X.F.; software, X.Z.; validation, X.Z. and X.F. and S.L.; formal analysis, X.Z.; investigation, X.Z. and X.F.; resources, S.L.; data curation, S.L.; writing—original draft preparation, X.Z.; writing—review and editing, X.Z. and S.L.; visualization, X.Z. and X.F.; supervision, X.F.; project administration, X.F.; funding acquisition, X.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by and Ministry of National Defense of China, grant number GFZX04014307 respectively.

Acknowledgments

Thanks to the experimental data provided by University of Ljubljana, SICK, Hiar, King Abdullah University of Science and Technology. Thanks to the experimental facilities provided by Xi’an Institute of Optics and Precision Mechanics of CAS.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Müller, M.; Smith, N. A benchmark and simulator for uav tracking. In Proceedings of the IEEE European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 445–461. [Google Scholar]
  2. Ning, G.; Huang, H. LightTrack: A generic framework for online top-down human pose tracking. arXiv 2019, arXiv:1905.02822. [Google Scholar]
  3. Wu, Y.; Lim, J.; Yang, M.H. Online object tracking benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 37, 1834–1848. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Zhang, X.; Wang, M. Robust visual tracking based on adaptive convolutional features and offline siamese tracker. Sensors 2018, 187, 2359. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Sun, Z.; Wu, J.; Wang, L. SRDT: A novel robust rgb-d tracker based on siamese region proposal network and depth information. Int. J. Pattern Recognit. Artif. Intell. 2019, 37, 437–452. [Google Scholar] [CrossRef]
  6. Gao, P.; Ma, Y.; Yuan, R. Learning cascaded siamese networks for high performance visual tracking. arXiv 2019, arXiv:1905.02857. [Google Scholar]
  7. Li, B.; Yan, J.; Wu, W.; Zhu, Z.; Hu, X. High performance visual tracking with siamese region proposal network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8971–8980. [Google Scholar]
  8. Zhu, Z.; Wang, Q.; Li, B. Distractor-aware siamese networks for visual object tracking. arXiv 2018, arXiv:1808.06048. [Google Scholar]
  9. Li, B.; Wu, W.; Wang, Q. SiamRPN++: Evolution of siamese visual tracking with very deep networks. arXiv 2018, arXiv:1812.11703. [Google Scholar]
  10. Zhang, H.; Ni, W.; Yan, W. Visual tracking using siamese convolutional neural network with region proposal and domain specific updating. Neurocomputing 2018, 275, 2645–2655. [Google Scholar] [CrossRef]
  11. Gidaris, S.; Komodakis, N. Object detection via a multi-region and semantic segmentation-aware CNN model. arXix 2015, arXiv:1505.01749. [Google Scholar]
  12. Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. arXiv 2017, arXiv:1712.00726. [Google Scholar]
  13. Smeulders, A.W.; Chu, D.M.; Cucchiara, R.; Calderara, S.; Dehghan, A.; Shah, M. Visual tracking: An experimental survey. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1442–1468. [Google Scholar] [PubMed] [Green Version]
  14. Li, P.; Wang, D.; Wang, L.; Lu, H. Deep visual tracking: Review and experimental comparison. Pattern Recognit. 2018, 76, 323–338. [Google Scholar] [CrossRef]
  15. Krizhevsky, A.; Sutskever, I.; Hinton, G. ImageNet classification with deep convolutional neural networks. In Proceedings of the International Conference on Neural Information Processing Systems, Doha, Qatar, 12–15 November 2012; pp. 1097–1105. [Google Scholar]
  16. Wang, L.; Ouyang, W.; Wang, X. Visual tracking with fully convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 11–18 December 2015; pp. 1093–1112. [Google Scholar]
  17. Danelljan, M.; Robinson, A.; Khan, F. Beyond correlation filters: Learning continuous convolution operators for visual tracking. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 472–488. [Google Scholar]
  18. Danelljan, M.; Bhat, G.; Khan, F. ECO: Efficient convolution operators for tracking. arXiv 2016, arXiv:1611.09224. [Google Scholar]
  19. Song, Y.; Ma, C.; Wu, X. VITAL: Visual tracking via adversarial learning. arXiv 2018, arXiv:1804.04273. [Google Scholar]
  20. Fiaz, M.; Mahmood, A.; Jung, S.K. Learning soft mask based feature fusion with channel and spatial attention for robust visual object tracking. Sensors 2020, 20, 4021. [Google Scholar] [CrossRef]
  21. Gordon, D.; Farhadi, A.; Fox, D. Re 3: Real-time recurrent regression networks for visual tracking of generic objects. IEEE Robot. Autom. Lett. 2018, 3, 788–795. [Google Scholar] [CrossRef] [Green Version]
  22. Guo, Q.; Feng, W.; Zhou, C.; Huang, R.; Wan, L.; Wang, S. Learning dynamic siamese network for visual object tracking. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1763–1771. [Google Scholar]
  23. Ren, S.; He, K.; Girshick, R. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 1137–1149. [Google Scholar] [CrossRef] [Green Version]
  24. Yang, K.; Song, H.; Zhang, K. Hierarchical attentive Siamese network for real-time visual tracking. Neural Comput. Appl. 2019, 2, 342–356. [Google Scholar] [CrossRef]
  25. Ma, C.; Huang, J.; Yang, X. Hierarchical convolutional features for visual tracking. IEEE Trans. Image Process. 2015, 25, 1834–1848. [Google Scholar]
  26. Danelljan, M.; Bhat, G.; Khan, F. Convolutional features for correlation filter based visual tracking. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 11–18 December 2015; pp. 763–772. [Google Scholar]
  27. Huang, C.; Li, Y.; Loy, C.; Tang, X. Learning deep representation for imbalanced classification. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 5375–5384. [Google Scholar]
  28. Khan, S.H.; Hayat, M.; Bennamoun, M.; Sohel, F.; Togneri, R. Cost-sensitive learning of deep feature representations from imbalanced data. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 3573–3587. [Google Scholar]
  29. Bertinetto, L.; Valmadre, J.; Henriques, J.F.; Vedaldi, A.; Torr, P.H. Fully-convolutional siamese networks for object tracking. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 850–865. [Google Scholar]
  30. Li, H.; Li, Y.; Porikli, F. Robust online visual tracking with a single convolutional neural network. In Proceedings of the IEEE Asian Conference on Computer Vision, Singapore, 1–5 November 2014; pp. 1392–1403. [Google Scholar]
  31. Jaderberg, M.; Simonyan, K.; Zisserman, A. Spatial transformer networks. arXiv 2015, arXiv:1506.02025. [Google Scholar]
  32. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 11–18 December 2015; pp. 1440–1448. [Google Scholar]
  33. Lin, T.; Goyal, P.; Girshick, R.; He, K.; Dollr, P. Focal loss for dense object detection. arXiv 2017, arXiv:1708.02002. [Google Scholar]
  34. Karamikabir, H.; Afshari, M.; Arashi, M. Shrinkage estimation of non-negative mean vector with unknown covariance under balance loss. J. Inequalitiesappl. 2018, 1, 124–135. [Google Scholar] [CrossRef] [PubMed]
  35. Guo, G.; Huang, H.; Yan, Y.; Liao, H.; Li, B. A new target-specific object proposal generation method for visual tracking. IEEE Trans. Cybern. 2017, 2, 132–149. [Google Scholar]
  36. Kristan, M.; Leonardis, A.; Matas, J. The sixth visual object tracking vot2018 challenge results. In Proceedings of the European Conference on Computer Vision Workshop, Munich, Germany, 8–14 August 2018; pp. 1453–1484. [Google Scholar]
  37. Wu, Y.; Lim, J.; Yang, M.H. Object tracking benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 4, 112–135. [Google Scholar] [CrossRef] [Green Version]
  38. Fan, H.; Lin, L.; Yang, F.; Chu, P.; Deng, G.; Yu, S.; Bai, H.; Xu, Y.; Liao, C.; Ling, H. Lasot: A high-quality benchmark for large-scale single object tracking. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2018; pp. 2012–2048. [Google Scholar]
  39. Müller, M.; Bibi, A.; Giancola, S.; Al-Subaihi, S.; Ghanem, B. Trackingnet: A large-scale dataset and benchmark for object tracking in the wild. arXiv 2018, arXiv:1803.10794. [Google Scholar]
  40. Vedaldi, A.; Lenc, K. Matconvnet: Convolutional neural networks for matlab. arXiv 2014, arXiv:1412.4564. [Google Scholar]
  41. Nam, H.; Han, B. Learning multi-domain convolutional neural networks for visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4293–4302. [Google Scholar]
  42. Xiao, Y.; Lu, C.; Tsougenis, E.; Lu, Y.; Tang, C. Complexity adaptive distance metric for object proposals generation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 778–786. [Google Scholar]
  43. Chen, X.; Ma, H.; Wang, X.; Zhao, Z. Improving object proposals with multi-thresholding straddling expansion. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 2587–2595. [Google Scholar]
  44. Zitnick, C.; Dollar, P. Edge boxes: Locating object proposals from edges. In Proceedings of the IEEE European Conference on Computer Vision, Zurich, Switzerland, 5–12 September 2014; pp. 391–405. [Google Scholar]
  45. Uijlings, J.; Sande, K.; Gevers, T.; Smeulders, A. Selective search for object recognition. Int. J. Comput. Vis. 2013, 104, 154–171. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Comparison with the target groundtruth, SiamRPN and Spatial Cascaded Transformed Region Proposal Network (SCTRPN) on two sequences: It performs that the comparison between two different region proposal networks (RPN)-based methods in UAV123 dataset. The deep convolutional features can only be extracted to obtain the semantic information, it may easily drift away when suffering from the similar distractors and heavy occlusions.
Figure 1. Comparison with the target groundtruth, SiamRPN and Spatial Cascaded Transformed Region Proposal Network (SCTRPN) on two sequences: It performs that the comparison between two different region proposal networks (RPN)-based methods in UAV123 dataset. The deep convolutional features can only be extracted to obtain the semantic information, it may easily drift away when suffering from the similar distractors and heavy occlusions.
Sensors 20 04810 g001
Figure 2. Flow diagram of the proposed tracking algorithm. The improvements which should be noticed are the feature extraction model, shrinkage classification loss, and proposal re-ranking strategy.
Figure 2. Flow diagram of the proposed tracking algorithm. The improvements which should be noticed are the feature extraction model, shrinkage classification loss, and proposal re-ranking strategy.
Sensors 20 04810 g002
Figure 3. A brief structure of one-stage region proposal network in our networks. The main structure of one-stage region proposal network contains two branches, including classification and regression, which help the network seizing the proposals for detection or tracking.
Figure 3. A brief structure of one-stage region proposal network in our networks. The main structure of one-stage region proposal network contains two branches, including classification and regression, which help the network seizing the proposals for detection or tracking.
Sensors 20 04810 g003
Figure 4. The comparison between the original image, color scores and different-stage response maps. (a) Contains the original image and the region of interest, (b) illustrates the color scores, and (ce) clearly provide the multi-stage response maps.
Figure 4. The comparison between the original image, color scores and different-stage response maps. (a) Contains the original image and the region of interest, (b) illustrates the color scores, and (ce) clearly provide the multi-stage response maps.
Sensors 20 04810 g004
Figure 5. Flow chart of the feature extraction model though the spatial transformer network (STN). We employ the STN in order to transfer features to next stage and ensure dimension of feature simultaneously.
Figure 5. Flow chart of the feature extraction model though the spatial transformer network (STN). We employ the STN in order to transfer features to next stage and ensure dimension of feature simultaneously.
Sensors 20 04810 g005
Figure 6. Qualitative proposal generation results obtained by the proposal re-ranking method on some videos of the OTB-2015 dataset. Only top ten generated proposals are shown in the test images.
Figure 6. Qualitative proposal generation results obtained by the proposal re-ranking method on some videos of the OTB-2015 dataset. Only top ten generated proposals are shown in the test images.
Sensors 20 04810 g006
Figure 7. Precision and success plots show a comparison of our SCTRPN tracker with state-of-the-art trackers on the OTB2015 dataset. (a) Precision plots of OPE on OTB2015; (b) Success plots of OPE on OTB2015.
Figure 7. Precision and success plots show a comparison of our SCTRPN tracker with state-of-the-art trackers on the OTB2015 dataset. (a) Precision plots of OPE on OTB2015; (b) Success plots of OPE on OTB2015.
Sensors 20 04810 g007
Figure 8. A comparison of the quality and the speed of state-of-the-art tracking methods on VOT2018. The expected average overlap (EAO) with respect to the frames-per-second (FPS) are visualized in the Figure. Note that the FPS axis is in the log scale.
Figure 8. A comparison of the quality and the speed of state-of-the-art tracking methods on VOT2018. The expected average overlap (EAO) with respect to the frames-per-second (FPS) are visualized in the Figure. Note that the FPS axis is in the log scale.
Sensors 20 04810 g008
Figure 9. Normalized precision and success plots show a comparison of our SCTRPN tracker with state-of-the-art trackers on the LaSOT dataset. (a) Normalized Precision plots of OPE on LaSOT Testing Set; (b) success plots of OPE on the LaSOT Testing Set.
Figure 9. Normalized precision and success plots show a comparison of our SCTRPN tracker with state-of-the-art trackers on the LaSOT dataset. (a) Normalized Precision plots of OPE on LaSOT Testing Set; (b) success plots of OPE on the LaSOT Testing Set.
Sensors 20 04810 g009
Figure 10. Precision and success plots show a comparison of our SCTRPN tracker with state-of-the-art trackers on the UAV123 dataset. (a) Precision plots of OPE on UAV123; (b) Success plots of OPE on UAV123.
Figure 10. Precision and success plots show a comparison of our SCTRPN tracker with state-of-the-art trackers on the UAV123 dataset. (a) Precision plots of OPE on UAV123; (b) Success plots of OPE on UAV123.
Sensors 20 04810 g010
Figure 11. Attribute-based evaluation of the trackers. The experimental results show the validation of six attributes including deformation, occlusion, in-plane rotation, off-plane rotation, illumination change, and background clutter. The numbers under the name represent the range of the attributes respectively.
Figure 11. Attribute-based evaluation of the trackers. The experimental results show the validation of six attributes including deformation, occlusion, in-plane rotation, off-plane rotation, illumination change, and background clutter. The numbers under the name represent the range of the attributes respectively.
Sensors 20 04810 g011
Figure 12. (a) Modulating factors in (5) with different hyper-parameters. (b) Comparison between the square loss (L2), focal loss (L3), and the proposed shrinkage loss.
Figure 12. (a) Modulating factors in (5) with different hyper-parameters. (b) Comparison between the square loss (L2), focal loss (L3), and the proposed shrinkage loss.
Sensors 20 04810 g012
Table 1. The reliability ablation study on multi-features fusion considering the number of stages and STN. The best two results are highlighted in red and blue fonts, respectively.
Table 1. The reliability ablation study on multi-features fusion considering the number of stages and STN. The best two results are highlighted in red and blue fonts, respectively.
StageOne StageTwo StagesThree StagesOne Stage without STNTwo Stages without STNThree Stages without STN
Accuracy 0.5230.5650.5770.5080.5380.566
Robustness 1.231.020.971.341.191.04
EAO 0.3210.3520.3610.3140.3420.349
FPS 453022543625
Table 2. The recall obtained by the proposal re-ranking method and the other four competing methods when varying the number of object proposals on the OTB-2015 dataset.
Table 2. The recall obtained by the proposal re-ranking method and the other four competing methods when varying the number of object proposals on the OTB-2015 dataset.
MethodsNumber of Proposals
501002005001000
CADM03250.4360.5740.7060.735
MSTE0.2530.4240.5670.6320.653
EdgeBoxes0.6030.7430.8130.9240.929
SelectiveSearch0.2460.3920.5210.7320.841
SCTRPN0.9210.9320.9530.9830.991
Table 3. Comparison with the state-of-the-art in terms of accuracy, robustness (failure rate), expected average overlap (EAO) and no-reset-based average overlap (AO) on the VOT2018 dataset. The best two results are highlighted in red and blue fonts, respectively.
Table 3. Comparison with the state-of-the-art in terms of accuracy, robustness (failure rate), expected average overlap (EAO) and no-reset-based average overlap (AO) on the VOT2018 dataset. The best two results are highlighted in red and blue fonts, respectively.
TrackerSCTRPNSCTRPN-No STNMFTLADCFDRTSiamRPNDaSiamRPN
Accuracy 0.5830.5640.5250.5030.5190.5860.569
Robustness 0.2430.2690.1400.1590.2010.2760.323
EAO 0.3950.3810.3850.3890.3570.3820.327
AO 0.4780.4530.3930.4210.4260.4620.439
Table 4. Comparison on the TrackingNet in terms of success, precision, and normalized precision. The best two results are highlighted in red and blue fonts, respectively.
Table 4. Comparison on the TrackingNet in terms of success, precision, and normalized precision. The best two results are highlighted in red and blue fonts, respectively.
TrackerSCTRPNSCTRPN-No STNECOMDNetSiamFCSiamRPNDaSiamRPN
A(%) 69.762.755.460.657.262.463.8
P(%) 66.459.449.356.853.658.459.2
Pnorm(%) 76.373.962.171.266.674.173.2
Table 5. Tracking results of EAO on the VOT-2018 benchmark using different values of ξ in Equation (4). The best results are highlighted in red fonts.
Table 5. Tracking results of EAO on the VOT-2018 benchmark using different values of ξ in Equation (4). The best results are highlighted in red fonts.
ξ ξ = 0.1 ξ = 1 ξ = 10 ξ = 100
EAO 0.3890.3950.3920.387

Share and Cite

MDPI and ACS Style

Zhang, X.; Luo, S.; Fan, X. Proposal-Based Visual Tracking Using Spatial Cascaded Transformed Region Proposal Network. Sensors 2020, 20, 4810. https://doi.org/10.3390/s20174810

AMA Style

Zhang X, Luo S, Fan X. Proposal-Based Visual Tracking Using Spatial Cascaded Transformed Region Proposal Network. Sensors. 2020; 20(17):4810. https://doi.org/10.3390/s20174810

Chicago/Turabian Style

Zhang, Ximing, Shujuan Luo, and Xuewu Fan. 2020. "Proposal-Based Visual Tracking Using Spatial Cascaded Transformed Region Proposal Network" Sensors 20, no. 17: 4810. https://doi.org/10.3390/s20174810

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop