Next Article in Journal
Synthesis of LaXO3 (X = Fe, Mn, Cr, Ni) Thin Films Using a Simple Spin Coating Set-Up for Resistive Switching Memory Devices
Previous Article in Journal
Classification Algorithm of 3D Pattern Film Using the Optimal Widths of a Histogram
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Siamese Trackers Based on Deep Features for Visual Tracking

1
Department of Computer Engineering, Sunchon National University, Suncheon 57992, Republic of Korea
2
Department of Data Science, National Korea Maritime and Ocean University, Busan 49112, Republic of Korea
3
Interdisciplinary Major of Ocean Renewable Energy Engineering, National Korea Maritime and Ocean University, Busan 49112, Republic of Korea
*
Authors to whom correspondence should be addressed.
Electronics 2023, 12(19), 4140; https://doi.org/10.3390/electronics12194140
Submission received: 20 August 2023 / Revised: 17 September 2023 / Accepted: 25 September 2023 / Published: 4 October 2023

Abstract

:
Visual object tracking poses challenges due to deformation of target object appearance, fast motion, brightness change, blocking due to obstacles, etc. In this paper, a Siamese network that is configured using a convolutional neural network is proposed to improve tracking accuracy and robustness. Object tracking accuracy is dependent on features that can well represent objects. Thus, we designed a convolutional neural network structure that can preserve feature information that is produced in the previous layer to extract spatial and semantic information. Features are extracted from the target object and search area using a Siamese network, and the extracted feature map is input into the region proposal network, where fast Fourier-transform convolution is applied. The feature map produces a probability score for the presence of an object region and an object in a region, where the similarities are high to search the target. The network was trained with a video dataset called ImageNet Large Scale Visual Recognition Challenge. In the experiment, quantitative and qualitative evaluations were conducted using the object-tracking benchmark dataset. The evaluation results indicated competitive results for some video attributes through various experiments. By conducting experiments, the proposed method achieved competitive results for some video attributes, with a success metric of 0.632 and a precision metric of 0.856 as quantitative values.

1. Introduction

Visual object tracking (VOT) is one of the popular research subjects in the computer vision field due to the advantages that can be applied to visual-based application programs, such as factory automation monitoring, autonomous driving, intruder monitoring, and drone work [1,2,3,4]. VOT is regarded as the most challenging and fundamental field because it has to steadily search and track a specific target in video frames. The general VOT process serves to track a target object using a bounding box, which is given in the first frame [5,6,7]. However, information that is obtained from the first frame is not sufficient to track a target object that is present in all frames. With a lack of feature information, object tracking is likely to fail [8,9]. Thus, high-level feature extraction is needed to represent objects well.
Although studies on object tracking have been conducted over the past decade, object tracking is still difficult due to many shortcomings in videos that capture the real world, such as shape conversion, illumination variation, and occlusion. The success of object tracking is dependent on how the information representing objects can robustly represent objects against various problems. Because of this, various approaches have been proposed to solve these problems in object tracking. The existing appearance-model-based tracking method employs a creation or identification model to separate the foreground and background [10].
This method depends on features created using hand-crafted methods. There are related drawbacks, such as not being able to exhibit the key information of the target object or not responding to a change in appearance robustly. To solve these problems, robust features that can represent the attributes of the target object should be extracted, and an appearance model needs to be created. The appearance model created through this process searches for a target in the image frame region and removes the external noise elements.
The two categories in this study are a generative method that focuses on appearance model creation and a discriminative method. In the generative method, the appearance of the target object is configured through the statistical model using object region information estimated from the previous frame. To maintain appearance information, studies on sparse representation and cellular automata have been conducted [11,12]. In contrast, the discriminative method aims to train a classifier that distinguishes objects and surrounding backgrounds. Studies on support vector machines and multiple instance learning have been conducted for classification [13,14]. However, since such methods employ hand-crafted features such as color histograms, poor information is only extracted, which cannot effectively respond to various changes in environments contained in videos.
Deep learning has shown outstanding results in the field of computer vision by introducing powerful algorithms that can automatically extract and learn complex patterns and features from visual data. By applying deep learning, various advantages can be obtained, such as improved accuracy and robustness through the learning and extraction of hierarchical representations of visual features from large datasets. Additionally, deep-learning models learn end-to-end mappings from raw inputs to outputs, greatly reducing complexity and making the models easy to implement. Deep learning also has scalability, making it suitable for applications beyond computer vision, such as the medical field [15,16]. This scalability allows for the efficient processing of large amounts of data, leveraging the parallel processing capabilities of deep-learning models. As a result, models can be developed to handle increasingly complex tasks and data in various domains.
Recent study methods have focused on deep features based on deep learning, shifting from existing hand-crafted methods. Deep features can be mainly obtained using a convolutional neural network (CNN). Features based on CNNs exhibit good performances in a wide range of visual recognition tasks. Since high-level information can be extracted through the multilayers in a CNN, it is gaining ground as a key method that can overcome the limitations of the tracking algorithm applied via the hand-crafted method. A CNN is trained using a large amount of image data and numerous object class types. Features extracted with a CNN show good performances in representing high-level information and distinguishing objects in various categories. Thus, it is important to use deep features extracted from a CNN for VOT applications.
In this paper, unique features of the target object and search region are extracted using a CNN and are used in the object-tracking algorithm by comparing similarities. The tracking problem is regarded as a method to search specific objects inside an image through a similarity comparison rather than considering this as a problem to classify target objects. Image similarity involves a task to compare features of the target object and features of objects that are present in the image plane. Existing CNNs have focused on generalization performance to classify a large number of classes in many types. Because of this, it has caused a low performance for object location identification. Thus, a customized CNN, in which all layers are convolutional, is produced by removing a fully connected layer to preserve object location information. For the similarity comparison, a customized CNN is configured as a Siamese network consisting of a Y-shaped branch network. Since this network is composed of the same weight and shape, similar features are extracted if similar images are input. These features are then input into a fast Fourier-transform (FFT) layer, thus performing a similarity comparison. A region proposal network (RPN) is used to infer a region where the target object is present from the region with the highest similarity.
The main contributions of this paper are as follows. First, we propose a method to increase the robustness of the tracking algorithm by applying FFT to capture global frequency information of the feature map, which can reduce sensitivity to image distortions and noise. Second, we employ a method to leverage the hierarchical characteristics of a Fully Deep CNN, which can effectively utilize both spatially detailed and semantic information. Third, we suggest a region regression technique that examines feature maps generated from various layers and uses deep convolution features and an RPN.
The present paper is organized as follows. In Section 2, studies on VOT are summarized. Section 3 describes a fully convolutional Siamese network for object tracking, while Section 4 describes the experimental results of the proposed tracking algorithm and performance comparison. Lastly, Section 5 presents the conclusion of this study and future research.

2. Related Studies

2.1. Tracking Algorithm Based on Correlation Filter

A correlation filter that is applied to VOT creates an appearance model using features extracted through hand-crafted filters. This appearance model is trained while renewing appearance weights in the partial region of the object obtained in each video frame. Keen attention has been paid to this process due to the high computation efficiency as a result of using FFT. To guarantee the speed of the algorithm, the minimum output sum of the squared error methodology that learns a minimum output sum of a squared error filter on the luminance channel was studied [17]. Furthermore, extended studies to improve the tracking accuracy have been proposed using context learning and a kernelized correlation filter [18,19]. Generally, a correlation filter that exhibits strong signals generates a correlation peak in each interested patch of the frame and produces a low response in the background region. A spatially regularized discriminative correlation filter (SRDCF) tracker imposes constraints on the correlation filter coefficients according to locations using a spatial regularization component in training to induce boundary effects [20]. The MCCTH-Staple tracker combined various types of features and configured various experts through DCF, thus independently tracking the target object with each expert [21]. Channel and spatial reliability concepts were applied to discriminative correlation filter (DCF) tracking, and the spatial reliability map was used for the filter adjustment in the partial region of the target object [22]. Improved kernelized correlation filters employ multichannel features, and they are the most widely used filters because of their overall outstanding performance and high frame-per-second rate [23].

2.2. Tracking Algorithm Based on CNN

CNNs have been used to obtain technical features, being validated in various works as an emerging network with excellent capability in computer vision and pattern recognition, such as image and video classification and object recognition [24,25]. It is important to use features that can be obtained from each convolution layer in the CNN for VOT. Visual representation plays the role of the key element in object tracking. Existing tracking algorithms employ a large number of hand-crafted features to represent a partial space and target shape, such as a color histogram [26,27].
More recently, studies on the application of CNNs’ deep features in tracking algorithms have been conducted. A deep-learning tracker that used a multilayer auto-encoder network was proposed [28]. DeepTrack consists of two-layer CNN classifiers using binary samples, while model updates are conducted via fine adjustment online [29]. Studies on the use of neural networks that learn a target-specific saliency map for tracking have also been conducted [30,31]. A tracking algorithm that adopted hierarchical features that were individually output from the convolution layer in the network was proposed to guarantee the tracking accuracy and robustness of VOT [32].

3. Proposed Method

In this section, the proposed tracking algorithm is described, as shown in Figure 1. The target object and search region images are used for the network input. The main features are extracted from the target object and search region in the Fully Deep CNN. This network is a Siamese network consisting of a Y-shaped branch network. Each feature passes through an FFT layer included in the RPN, thus classifying objects and calculating the bounding box center coordinates.

3.1. Siamese Network with CNN for Feature Extraction

Studies using features obtained with a CNN have been positioned as the key important element in the computer vision area. It is important to use features obtained with a CNN for robust VOT. In a standard CNN, features are extracted using a convolutional layer, and results are produced in a fully connected layer. However, there is a limitation in the fully connected layer from the VOT viewpoint. It is a problem of the disappearance of spatial location information.
Figure 2 shows that spatial information is maintained when only a convolutional layer is used. It is effective to use a fully connected layer because generalization should be performed inside the same class in a simple-class classification problem, and variables such as location information should not change. However, the purpose of VOT is not to infer a specific class but, rather, to infer the location of the target object that is present in a video frame. Thus, it is not appropriate to use a fully connected layer where location information disappears. In this paper, a customized network in which the fully connected layer is removed is developed. By deeply stacking convolutional layers, spatial information is maintained, and semantic information is used, as shown in Figure 2. To configure a deep convolutional layer, a convolutional block consisting of 1 × 1 and 3 × 3 filters was applied. The input feature map was compressed using a 1 × 1 filter, and the feature map was expanded using a 3 × 3 filter. A high-level feature map could be obtained because more convolutional layers could be stacked, even if the same number of parameters was used, by applying a convolutional block.
Figure 3 shows the feature map produced in the visualized convolutional layer. Figure 3a shows the input image, and 3b–d show the output results using the same layer and filter. The figures demonstrate that the output feature map, after passing through the convolutional layer, represented the location information and feature owned by the object. In this paper, a custom-tailored CNN was configured with a Siamese network for feature extraction.
Siamese networks are a type of deep-learning architecture specialized in tasks related to comparing the similarity between two pairs of input data. In particular, it serves as a core method in data-comparison-based applications, such as face recognition and signature verification systems. The fundamental concept of Siamese networks involves training the network on pairs of datapoints to learn a similarity metric between two inputs. A Siamese network encompasses four main features. Firstly, two pairs of data are input to the network in the form of pairs. If the input data type is an image, the reference image and the image to be compared are configured as a pair and fed into the network. Secondly, all parameters of the network are shared with each other. A Siamese network consists of a Y-shaped branch network with two identical structures, as shown in Figure 4. Since a Siamese network employs the same CNN, it is characterized by parameter and weight sharing. Although data pairs are input individually, the same parameters are used throughout the process. Thirdly, data features are extracted through the same network. The structure used for feature extraction may vary depending on the architectural layer that constitutes the network. If similar images are input, a similar feature map is produced. Images pass through the network to extract more detailed features. Lastly, a distance function is utilized to measure the similarity between the features of the extracted data. It quantifies the degree of similarity or distance between features of data pairs extracted from the network. General neural networks train a method to predict multiple classes, whereas a Siamese network can train the comparison of similarity between two images. The proposed architecture of the Siamese network is shown in Figure 5.
Figure 5 shows the network structure used to solve the tracking problem. It was designed by stacking a convolutional block consisting of convolutional layers that included kernels measuring 1 × 1 and 3 × 3 in size to increase the number of kernels that extracted features. The final output feature map in the tracking object region measured 18 × 18 × 256 in size, and the final output feature map in the search region measured 34 × 34 × 256 in size.
Meanwhile, Figure 6 shows the visualized heat map measuring image similarity using the output feature map. The brighter the section, the higher the similarity. This result verifies that the target object region could be approximated. However, we needed to obtain high-level information, such as coordinates, using a feature map to infer an accurate region.

3.2. Region Proposal Network for Estimating Object Area and Coordinates

In Faster R-CNN, an RPN (region proposal network) [33] is introduced to predict bounding box coordinates around objects present in an image. The RPN takes as input a feature map generated via passing the image through a convolutional neural network (CNN). This feature map encodes structural features and spatial information related to target objects. For the object region, feature maps of the target object and search region, which are finally produced in a Siamese network, are used.
Figure 7 shows the anchor box structure that infers coordinates in an RPN. An anchor box is arranged in every cell of the feature map. The number of anchor boxes can be arbitrarily set. It can be advantageous to infer more accurate object regions if the number of anchor boxes whose sizes are different increases. Conversely, as the number of anchor boxes increases, so does the number of computations.
The primary function of the RPN is to predict anchor box coordinates through regression and determine whether an object is present within each anchor box. Each anchor box is defined by four values: center X, center Y, width, and height. The number of anchor boxes varies based on the chosen box scale and aspect ratio. For instance, with a scale of three and an aspect ratio of two, six anchor boxes are generated. Each anchor box is placed individually in each cell of the feature map. Each anchor box is associated with one of three labels. A positive number indicates that there is significant area overlap between the object and the anchor box. A negative number indicates little or no overlap with the object, and −1 is data that do not fall into either the positive or negative categories. Data with −1 labels are ignored during the training process of the RPN to avoid interference.
The RPN performs binary classification to determine whether each anchor box contains an object or not. This classification yields probabilities ranging from 0 to 1. Boxes with probabilities close to 0 are classified as background, while those approaching 1 are considered to contain a substantial portion of an object.
The object existence in the anchor box and the inference of boxes are conducted in the converted frequency domain using FFT. The advantages of applying FFT in convolution are as follows. Firstly, in terms of computational speed, FFT-based convolution exhibits higher computational efficiency compared to traditional spatial domain convolution. This efficiency is particularly pronounced when processing large input data or kernels. While the computational complexity of conventional convolution is O ( N 2 ) , FFT-based convolution reduces it to O ( N log N ) . Here, N denotes the size of the input data or feature map. Secondly, FFT-based convolution excels in handling large kernels. As the kernel size increases, the computational cost of conventional convolution escalates. In contrast, FFT-based convolution maintains a relatively consistent performance, making it advantageous for operations involving large kernels. Consequently, performance gains can be achieved by integrating both conventional and FFT-based convolutions. Lastly, from the perspective of convolution theory, convolution is equated in the spatial domain to multiplication in the frequency domain. This equivalence offers the advantage of easily modifying and applying algorithms. The convolution in the spatial domain can be simply represented by the Hadamard product in the frequency domain that is obtained through FFT. It is converted into a frequency domain by applying FFT to the feature map of the target object and search region. In the feature map, which is used to determine whether an object is present in the anchor box region, FFT is applied as represented in Equations (1) and (2). In the feature map, which is used to infer the center coordinates of the anchor box, FFT is applied as represented in Equations (3) and (4).
T c l s u , v = F F T ( t c l s x , y )
S c l s u , v = F F T ( s c l s x , y )
T a n c h u , v = F F T ( t a n c h x , y )
S a n c h u , v = F F T ( s a n c h x , y )
In these equations, c l s refers to the object classification, and a n c h refers to the anchor box. t and s refer to the feature maps of the target object and search region, respectively. x and y refer to the location of each cell in the feature map. u and v refer to the coordinates in the frequency domain. T and S refer to the feature maps, which are represented in the frequency domain.
Figure 8 shows the FFT convolution process. To multiply each component in two feature maps, the size should be the same. However, the feature map of the target object is smaller than the feature map of the search region. Thus, it is necessary to match the size of the feature map of the target object with the size of the feature map of the search region. The reference region of the feature map is located at the edge of the upper end on the left side. For the other regions, zero padding is applied. Zero padding refers to filling the space with a zero. By filling it with a zero, it does not influence the FFT calculation and increases the resolution of the frequency domain. The input feature map and kernel that are converted into the frequency domain are calculated using the Hadamard product, as represented in Equations (5) and (6). This process offers more advantages in calculation speed than standard convolution because the product is conducted between the elements, which is different from that of standard convolution.
O c l s u , v = T c l s ( u , v ) S c l s ( u , v )
O a n c h u , v = T a n c h ( u , v ) S a n c h ( u , v )
In Equations (5) and (6), refers to the Hadamard product, which acquires O c l s u , v for object classification and O a n c h u , v for anchor box inference as the output. Each output is restored to the spatial domain by applying I n v e r s e F F T to each output. The restored final output includes the x , y , w i d t h , h e i g h t of each anchor box and a probability for object classification.

4. Experiments

4.1. Experiment Environment

The hardware specifications used in the experiment are described in the following. For the CPU, Intel Core i7 8 Generation 8700K and, for the GPU, NVIDIA TITAN X pascal 12GB were used. The proposed algorithm was implemented using PyTorch version 1.8. In this study, the ILSVRC2015 VID dataset [34] was used for training the tracking network as training data. ILSVRC VID is a dataset constructed for the object detection field. It is an expanded version of the ILSVRC dataset used for image classification and object detection. Unlike classification datasets, ILSVRC VID consists of video sequences and frame data and is specialized and suitable for object-tracking tasks. To quantitatively evaluate the algorithm, the object-tracking benchmark (OTB) dataset [35] was used. The ILSVRC 2017 VID dataset was divided into training and validation sets that consisted of 3862 video snippets and 555 video snippets, respectively. For the network training, extracted images with one frame were used, with the number of frames that made up each video being different.
The object region could be acquired using the annotation that was assigned for each frame. The annotation was composed of the bounding box coordinates (xmin, ymin, xmax, ymax) and frame size. The OTB dataset, which was used to quantitatively evaluate the tracking algorithm, consisted of around 100 video datasets, including 11 different attributes such as illumination variation (IV), scale variation (SV), and occlusion (OCC). The detailed attributes are presented in Table 1. A video contained one or more attributes. The evaluation was conducted in this study using OTB-100, which consisted of 100 video datasets, and OTB-50, which consisted of 50 video datasets containing videos that were relatively difficult to track. The OTB dataset also included an annotation. The target object region was initialized using the annotation of the first frame in the video for the qualitative evaluation. The annotation was not used in the tracking process but was used as the ground truth when conducting the performance evaluation.

4.2. Network Training

The ILSVRC 2015 dataset was used to train the proposed Siamese network. A pair of two images that were arbitrarily extracted from the dataset was used for the network input. Each image was extracted from the same video, which was used as the target object and search region. The sequence order was ignored in the image extraction process because the training was conducted through the similarity comparison of the objects in the network. The images used in learning were passed through preprocessing and then employed in the network training along with the normalized anchor box coordinate labels and object classification labels.
Figure 9 shows the pair of preprocessed learning images. Figure 9a,b show the original and preprocessed images of the target object, while Figure 9c,d show the original image of the search region and the final image after completing preprocessing, respectively. They were reconfigured so that the center point in the object region is positioned in the center of the image. The target image and search region were converted into dimensions of 127 × 127 and 255 × 255 in size, respectively, for the network input. In the size conversion, the margin was cut while maintaining the image ratio to preserve the shape of the object. Preprocessed images were reprocessed because the coordinates in the region where the object was located changed according to the conversion ratio, thus producing the anchor box coordinate labels. The classification label was used to determine whether the target object existed inside the anchor box. If the object existed, a one (otherwise a zero or −1) was assigned. The object’s existence was determined by the intersection result between the created anchor box region and the object region specified in the annotation of the training image.
The intersection over union (IOU) was used to calculate the intersection ratio. If the IOU was more than 60%, it was determined that the object existed by assigning a one to the anchor box. If the IOU was less than 50%, it was determined that the object did not exist in the anchor box by assigning a zero to the anchor box. If the IOU was between 50% and 60%, it was determined that the object’s existence was unclear. In that case, a − 1 was assigned so as not to affect the weight training. The number of classification labels was created, which was the same as the number of anchor boxes.
The loss function used in the network training had two types, namely the SmoothL1Loss function used to estimate the anchor box coordinate and the cross-entropy function used to classify objects. Equation (7) presents the SmoothL1Loss equation. In this equation, β refers to the hyperparameter, which is generally defined as one.
S m o o t h L 1 = 0.5 x n y n 2 β ,     i f   x n y n < β x n y n 0.5 × β ,     o t h e r w i s e
In Equation (7), if the x n y n value is smaller than the β term, a square term is used. Otherwise, the following L1 term is used. Due to this characteristic, it is less sensitive to abnormal values, and gradient exploding can be prevented. Equation (8) presents the cross-entropy function.
l o s s c l s = l o g e x p ( x c l a s s ) j e x p ( x j ) = x c l a s s + l o g j e x p ( x [ j ] )
The final loss function is calculated by summing Equations (7) and (8), which are aggregately represented in Equation (9).
l o s s t o t a l = S m o o t h L 1 + l o s s c l s

4.3. Quantitative Evaluation Metrics

In this paper, the performance evaluation of the proposed tracking algorithm was conducted using the OTB-50 and OTB-100 benchmark datasets that contained different video attributes. OTB-50 and OTB-100 consisted of 50 and 100 types of videos, respectively. The number of frames in each video was different, and in the evaluation of the proposed algorithm, precision and success plots were used. The precision plot calculated a difference in the center coordinates from the ground truth (GT) coordinates manually obtained in the annotation of the frame and predicted object region(PR). The higher the index, the more robust the algorithm tracking without drifting. The success plot referred to the index that showed an intersection ratio of bounding boxes that surrounded the object region. The intersection ratio was calculated as shown in Figure 10.
The intersection ratio was calculated using the GT coordinates that were manually obtained in the annotation of the frame and predicted region coordinates. As shown in Figure 10, GT refers to a bounding box region consisting of GT coordinates, and PR refers to a bounding box region of the object produced via tracking with the user’s tracking algorithm. Equation (10) is used to calculate the intersection ratio.
I o U G T , P R = A r e a ( G T P R ) A r e a ( G T P R )
The denominator of Equation (10) is the union region of GT and PR in Figure 10b. The numerator is the intersection region of GT and ST in Figure 10a. To express the performance rank, the area under the curve was used.

4.4. Experiment Results

In this paper, two quantitative evaluation indices of precision and success plots were used to evaluate the performance of the tracking algorithm. The performance verification of the proposed algorithm was conducted using the BACF [8], MCCTH-Staple [19], and CSRDCF-LP [20] tracking algorithms. The colors in the produced bar graph consisted, in order, of red, yellow-green, sky blue, and purple from the first to fourth ranks.
Figure 11 shows the performance evaluation results using the OTB-50 benchmark dataset. The proposed algorithm achieved a 0.581 success score and a 0.811 precision score, which were the highest scores.
Table 2 presents individual results of 11 attributes. The highest value is expressed in bold font. The proposed algorithm exhibited better results in the OCC, OV, FM, MB, SV, DEF, and OPR attributes than those of the comparable algorithms. However, it showed a worse performance in the IPR, IV, and BC attributes, given by a 0.535 success score in the LR attribute, which was lower than that of MCCTH-Staple. However, it did achieve a 0.879 precision score, which was the best result. These results indicated that the error of the intersection ratio between the GT and the predicted box was relatively higher than those of other algorithms, whereas the GT and center point error were lower. It can be deduced that the tracking success rate of the proposed algorithm was high.
Figure 12 shows the performance evaluation results using the OTB-100 benchmark dataset. The proposed algorithm achieved a 0.632 success score and a 0.856 precision score, which were the highest scores. Table 3 presents the scores for each attribute. The proposed algorithm exhibited higher scores in the OCC, OV, FM, MB, SV, DEF, and OPR attributes, showing the robustness of the algorithm. It also showed a weakness in the BC attribute, which was also shown using OTB-50. However, it showed strength in both the success score for the IV attribute and the precision score for the LR attribute.
The highest success scores were found to be 0.578 for the FM attribute using OTB-50 and 0.638 for the IV attribute using OTB-100. The highest precision scores were found to be 0.879 and 0.880 for the same LR attribute in both datasets. Figure 13 shows the tracking results of the highest-scoring attributes using OTB-50 and OTB-100. In the figure, the green box indicates the GT region, and the red box indicates the object region extracted using the proposed algorithm. Figure 13a depicts the FM attribute results, in which the target object that appeared in frame Nos. 10 and 11 moved fast, with the moving distance being long between the continuous frames. In this process, this figure showed that a blur phenomenon occurred, but the object region was well-preserved and tracked.
Figure 13b shows the IV attribute results. Although illumination variation occurred over the entire target object and partial region due to shade, the proposed algorithm tracked a target object region similar to that of the GT region. Figure 13c shows the LR attribution results. This figure depicts that the object regions of frame 1530 and 1904 were smaller than the initialized object region in frame No. 0. Although errors occurred in the intersection region, the object was well-tracked without a drift phenomenon.
Meanwhile, Figure 14 shows the qualitative evaluation results of the proposed algorithm. Figure 14a verifies that both the proposed and compared algorithms were robust in sequences with partial deformation and mild occlusion. Figure 14b,c show that object deformation was more severe than that of Figure 14a. However, the proposed tracking algorithm (red box) responded more robustly compared to comparable algorithms, in which tracking failed or the intersection ratio error was larger. Figure 14d shows the blocking of the object with a specific obstacle. Only the proposed algorithm and the BACF algorithm tracked the target object continuously, but tracking failed with the other algorithms in the No. 78 frame when the target object passed through the obstacle.
Tracking failure results of the proposed algorithm are shown in Figure 15. Example video sequences of failure results are Biker and Birds. The target object area was initialized using the object area given in frame 0 of Biker and Birds. In the case of Biker, the helmet part of the object was set as the initial area, and in the case of Birds, the body part of the bird was selected.
In the Biker sequence, object tracking proceeded normally until frame 66, but the shape of the object rapidly changed in the process of moving to frame 80. In the com-parison algorithm, object drift results were shown, and tracking failed, but the proposed algorithm showed continuous tracking results. However, the area of the bounding box extended to the green upper body area, not the helmet area, resulting in lower overlap accuracy. During the process of inferring the object’s region, the candidate region’s range was expanded between frames 66 and 79, leading to an erroneous detection.
The Birds sequence showed correct inference in situations where objects appeared, but tracking failed because the object was obscured by clouds from frames 129 to 183. As a result, even if the object reappeared in frame 184, it was judged that the position of the candidate area designated an area other than the object, causing retracking to fail. This phenomenon could be addressed by resetting the starting position via retrieving the object’s position if it is missed in the future using an algorithm such as a sliding window.

5. Conclusions

In this paper, an algorithm for object tracking was proposed for two items, namely success and precision plots, using a Siamese network that was trained with the ILSVRC 2015 dataset. The CNN that made up the Siamese network was developed from convolutional layers to maintain spatial information. The feature maps of the target object and search region, which included spatial and semantic information, were commonly used in the RPN, where the Siamese network and FFT convolution were contained for tracking. In this study, tracking was conducted, as well as inferring the object region in regions with a high object similarity. The evaluation results indicated competitive results for some video attributes based on various experiments. However, the proposed algorithm exhibited a limitation in the in-plane rotation and background clutter attributes. To address this limitation, future studies should be conducted to enhance tracking performance by extracting only the necessary information required for object tracking within a small region. Specifically, we plan to investigate a novel feature map selection method using graph concepts to choose discriminative features while discarding noisy or irrelevant ones and to analyze the interrelationship of information surrounding objects. This future study aims to further improve tracking performance and overcome the challenges faced in the current study.

Author Contributions

Conceptualization, S.-C.L., J.-H.H. and J.-C.K.; data curation, S.-C.L. and J.-C.K.; formal analysis, S.-C.L. and J.-H.H.; investigation, S.-C.L., J.-H.H. and J.-C.K.; methodology, S.-C.L. and J.-C.K.; project administration, S.-C.L., J.-H.H. and J.-C.K.; resources, S.-C.L., J.-H.H. and J.-C.K.; software, S.-C.L., J.-H.H. and J.-C.K.; supervision, S.-C.L. and J.-C.K.; validation, S.-C.L. and J.-C.K.; visualization, S.-C.L., J.-H.H. and J.-C.K.; writing—original draft, S.-C.L., J.-H.H. and J.-C.K.; writing—review and editing, S.-C.L., J.-H.H. and J.-C.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The dataset used in the experiment is [34] (datasets publicly available at https://www.image-net.org/challenges/LSVRC/, accessed on 26 September 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Koubâa, A.; Qureshi, B. DroneTrack: Cloud-Based Real-Time Object Tracking Using Unmanned Aerial Vehicles over the Internet. IEEE Access 2018, 6, 13810–13824. [Google Scholar] [CrossRef]
  2. Jia, X.; Lu, H.; Yang, M.H. Visual Tracking via Adaptive Structural Local Sparse Appearance Model. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 1822–1829. [Google Scholar]
  3. Kim, S.H.; Lim, S.C.; Kim, D.Y. Intelligent Intrusion Detection System Featuring a Virtual Fence, Active Intruder Detection, Classification, Tracking. Ann. Nucl. Energy 2018, 112, 845–855. [Google Scholar] [CrossRef]
  4. Ma, B.; Huang, L.; Shen, J.; Shao, L. Discriminative Tracking Using Tensor Pooling. IEEE Trans. Cybern. 2016, 46, 2411–2422. [Google Scholar] [CrossRef] [PubMed]
  5. Wang, N.; Shi, J.; Yeung, D.Y.; Jia, J. Understanding and Diagnosing Visual Tracking Systems. In Proceedings of the IEEE International Conference on Computer Vision 2019, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3101–3109. [Google Scholar]
  6. Ma, C.; Huang, J.B.; Yang, X.; Yang, M.H. Hierarchical Convolutional Features for Visual Tracking. In Proceedings of the IEEE International Conference on Computer Vision, ICCV 2015, Santiago, Chile, 7–13 December 2015; pp. 3074–3082. [Google Scholar]
  7. Li, B.; Wu, W.; Wang, Q.; Zhang, F.; Xing, J.; Yan, J. Siamrpn++: Evolution of siamese visual tracking with very deep networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, 15–20 June 2019; pp. 4282–4291. [Google Scholar]
  8. Kiani Galoogahi, H.; Fagg, A.; Lucey, S. Learning Background-Aware Correlation Filters for Visual Tracking. In Proceedings of the IEEE International Conference on Computer Vision, ICCV 2017, Venice, Italy, 22–29 October 2017; pp. 1135–1143. [Google Scholar]
  9. Du, X.; Clancy, N.; Arya, S.; Hanna, G.B.; Kelly, J.; Elson, D.S.; Stoyanov, D. Robust Surface Tracking Combining Features, Intensity and Illumination Compensation. Int. J. Comput. Assist. Radiol. Surg. 2015, 10, 1915–1926. [Google Scholar] [CrossRef] [PubMed]
  10. Wang, L.; Ouyang, W.; Wang, X.; Lu, H. Visual tracking with fully convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, ICCV 2015, Santiago, Chile, 7–13 December 2015; pp. 3119–3127. [Google Scholar]
  11. Bao, C.; Wu, Y.; Ling, H.; Ji, H. Real-Time Robust L1 Tracker using Accelerated Proximal Gradient Approach. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2012, Providence, RI, USA, 16–21 June 2012; pp. 1830–1837. [Google Scholar]
  12. Xue, X.; Li, Y.; Shen, Q. Unmanned Aerial Vehicle Object Tracking by Correlation Filter with Adaptive Appearance Model. Sensors 2018, 18, 2751. [Google Scholar] [CrossRef] [PubMed]
  13. Hare, S.; Golodetz, S.; Saffari, A.; Vineet, V.; Cheng, M.M.; Hicks, S.L.; Torr, P.H. Struck: Structured Output Tracking with Kernels. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 2096–2109. [Google Scholar] [CrossRef] [PubMed]
  14. Babenko, B.; Yang, M.H.; Belongie, S. Robust Object Tracking with Online Multiple Instance Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 1619–1632. [Google Scholar] [CrossRef] [PubMed]
  15. Zhu, Y.; Tang, H. Automatic Damage Detection and Diagnosis for Hydraulic Structures Using Drones and Artificial Intelligence Techniques. Remote Sens. 2023, 15, 615. [Google Scholar] [CrossRef]
  16. Zhu, Y.; Xie, M.; Zhang, K.; Li, Z. A Dam Deformation Residual Correction Method for High Arch Dams Using Phase Space Reconstruction and an Optimized Long Short-Term Memory Network. Mathematics 2023, 11, 2010. [Google Scholar] [CrossRef]
  17. Bolme, D.S.; Beveridge, J.R.; Draper, B.A.; Lui, Y.M. Visual object tracking using adaptive correlation filters. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition 2010, San Francisco, CA, USA, 13–18 June 2010; pp. 2544–2550. [Google Scholar]
  18. Zhang, K.; Zhang, L.; Liu, Q.; Zhang, D.; Yang, M.H. Fast Visual Tracking via Dense Spatio-Temporal Context Learning. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 127–141. [Google Scholar]
  19. Henriques, J.F.; Rui, C.; Martins, P.; Batista, J. Exploiting the Circulant Structure of Tracking-By-Detection with Kernels. In Proceedings of the 12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; pp. 702–715. [Google Scholar]
  20. Danelljan, M.; Hager, G.; Shahbaz Khan, F.; Felsberg, M. Learning spatially regularized correlation filters for visual tracking. In Proceedings of the IEEE International Conference on Computer Vision 2015, Santiago, Chile, 7–13 December 2015; pp. 4310–4318. [Google Scholar]
  21. Wang, N.; Zhou, W.; Tian, Q.; Hong, R.; Wang, M.; Li, H. Multi-cue correlation filters for robust visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4844–4853. [Google Scholar]
  22. Lukezic, A.; Vojir, T.; Cehovin Zajc, L.; Matas, J.; Kristan, M. Discriminative correlation filter with channel and spatial reliability. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 6309–6318. [Google Scholar]
  23. Henriques, J.F.; Caseiro, R.; Martins, P.; Batista, J. High-speed tracking with kernelized correlation filters. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 583–596. [Google Scholar] [CrossRef] [PubMed]
  24. Kim, J.-C.; Lim, S.-C.; Choi, J.; Huh, J.-H. Review for Examining the Oxidation Process of the Moon Using Generative Adversarial Networks: Focusing on Landscape of Moon. Electronics 2022, 11, 1303. [Google Scholar] [CrossRef]
  25. Liu, T.; Tao, D.; Song, M.; Maybank, S.J. Algorithm dependent generalization bounds for multi-task learning. Proc. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 227–241. [Google Scholar] [CrossRef] [PubMed]
  26. Ross, D.A.; Lim, J.; Lin, R.S.; Yang, M.H. Incremental learning for robust visual tracking. Int. J. Comput. Vis. 2008, 77, 125–141. [Google Scholar] [CrossRef]
  27. Hao, Z.; Liu, G.; Gao, J.; Zhang, H. Robust Visual Tracking Using Structural Patch Response Map Fusion Based on Complementary Correlation Filter and Color Histogram. Sensors 2019, 19, 4178. [Google Scholar] [CrossRef] [PubMed]
  28. Wang, N.; Yeung, D. Learning a Deep Compact Image Representation for Visual Tracking. In Proceedings of the Advances in Neural Information Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems 2013, Lake Tahoe, NV, USA, 5–10 December 2013; pp. 809–817. [Google Scholar]
  29. Li, H.; Li, Y.; Porikli, F. Deeptrack: Learning discriminative feature representations online for robust visual tracking. IEEE Trans. Image Process. 2015, 25, 1834–1848. [Google Scholar] [CrossRef] [PubMed]
  30. Zou, W.; Zhu, S.; Yu, K.; Ng, A. Deep learning of invariant features via simulated fixations in video. Adv. Neural Inf. Process. Syst. 2012, 25, 3203–3211. [Google Scholar]
  31. Hong, S.; You, T.; Kwak, S.; Han, B. Online tracking by learning discriminative saliency map with convolutional neural network. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 597–606. [Google Scholar]
  32. Zhang, H.; Duan, R.; Zheng, A.; Zhang, J.; Li, L.; Wang, F. Discriminative Siamese Tracker Based on Multi-Channel-Aware and Adaptive Hierarchical Deep Features. Symmetry 2021, 13, 2329. [Google Scholar] [CrossRef]
  33. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. Adv. Neural Inf. Process. Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef] [PubMed]
  34. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
  35. Wu, Y.; Lim, J.; Yang, M.-H. Online object tracking: A benchmark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2013, Portland, OR, USA, 23–28 July 2013; pp. 2411–2418. [Google Scholar]
Figure 1. Proposed tracking algorithm with Siamese network: (a) input image, (b) Siamese network consisting of fully convolutional neural network for feature extraction, (c) RPN where FFT computation is applied to predict the object region and bounding box coordinates, and (d) tracking result.
Figure 1. Proposed tracking algorithm with Siamese network: (a) input image, (b) Siamese network consisting of fully convolutional neural network for feature extraction, (c) RPN where FFT computation is applied to predict the object region and bounding box coordinates, and (d) tracking result.
Electronics 12 04140 g001
Figure 2. Spatial and semantic information extracted from convolution layer.
Figure 2. Spatial and semantic information extracted from convolution layer.
Electronics 12 04140 g002
Figure 3. Feature map produced in convolutional layer; (a) Input image, (bd) Output feature map of first convolution block.
Figure 3. Feature map produced in convolutional layer; (a) Input image, (bd) Output feature map of first convolution block.
Electronics 12 04140 g003
Figure 4. Standard architecture of a Siamese network.
Figure 4. Standard architecture of a Siamese network.
Electronics 12 04140 g004
Figure 5. Proposed architecture of the Siamese network.
Figure 5. Proposed architecture of the Siamese network.
Electronics 12 04140 g005
Figure 6. Visualized similarity map of feature maps produced with the Siamese network.
Figure 6. Visualized similarity map of feature maps produced with the Siamese network.
Electronics 12 04140 g006
Figure 7. Anchor boxes arranged in the feature map.
Figure 7. Anchor boxes arranged in the feature map.
Electronics 12 04140 g007
Figure 8. FFT convolution process using the feature map converted into the frequency domain.
Figure 8. FFT convolution process using the feature map converted into the frequency domain.
Electronics 12 04140 g008
Figure 9. Preprocessed training dataset: (a) original target image, (b) converted target image, (c) original search region, and (d) converted search image.
Figure 9. Preprocessed training dataset: (a) original target image, (b) converted target image, (c) original search region, and (d) converted search image.
Electronics 12 04140 g009
Figure 10. Calculation method of intersection ratio using bounding box: (a) intersection and (b) union.
Figure 10. Calculation method of intersection ratio using bounding box: (a) intersection and (b) union.
Electronics 12 04140 g010
Figure 11. Evaluation graph of OTB-50 for all attributes: (a) success plot graph of OTB-50 and (b) precision plot graph of OTB-50.
Figure 11. Evaluation graph of OTB-50 for all attributes: (a) success plot graph of OTB-50 and (b) precision plot graph of OTB-50.
Electronics 12 04140 g011
Figure 12. Evaluation graph of OTB-100 for all attributes: (a) success plot graph of OTB-100 and (b) precision plot graph of OTB-100.
Figure 12. Evaluation graph of OTB-100 for all attributes: (a) success plot graph of OTB-100 and (b) precision plot graph of OTB-100.
Electronics 12 04140 g012
Figure 13. Tracking results of highest-scoring attributes: (a) FM attribute in the success score using OTB-50; (b) IV attribute in the success score using OTB-100; (c) LR attribute in the precision score using OTB-50 and OTB-100.
Figure 13. Tracking results of highest-scoring attributes: (a) FM attribute in the success score using OTB-50; (b) IV attribute in the success score using OTB-100; (c) LR attribute in the precision score using OTB-50 and OTB-100.
Electronics 12 04140 g013
Figure 14. Qualitative evaluation of proposed tracking algorithm: (a) bolt; (b) board; (c) lemming; (d) jogging.
Figure 14. Qualitative evaluation of proposed tracking algorithm: (a) bolt; (b) board; (c) lemming; (d) jogging.
Electronics 12 04140 g014
Figure 15. Result images of failed tracking: (a) failed overlapping of bounding box due to rapid changes in target object; (b) failure case where the target was lost due to the object disappearing behind an obstacle.
Figure 15. Result images of failed tracking: (a) failed overlapping of bounding box due to rapid changes in target object; (b) failure case where the target was lost due to the object disappearing behind an obstacle.
Electronics 12 04140 g015
Table 1. Description of attributes contained in the videos of the OTB dataset.
Table 1. Description of attributes contained in the videos of the OTB dataset.
AttributeDescription
Background Clutter (BC)Color or texture created similar to the object
Deformation (DEF)Non-rigid deformation of the object
Fast Motion (FM)Fast motion of the object detected
In-Plane Rotation (IPR)Object rotation detected in the image
Illumination Variation (IV)Illumination variation in the target object region
Low Resolution (LR)The low resolution of the object
Motion Blur (MB)Motion blur occurred in the target object
Occlusion (OCC)Occlusion generated in the target object region
Out-of-Plane Rotation (OPR)Object rotation detected outside the image
Out-of-View (OV)Some regions of the object moved outside the image
Scale Variation (SV)Scale variation in the tracking object
Table 2. Quantitative evaluation results of proposed tracking algorithm using OTB-50 dataset.
Table 2. Quantitative evaluation results of proposed tracking algorithm using OTB-50 dataset.
MetricsProposedMCCTH-
Staple
BACFCSRDCF-LP
TotalSuccess0.5810.5490.5500.502
Precision0.8110.7130.7570.662
IPRSuccess0.5210.5140.5400.462
Precision0.7450.6830.7480.607
OCCSuccess0.5670.5520.5160.464
Precision0.8210.7150.7080.608
OVSuccess0.5600.4910.4830.435
Precision0.8040.6710.7040.624
IVSuccess0.5660.5490.5870.466
Precision0.7660.7240.7920.607
LRSuccess0.5350.5710.4370.486
Precision0.8790.8340.6950.711
BCSuccess0.5720.5170.5850.445
Precision0.7720.6790.7970.575
FMSuccess0.5780.5240.5340.536
Precision0.7700.6460.7490.693
MBSuccess0.5780.4920.5420.535
Precision0.7910.6250.7560.692
SVSuccess0.5770.5250.5060.470
Precision0.8010.6800.7100.622
DEFSuccess0.5300.5300.5140.493
Precision0.7530.6920.7100.688
OPRSuccess0.5430.5360.5180.432
Precision0.7800.6940.7190.562
Table 3. Quantitative evaluation results of proposed tracking algorithm using OTB-100 dataset.
Table 3. Quantitative evaluation results of proposed tracking algorithm using OTB-100 dataset.
MetricsProposedMCCTH-
Staple
BACFCSRDCF-LP
TotalSuccess0.6320.6160.6020.585
Precision0.8560.7990.8170.764
IPRSuccess0.5590.5650.5670.534
Precision0.7910.7540.7920.716
OCCSuccess0.6180.5950.5600.527
Precision0.8430.7560.7560.671
OVSuccess0.6040.5290.5290.508
Precision0.8300.7300.7400.685
IVSuccess0.6380.5920.6270.564
Precision0.8290.7550.8300.721
LRSuccess0.5140.5720.4460.495
Precision0.8800.8510.7290.743
BCSuccess0.6290.6100.6460.553
Precision0.8430.7880.8630.715
FMSuccess0.6190.5800.5720.593
Precision0.8140.7130.7820.759
MBSuccess0.6310.5700.5840.584
Precision0.8230.7050.7770.732
SVSuccess0.5990.5850.5380.537
Precision0.8170.7660.7640.710
DEFSuccess0.5910.5860.5550.544
Precision0.8180.7790.7770.747
OPRSuccess0.5980.5850.5660.524
Precision0.8310.7690.7880.694
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lim, S.-C.; Huh, J.-H.; Kim, J.-C. Siamese Trackers Based on Deep Features for Visual Tracking. Electronics 2023, 12, 4140. https://doi.org/10.3390/electronics12194140

AMA Style

Lim S-C, Huh J-H, Kim J-C. Siamese Trackers Based on Deep Features for Visual Tracking. Electronics. 2023; 12(19):4140. https://doi.org/10.3390/electronics12194140

Chicago/Turabian Style

Lim, Su-Chang, Jun-Ho Huh, and Jong-Chan Kim. 2023. "Siamese Trackers Based on Deep Features for Visual Tracking" Electronics 12, no. 19: 4140. https://doi.org/10.3390/electronics12194140

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop