Next Article in Journal
Universal Filter Based on Compact CMOS Structure of VDDDA
Next Article in Special Issue
DDoS Flood and Destination Service Changing Sensor
Previous Article in Journal
Proof of Concept for a Quick and Highly Sensitive On-Site Detection of SARS-CoV-2 by Plasmonic Optical Fibers and Molecularly Imprinted Polymers
Previous Article in Special Issue
Digitally Controlled Oscillator with High Timing Resolution and Low Complexity for Clock Generation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Siamese Networks-Based People Tracking Using Template Update for 360-Degree Videos Using EAC Format †

1
InnoFusion Technology, 3F-10, No. 5, Taiyuan 1st street, Zhubei, Hsinchu 30288, Taiwan
2
Communication Engineering Department, National Central University, Taoyuan 320317, Taiwan
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper Siamese Networks Based People Tracking for 360-degree Videos with Equi-angular Cubemap Format, In Proceedings of the 2020 IEEE International Conference on Consumer Electronics—Taiwan (ICCE-Taiwan), Taoyuan, Taiwan, 28–30 September 2020.
Sensors 2021, 21(5), 1682; https://doi.org/10.3390/s21051682
Submission received: 10 January 2021 / Revised: 16 February 2021 / Accepted: 23 February 2021 / Published: 1 March 2021

Abstract

:
Rich information is provided by 360-degree videos. However, non-uniform geometric deformation caused by sphere-to-plane projection significantly decreases tracking accuracy of existing trackers, and the huge amount of data makes it difficult to achieve real-time tracking. Thus, this paper proposes a Siamese networks-based people tracker using template update for 360-degree equi-angular cubemap (EAC) format videos. Face stitching overcomes the problem of content discontinuity of the EAC format and avoids raising new geometric deformation in stitched images. Fully convolutional Siamese networks enable tracking at high speed. Mostly important, to be robust against combination of non-uniform geometric deformation of the EAC format and partial occlusions caused by zero padding in stitched images, this paper proposes a novel Bayes classifier-based timing detector of template update by referring to the linear discriminant feature and statistics of a score map generated by Siamese networks. Experimental results show that the proposed scheme significantly improves tracking accuracy of the fully convolutional Siamese networks SiamFC on the EAC format with operation beyond the frame acquisition rate. Moreover, the proposed score map-based timing detector of template update outperforms state-of-the-art score map-based timing detectors.

1. Introduction

Visual tracking aims at locating and covering the target with an accurate bounding box. Among various kinds of videos, 360-degree videos feature panoramic information. Thus, they provide applications of computer vision rich environmental information and provide users an immersive viewing experience. Examples of applications of visual tracking on 360-degree videos are virtual reality (VR), augmented reality (AR), autonomous vehicles, unmanned autonomous vehicles (UAVs), health care, and medical image analysis. Although many visual trackers have been proposed, sphere-to-plane projection of 360-degree videos fails most in existing trackers since these trackers do not consider the characteristics of 360-degree videos.
The following subsections provide overviews of visual trackers on 360-degree videos (Section 1.1), Siamese networks-based visual trackers (Section 1.2), and template update for Siamese networks-based visual trackers (Section 1.3). Section 1.4. provides the motivation and contributions of the proposed scheme. All described trackers in Section 1 are summarized in Table 1.

1.1. Visual Tracking on 360-Degree Videos

360-degree cameras have started to be commercialized in recent years. Although 360-degree videos-based novel applications are expected to become widespread in the near future, many challenges of 360-degree signal processing remain to be overcome [1]. In the latest video coding standard Versatile Video Coding (VVC) that was finalized in July 2020, several novel sphere-to-plane projection formats (e.g., equiangular cubemap (EAC) [2]) of 360-degree videos were considered [3]. These formats enable higher coding gain and higher tracking accuracy since they have less non-uniform geometric distortions. Few trackers have been designed for 360-degree videos using these formats since these formats are novel. Liu et al. proposed a convolutional neural network (CNN)-based multiple pedestrian tracker for 360-degree equirectangular projection (ERP) format videos [4]. The boundary handler extends the left/right boundary of a cropped panoramic image of ERP format. Bounding boxes extracted by a CNN-based object detector and are incorporated into an extended DeepSORT tracker to estimate the target state [5], where DeepSORT is in the framework of a Kalman filter. The cubemap projection (CMP) format and its variants have the problem of boundary discontinuity and deformation (Figure 1a). To solve the problem of boundary discontinuity and deformation of the CMP of road panorama captured by vehicle-mounted cameras, the pre-processing phase of the mean-shift-based tracker of rigid objects (e.g., car, traffic sign, or building) extends faces using the epipolar geometry that indicates the moving direction of the scene in [6]. To preserve edge continuity, one quarter (triangle) of the top image and one quarter (triangle) of the bottom image are stitched to the left/front/right/back image with pixel expansion.

1.2. Siamese Networks Based Visual Trackers

A target can be represented by handcrafted features or deep features. Handcrafted features are extracted using feature detection schemes with predefined algorithms and fixed kernels. Evaluation of precision-recall curves of several local handcrafted feature detection and description algorithms have been provided in [8]. On the other hand, by minimizing a loss function, a deep network framework that consists of linear layers (e.g., a convolution layer) and non-linear layers (e.g., an activation function) learns from a large scale of training dataset in the training phase. In the test phase, deep features are extracted using the trained network. Deep learning-based trackers benefit from learning with a huge dataset. The first deep learning tracker DLT adopted the stacked denoising autoencoder with off-line training and on-line fine-tuning [9]. Deep learning-based trackers can be categorized into two groups. The first group is based on tracking-by-detection. This group incorporates deep features that are taken as detection into a tracking framework. The second group follows tracking-by-matching, where one sub-group off-line trains the Siamese networks for similarity learning or regression. Siamese networks were proposed by Jane Bromley et al. in 1993 [10]. In the tracking phase, similarity between deep features of the target template and the candidate proposal in the current frame are computed to predict the bounding box of the target. Thus, trackers of the second group work well even the target sample has not been learned off-line. Moreover, it benefits from the small number of convolutional layers and thus it achieves real-time tracking. The classic and pioneer trackers of the second group are GOTURN and SiamFC [11,12]. Fully connected layers-based similarity computation of GOTURN (generic object tracking using regression networks) predicts coordinates of anchors of the bounding box [11]. SiamFC adopts the fully convolution network (FCN) where similarity is computed efficiently using cross correlation [12]. Such strategy enables the similarity to be computed in a translation manner. The output is a score map defined on a finite grid that indicates the likelihood of the target measurement. In SiamRPN [13], the Siamese subnetwork for feature extraction is followed by the region proposal subnetwork that includes regression and classification branches [13]. In the tracking phase, the input of the template branch is the template from the first frame. Then one-shot detection in the other frames is based on the category information from the template branch. Deep analyses of Li et al. found the lack of strict translation invariance of Siamese trackers [14]. Thus SiamRPN++ was proposed where the effective sampling strategy breaks the restriction. SiamRPN++ successfully enables training of the Siamese tracker with ResNet as a backbone network to obtain performance improvement [14]. Considering most Siamese trackers suffer from drifting problems due to learning global representation of the entire object, Liang et al. proposed LSSiam. Local semantic features are learned with additional classification branch off-line and then the classification branch is removed in the tracking phase [15]. Considering that semantic features learned for classification and appearance features learned for similarity complement each other, SA-Siam is composed of a semantic branch and a similarity branch [16]. Both branches are in the framework of Siamese network. The discriminative power of the semantic branch is improved by being weighted by an attention mechanism. The overall heat map (i.e., score map) of SA-Siam is generated using the weighted average of heat maps of two branches. In adaDCF [17], Fisher linear discriminant analysis (FDA) helps the tracker enhance discrimination between the positive (i.e., foreground) and negative (i.e., background) samples. A convolutional FDA layer is incorporated into the network framework with on-line fine tuning.

1.3. Template Update for Siamese Networks Based Visual Trackers

Although improvement of Siamese networks-based framework and loss functions helps learn visual representation off-line, template update plays a key role in the success of on-line tracking. SiamFC lacks online adaptability since it neither adopts on-line fine tuning nor template update [12]. Thus, it is not robust against geometric deformation and partial occlusions over time. A template is the exemplar that represents the target model (e.g., appearance model). In the tracking phase, template update replaces the original template with a new one to improve tracking accuracy. However, for a target with significant change of visual features over time, tracking with the target template extracted from the first frame suffers from drifting problems. Motivated by this, several template update schemes have been designed for Siamese networks-based trackers. For example, Valmadre et al. proposed a correlation filter-based deep tracker CFNet for end-to-end training [18]. The baseline Siamese networks-based tracker was modified with a correlation filter block to construct a discriminative template for robustness against translation, where a new template is generated for each new frame [18]. In Yang et al. [19], attentional Long Short-Term Memory (LSTM) control memory controls memory reading and writing. At each time instant, a residual template retrieved from the memory is combined with the initial template extracted from the first frame to build the new template. After Siamese networks-based target prediction, features of the cropped new target template corresponding to the predicted bounding box are written into memory for model update. Although the aforementioned template update schemes significantly improve tracking accuracy of SiamFC, an aggressive template updating strategy that does not introspect tracking accuracy easily leads to extra computation cost and drift problems. Moreover, the focus of these schemes is how to generate a new template instead of when to update template. Detection of the right timing of template update should be more efficient and effective. To deal with the problem that the annotation is available only in the first frame in the on-line tracking phase, clues extracted from tracking results can reflect tracking accuracy to some extent. Only a few schemes have proposed Siamese networks-based timing detector of template update. For example, Xu et al. use the highest value of a score map to determine the timing of update for UAV-based visual tracking [20]. If the value is small, conditions of deformation and partial occlusions will be further differentiated by the template update module, comprised of a contour detection network and a target detection network. If outputs of two networks are consistent, the template will be updated since deformation may occur. Otherwise, the template will be not updated to avoid the new template containing partial occlusions. For a Siamese networks-based score map, Wang et al. proposed the average peak to correlation energy (APCE) that denotes the ratio of the difference between the maximum and minimum value to the variation [21]. A higher APCE indicates the higher confidence of prediction being the target. When both the maximum value of the score map and APCE are greater than their weighted historical averages, the template is updated. In Liang et al. [15], proposed to detect the update timing by taking account into both the update interval and APCE-based confidence score [21].

1.4. Motivation and Contributions of The Proposed Scheme

Generally, a 360-degree image is stored as an ERP image after image acquisition and image stitching. However, it can be noticed that videos occupy 80% of internet traffic. Equiangular cubemap (EAC) format is one of variants of CMP format and it has less non-uniform geometric deformation than ERP (Figure 1) [2]. Thus, it enables more efficient video compression than ERP and higher tracking accuracy [22]. Each EAC image has six faces. All angles in the sphere domain have equal pixel counts in EAC format. Problems with tracking in EAC format are stated as follows. First, sphere-to-plane projection may make the target in the real world be projected onto distant regions on an EAC image. Accordingly, content discontinuity of EAC fails most state-of-the-art trackers. Second, to provide immersive viewing experiences, a 360-degree video has a huge amount of data. It is difficult to achieve real-time tracking. Third, non-uniform geometric deformation of EAC format makes the target appearance vary with frame index. Tracking accuracy will decrease if the template is not updated. Few visual trackers have been proposed for 360-degree EAC format videos. Since most handcrafted feature descriptors are not robust against non-uniform geometric deformation in the EAC format, Tai et al. proposed face stitching for the EAC format and a preliminary work of a timing detector of template update for the EAC format by referring to statistics of the score map [23], generated by SiamFC [12]. A total of 1600 local windows of score maps of eight 360-degree videos are analyzed. Analyses of observations indicate that a local window with high mean and low standard deviation implies the higher overlap ratio between the predicted bounding box and the ground truth of the target. However, thresholds of mean and standard deviation of a score map are selected for individual test videos heuristically in Tai et al. [23]. Thus, this paper proposes a Siamese networks-based people tracker using machine learning-based template update for 360-degree EAC format videos. Major contributions of the proposed scheme include: (1) regardless of the content discontinuity, the proposed tracker with face stitching can track a target that moves between any adjacent faces of EAC format. (2) The proposed machine learning-based timing detector of template update is built using a binary Bayes classifier, a statistically optimal classifier [24]. Statistical features and the linear discriminant feature of a score map, generated by the fully convolutional Siamese networks, are extracted and referred to. Statistical features include the mean and standard deviation. The linear discriminant feature is extracted using a dimension reduction Fisher linear discriminant (FLD) algorithm for statistical optimization [25]. The timing detector helps the tracker evaluate tracking accuracy without ground truth. Thus, the tracker can determine whether the template should be updated or not. (3) The proposed tracker operates at 52.9 fps, beyond the frame acquisition rate.
The remainder of this paper is organized as follows. Section 2 introduces the ERP plane-to-sphere projection and sphere-to-EAC plane projection. It also provides overviews of the classic fully-convolutional Siamese networks-based tracker, SiamFC [12]. Section 3 proposes the Siamese networks-based people tracker using timing detector of template update for EAC format. Section 4 provides experimental results and analyses. Finally, Section 5 concludes this paper.

2. Overview of Projection Format Conversion and Fully-Convolutional Siamese Networks Based Tracker

Before visual tracking on 360-degree EAC format videos, an ERP image is converted into the sphere domain and re-projected onto the planar representation of the EAC format. Thus, this section first states format conversion from ERP to EAC (Section 2.1). Next, the fully-convolutional Siamese networks-based tracker SiamFC is introduced (Section 2.2).

2.1. Conversion from Equirectangular Projection (ERP) to Equi-Angular Cubemap (EAC)

Generally, a 360-degree image is stored as an ERP image after image acquisition and image stitching. To convert an ERP image into an EAC image, two phases are proceeded: ERP plane-to-sphere projection and sphere-to-EAC plane projection [2]. An example of conversion from the ERP format to the EAC format for the 1330th frame of Video #3 (London Park Ducks and Swans) is shown in Figure 2. In the first phase, a sampling point position (i, j) in an ERP image is converted to the position (u, v) in the 2-D projection plane by
u = ( i + 0.5 ) / W , 0 i < W
v = ( j + 0.5 ) / H , 0 j < H
where W and H are the width and height of an ERP image, respectively. Next, the longitude and latitude ( φ , θ ) in the unit sphere are calculated by
φ = ( u 0.5 ) × 2 π
θ = ( 0.5 v ) × π
X = cos ( θ ) cos ( φ )
Y = sin ( θ )
Z = cos ( θ ) sin ( φ )
In the second phase, given the point position (X, Y, Z) in the sphere domain, the position ( u , v ) on the 2-D projection plane and the face index of an EAC image are computed [2]. Thus, the 2-D position ( u ˜ , v ˜ ) on an EAC image is calculated by
u ˜ = ( 4.0 / π ) × tan 1 ( u )
v ˜ = ( 4.0 / π ) × tan 1 ( v )

2.2. Fully-Convolutional Siamese Networks Based Tracker

The classic fully-convolutional Siamese networks-based tracker, SiamFC, was proposed by L. Bertinetto et al. [12]. The advantage of the Siamese networks is that they learn the similarity function of any pair of image patches where the convolution phase includes only several convolution layers. Thus, it is advantageous that it can track a target that has not been seen in the off-line training phase at high speed. SiamFC adopts neither on-line fine-tuning nor template update. Instead of fully connected layers-based regression, the cross-correlation-based similarity between feature maps of the template and those of the sub-image of the search image is computed efficiently since the operation is equivalent to use the inner product. The input of SiamFC includes the exemplar image z (i.e., the target template) and a candidate image x. With the aid of the VID dataset of the 2015 edition of the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) dataset [26], the network learns the cross correlation-based similarity function f ( z , x ) that is defined as [12]
f ( z , x ) = g ( z ) ( x ) + b
where g ( ) is a convolutional embedding function, and b is a location independent real value. Implementation of g ( ) in [7] follows the convolution architecture that consists of five convolution layers, two max pooling layers, and four ReLU (Rectified Linear Unit) layers of AlexNet [27]. Padding is avoided to keep the property of translation invariance of convolution. In the training phase, each convolution layer is followed by batch normalization to avoid the problem of an exploding/vanishing gradient. The output of the network is a score map defined on a finite grid D Z 2 .
The training phase aims at minimizing the mean of the logistic loss of training samples by updating parameters θ of the network:
arg m i n θ E ( z , x , a ) L ( a , f ( z , x ; θ ) )
where the logistic loss of a score map is defined as
L ( a , c ) = 1 | D | r D l o g ( 1 + e a [ r ] c [ r ] )
r is a 2-D position in the score map. a [ r ] { 1 , + 1 } is the ground truth label of the position r, determined by the distance between the ground truth of the target position and the center of the candidate image. c [ r ] is the score corresponding to the exemplar image and the candidate image centered at the rth position within the search image. The stochastic gradient decent is adopted for optimization of Equation (11). In the tracking phase, the 2-D position with the highest score is taken as the predicted center of the target in the current frame.

3. The Proposed Tracking Scheme for EAC Format of 360-Degree Videos

At the off-line training phase, score maps are generated by applying Siamese networks-based tracker SiamFC to the dataset that consists of single-view videos [28]. Then the training dataset of the proposed timing detector is constructed using feature vectors of 100 × 100 sub-images, cropped from the image patch centering at the location with the highest score on a score map. Each training sample is labeled according to the overlap ratio between the predicted bounding box and the ground truth. After training the Fisher Linear Discriminant (FLD)-based projection vector for dimensionality reduction [25,29], each sub-image of a score map is represented by a feature vector including statistics of the sub-image and the FLD-based feature (Section 2.2). Then a binary Bayes filter-based timing detector is trained. Both the Fisher linear discriminant (FLD) and Bayes filter are supervised learning-based methods.
In the tracking phase of the proposed scheme (Figure 3), the target is manually detected with a tight bounding box drawn around the target in the first frame. Then the image patch with the context of the target is taken as the exemplar image (i.e., template). To be robust against content discontinuity between some adjacent faces, face stitching is applied before tracking (Section 3.1). SiamFC (Section 2.2) predicts the bounding box of the target in stitched images instead of original EAC images (Section 3.2). By referring to the predicted bounding box in the previous frame, a 100 × 100 sub-image of the current score map is cropped. Instead of an aggressive update strategy, the proposed machine learning-based timing detector of template update selects the right timing of the update (Section 3.3).

3.1. Face Stitching of EAC Format

Tai et al. proposed a face stitching scheme that successfully helps existing trackers to keep tracking on EAC images in [23]. Due to the sphere-to-EAC plane projection, the target in the sphere domain may be projected onto regions that do not belong to the same connected component in an EAC image. An example is shown in Figure 3, where the target (blue bounding box) in the sphere domain (real world) is projected onto disjoint regions in an EAC image (the 249th frame of Doi Sutheup [7]). Moreover, these disconnected components may be distant from each other. Accordingly, EAC images of 360-degree videos make most existing high-accuracy trackers fail. However, face stitching helps a tracker keep tracking across faces where borders are discontinuous. Moreover, the scheme has low computation complexity and avoids raising new geometric deformation in stitched images. The scheme is stated as follows.
An EAC image with six faces is indexed by {0,1,2,3,4,5}, where 0 denotes front, 1 denotes back, 2 denotes top, 3 denotes bottom, 4 denotes right, and 5 denotes left. In an EAC image acquired at frame rate either 30 fps or 60 fps [2], people usually cannot move from the current face to the opposite face of a cube in two consecutive frames. Motivated by this, the face stitching strategy only stitches four semantic neighboring faces in the sphere domain with the face indexed by N, where the center of the predicted bounding box of the target is located. Figure 4 shows the block diagram of face stitching. Let S denote the side length of a face. For the current frame, the central region (i.e., the central face) of a stitched square image with side length 3S is the face N that contains the center of the predicted bounding box in the previous frame. The central face indexed by 1, 2, or 3 is rotated upwards in stitched images. The four neighboring faces of the central face in the stitched image are the neighboring regions of the central face in the sphere domain. Regions containing four corners of the stitched image are filled with black pixels instead of extrapolated pixels from the existing faces. This is to avoid raising new geometric deformation in such regions. Figure 5 shows examples of stitched images that center at individual faces of the 1360th frame of London Park Ducks and Swans [7]. The face containing the smiling face indicates the target location. The orientation of the index number corresponds to the north–south axis in the sphere domain.
At the tracking phase, the exemplar image (i.e., the target) and a search window (both in the stitched image domain) are input into the Siamese networks-based tracker. The coordinate of the predicted bounding box of the target in the stitched image is converted back to the coordinate in the EAC image by the function T N ( ) , where the EAC image is packed in 3 × 2 format. That is,
x N , e = T N ( x s )
where x s = [ u ˜ s , v ˜ s ] t is the 2-D coordinate in the stitched image and x N , e = [ u ˜ N , e , v ˜ N , e ] t is the 2D coordinate in the Nth face of an EAC image. Figure 6 shows conversion for the target located in different faces. In some cases, the center of the predicted bounding box is not located in the central face of the stitched image. It indicates that the target just moved across different faces.

3.2. Feature Extraction of Score Maps Using Machine Learning Based Dimensionality Reduction

The adequate timing of template update is when the predicted bounding box of the target has high tracking accuracy. Thus, the problem of how to decide the update timing turns into the problem of how a tracker identifies the tracking result is good without knowledge of ground truth. Although Tai et al. found that mean and standard deviation of a sub-image of the score map of Siamese networks [23], SiamFC [12], were related to tracking accuracy, the mean and standard deviation-based update timing was decided for individual test videos manually and heuristically in [23]. Thus, this subsection first provides analyses of conditional probability distributions of the mean, standard deviation, and linear discriminant feature of score maps given the tracking result with high/non-high tracking accuracy. Next, this subsection proposes a statistics-based feature vector that consists of a mean, standard deviation, and dimensionality reduction-based feature of score maps for a machine learning-based timing detector of template update in Section 3.3.
Face stitching solves the problem of content discontinuity between several adjacent faces of an EAC image. However, black padding pixels on four corners raised by face stitching lead to occlusions, similar to the one caused by the surroundings of the target (Figure 7b). In the meantime, non-uniform geometric deformation on boundaries of each face of EAC format still exists in a stitched image. However, compared with the ERP format and most variants of the CMP format, the EAC format features less non-uniform geometric deformations. Thus, this paper adopts single-view videos instead of 360-degree EAC format videos as training data of dimensionality reduction and a binary Bayes classifier for decision of update timing. Since the focus of this paper is people tracking, 21 videos containing people were selected from the OTB100 dataset [28]. These videos include Blurbody, Crossing, Crowds, Dancer, Dancer2, Girl, Gym, Human2, Human3, Human5, Human6, Human7, Human9, Jogging.1, Jogging.2, Singer1, Skating2.2, Subway, Walking, Walking2, and Woman. Videos containing people but with low tracking accuracy of SiamFC were not selected. The reason is that they cannot provide sufficient positive samples (i.e., score maps with high tracking accuracy). For selected videos, the number of training samples was approximately 10,000 while all 11 attributes (illumination variation, scale variation, occlusion, deformation, motion blur, fast motion, in-plane rotation, out-of-plane rotation, out-of-view, background clutters, low resolutions) of the OTB dataset were included. The trained SiamFC tracker was applied to each selected video. The tracker generates a score map for each video frame. A 100 × 100 sub-image centering at the location with the highest score on the score map is cropped. Each sub-image is taken as a training sample. A training sample will have the positive label if the tracking result of the corresponding video frame has a high overlap ratio (Section 3.2). Otherwise, it will have the negative label. Accordingly, the training dataset that contains 10,407 training sample is generated. This paper adopts the overlap ratio-based tracking accuracy as follows.
First, the overlap ratio is defined by [28]
O r = | r p r g | / | r p r g |
where O r [ 0 , 1 ] is the overlap ratio. r p and r g are bounding boxes of the tracking result and ground truth, respectively. To improve tracking accuracy using template update, a new template that has a dissimilar appearance model with the original target template should be avoided. For example, an image patch that contains an occluded target or cluttered background is improper to use as a new template. It is suggested that the tracking result with O r that is greater than 0.6 is taken as high tracking accuracy in [13]. Although it will be the better update timing if the tracking result has higher accuracy (e.g., overlap ratio is 0.9), higher overlap ratio leads to infrequent update. Thus, comparisons of success plots and precision plots among the proposed tracking scheme with overlap ratios 0.6, 0.7, and 0.8 are shown in Figure 8. The success plot is generated using the overlap ratio between the predicted bounding box and the ground truth. If the overlap ratio is greater than a given threshold, the target will be taken as being tracked successfully in one frame. In the success plot, the success rate represents the ratios of successful frames at different thresholds that range between 0 and 1. The precision plot is generated using the Euclidean distance-based location error between the center of the predicted bounding box and that of the ground truth. The precision rate indicates the ratio of frames with location error that is less than a given threshold [28]. Tracking accuracy of the proposed scheme with overlap ratio 0.7 (orange curve) is similar to that of the proposed scheme with overlap ratio 0.6 (green curve). However, the proposed scheme with overlap ratio 0.8 (green curve) significantly outperforms the proposed scheme with overlap ratio 0.7 (orange curve) and that with 0.6 (black curve). Thus, a training sample will be labelled as a positive sample (i.e., high tracking accuracy) if Or is greater than 0.8 and otherwise it will be labelled as a negative sample (i.e., non-high tracking accuracy). The training dataset is adopted for analyses of conditional probability distribution and training of the projection matrix of dimensionality reduction.
A score value on the score map of a fully convolutional Siamese networks represents similarity between the template and a candidate proposal that centers at the location in the search window. Figure 9 shows examples of tracking results of stitched images of EAC format and their corresponding score maps. Bounding boxes (i.e., tracking results) in Figure 9 have overlap ratios 0.74 and 0.89, respectively, and their corresponding score maps shown in Figure 9b,f both feature with high mean and high standard deviation. On the other hand, bounding boxes (i.e., tracking results) in Figure 9c,g have overlap ratios 0.14 and 0.45, respectively, and the corresponding score maps shown in Figure 9d,h both feature low mean and low standard deviation. That is, content characteristics of a score map are related to tracking accuracy in terms of overlap ratio. Thus, conditional probability mass functions (PMFs) of features (i.e., mean (Figure 10a) and standard deviation (Figure 10b) of the score map given high overlap ratio and those of features given non-high overlap ratio are analyzed. In Figure 10, the value of each feature is normalized to range between 0 and 1. Although the range of the feature value (e.g., mean) with high overlap ratio is similar to that of a feature with non-high overlap ratio, modes of these two groups distinctly separate from each other. In the meantime, the mode of the class with high overlap ratio has high mean (Figure 10a) and high standard deviation of scores (Figure 10b). Thus, it can be inferred that a probabilistic-based classifier is proper to classify an input sample that is represented by features of a score map.
In addition to statistics including mean and standard deviation of the score map providing clues to reflect tracking accuracy, this paper proposes to explore the discriminant feature of a score map by machine learning-based dimensionality reduction for binary classification of tracking accuracy. Dimensionality reduction transforms input data in high dimensional space to a low dimensional space without loss of information and aids in reducing the effect of outliers of the data [32]. Dimensionality reduction can be realized using either deep learning-based methods or non-deep learning-based methods. Considering dimension reduction and statistical optimization, this paper proposes to adopt a non-deep learning-based method, Fisher linear discriminant (FLD) [25], to explore the discriminant feature of a score map. Examples of non-deep learning-based methods are principle component analysis (PCA), independent component analysis (ICA), t-distributed stochastic neighbor embedding (t-sne), and Fisher linear discriminant (FLD) [25]. FLD was proposed by Ronald Aylmer Fisher in 1936 [25]. It is a supervised learning method used in machine learning, pattern recognition, and statistics. It aims at estimating the projection matrix from the training dataset off-line to well separate different classes. For two Gaussian classes where the covariance of one class is a multiple of the other, the Fisher discriminant yields the optimal linear discriminant. Otherwise, FLD produces a linear discriminant that optimizes the weighted classification error [33]. The objective function focuses on minimization of the within-class scatter matrix and maximization of between-class scatter matrix of training samples. In the test phase, the FLD-based feature vector of an input sample X can be computed by projecting the data X onto the projection matrix. The n dimensional data X is projected onto the m dimensional data Y by the projection matrix P ^ with m < n .
Y = P ^ T X
In the training phase of FLD, the divergence of training samples is measured by scatter matrices. Assume there are L classes in the training dataset. For training data X ^ , the within-class scatter matrix S w is defined by [29]
S w = i = 1 L P i E [ ( X ^ M i ) ( X ^ M i ) t | ω i ]
where P i is the prior probability of the ith class, and M i is the mean vector of the ith class, ωi. The between-class scatter matrix S b is defined by
S b = i = 1 L P i ( M i M o ) ( M i M o ) t
where M o is the mean vector of all training samples. The divergence of all training samples, mixture scatter matrix, is equal to the sum of S w and S b
S M = E [ ( X ^ M o ) ( X ^ M o ) t ] = S w + S b
where m is less than n. Thus, the between-class scatter matrix of the input data after dimensionality reduction Y ^ of FLD is computed by
S b Y ^ = i = 1 L P i ( P ^ t M i P ^ t M o ) ( P ^ t M i P ^ t M o ) t = P ^ t S b P ^
In addition, the within-class scatter matrix of Y ^ is computed by
S w Y ^ = P ^ t S w P ^
Thus, the training phase of FLD is to find P ^ o p t that maximizes the objective function t r ( S w Y ^ 1 S b Y ^ )
P ^ o p t = a r g   m a x P ^   t r ( S W Y ^ 1 S b Y ^ )
where P ^ o p t satisfies
( S W 1 S b ) P ^ o p t = P ^ o p t ( S W Y ^ 1 S b Y ^ )
By diagonalizing S w Y ^ and S b Y ^ ,
B T S b Y ^ B = U m
B T S w Y ^ B = I m
where Um and I m are diagonal matrices, and B is a m × m square matrix that is invertible. By introducing Equations (23) and (24) into Equation (22),
( S W 1 S b ) P ^ o p t B = ( P ^ o p t B ) U m
Finally, the optimal projection matrix P ^ o p t is constructed from the first m eigenvectors that correspond to the first m greatest eigenvalues. In the test phase of FLD, the optimal projection matrix P ^ o p t reduces an input sample (e.g., a score map) in the n-dimensional space to the m-dimensional space.
It is noticed that if there are insufficient training samples, FLD will fail due to the problem of small sample space (SSS) [29]. In addition, the n-dimensional data with L classes can be reduced to at most L 1 dimensions by FLD given that n is greater than L 1 [29]. In the proposed detector of tracking results with high overlap ratio, n is 10,000 and L is 2 (i.e., class of tracking results of high overlap ratio and class of tracking results of non-high overlap ratio). Accordingly, each sample (i.e., score map) in our proposed scheme has a one-dimensional FLD-based feature with the aid of a projection matrix that is trained off-line. Figure 11 shows conditional probability mass functions (PMFs) of the FLD-based feature using 10,407 training samples. The conditional PMM when the score map has high overlap ratio differs from the conditional PMF when the score map has non-high overlap ratio. A score map with a lower feature value (e.g., 0.4) has higher probability to have high overlap ratio. Thus, each score map can be represented by a three-dimensional feature vector that consists of the mean of a score map, the standard deviation of a score map, and an FLD-based feature of a score map. Figure 11 shows two clusters of 10,407 training samples, where a yellow dot denotes a sample with high overlap ratio and a purple dot denotes a sample with non-high overlap ratio.

3.3. The Proposed Machine Learning Based Timing Detector of Template Update

By referring to conditional probability distributions of features of a score map of SiamFC (Section 3.2), it is reasonable to adopt a probabilistic-based classifier to decide the timing of template update where the input sample is represented by features of a score map. Thus, this subsection proposes to adopt the optimum statistical classifier, the Bayes classifier, for determination of update timing. In the following, an overview of the Bayes classifier is provided. Then the proposed Bayes classifier-based timing detector of template update for the fully convolutional Siamese networks-based tracker is stated.
The Bayes classifier predicts the label of an input sample by referring to the pre-trained likelihood distributions p ( X | ω k ) and prior distributions (i.e., the occurrence of the kth class) p ( ω k ) , where X is input data, ωk is the kth class, and k ranges from 1 to W. The Bayes classifier aims at minimization of the total average loss [24]
r j ( X ) = k = 1 W L k j ( X ) p ( X | ω k ) p ( ω k ) / p ( X )
where L k j ( X ) is the loss caused by assigning the input sample X to the jth class while X belongs to the kth class. p ( ω k | X ) is the posterior probability that X comes from ω k .
For binary classification (i.e., W = 2) of our proposed scheme, ω 0 and ω 1 are the class with non-high overlap ratio and the class with high overlap ratio, respectively. The input data X = [ x M , x s , x F ] t , where x M [ 0 , 1 ] is a n × n sub-image of the mean of a score map, x s [ 0 , 1 ] is the standard deviation of the n × n sub-image of a score map, and x F [ 0 , 1 ] is Fisher linear discriminant-based 1-D feature of an n × n sub-image of a score map, where the unnormalized value of x F ranges from −5.6 to 2.4. p ( X | ω 0 ) is the likelihood of the data given that X belongs to the class with high overlap ratio, and p ( X | ω 1 ) is the likelihood of the data given that X belongs to the class with non-high overlap ratio. Both p ( X | ω 0 ) and p ( X | ω 1 ) are learned off-line. Since there will be no loss if the label of X and the predicted result of the classifier are the same, L 00 ( X ) = 0 , L 01 ( X ) = 1 , L 10 ( X ) = 1 , and L 11 ( X ) = 0 in tests. In the test phase of the proposed detector of update timing, the predicted result Y is
Y = { 0 i f L 01 ( X ) p ( X | ω 0 ) p ( ω 0 ) < L 10 ( X ) p ( X | ω 1 ) p ( ω 1 ) . 1 o t h e r w i s e .
That is, if r 1 ( X ) is less than r 0 ( X ) , the predicted result of X will be 1. That is, the score map represented by the feature vector X will be taken as having non-high overlap ratio. Thus, the template should not be updated since the tracking result is inaccurate. If r 1 ( X ) is greater than r 0 ( X ) , the predicted result will be 0. It suggests that the template can be updated since tracking accuracy is high. Since p ( X ) does not vary with r j ( X ) , it can be ignored during comparisons of decision functions of two classes.
In the tracking phase, the fully convolutional Siamese networks-based tracker SiamFC computes the score map using the template and the candidate proposal extracted from the search window. For the sub-image of the score map, the feature vector that consists of the mean, the standard deviation, and the Fisher linear discriminant-based feature extracted by the projection direction is constructed. By referring to the feature vector, the Bayes classifier-based timing detector of template update decides whether the original template is replaced by the estimated bounding box of the current frame or not. To the avoid template drifting problem caused by frequent update, the duration between two updates is constrained to be not less than T P frames. If the timing detector of template update suggests that the template can be updated after prediction of the bounding box in the tth frame and there is no template update from the (t − ( T P 1 ))th frame to the tth frame, the proposed tracker will take the action of template update.

4. Experimental Results

The proposed scheme is realized by revising the implementation of SiamFC [34]. Test results are divided into five parts. The first part (Section 4.1) provides performance analyses of individual components of the proposed scheme. Comparisons among SiamFC [12], the original SiamFC with face stitching proposed in [23], and our proposed scheme are made. The second part (Section 4.2) compares the proposed scheme with the only existing tracker designed for 360-degree EAC format videos, proposed by Tai et al. [23]. Since the focus of the proposed scheme is the score map-based timing detector of template update, the third part (Section 4.3) compares the proposed timing detector with a state of the art score map-based timing detector of template update [15,21]. For this, the authors implement the timing detector APCE (Section 1.3) which is proposed by Wang et al. and is adopted by LSSiam (Section 1.3) with the baseline of the Siamese networks-based tracker SiamFC [12,15,21]. The fourth part (Section 4.4) validates the effectiveness of the proposed template update scheme for Siamese networks-based trackers on 360-degree EAC format videos. The Siamese networks-based tracker SA-Siam [16] is revised to integrate face stitching and the proposed timing detector of template update. The fifth part provides analyses of time complexity of all considered trackers in Section 4. In addition, the execution time of individual components of the proposed scheme is analyzed.
To validate the effectiveness of the proposed scheme, the training dataset of the Fisher linear discriminant and the Bayes classifier-based timing detector of the proposed scheme does not overlap with the test dataset of the proposed tracking scheme. SiamFC is trained using the VID dataset of the ILSVRC2015 dataset that consists of a variety of videos with more than one million frames [26]. Implementation details are as suggested in [12]. The training dataset of the Fisher linear discriminant and the Bayes classifier-based timing detector of the proposed scheme is generated using 21 single-view videos in [28], as described in Section 3. The test dataset consists of eight 360-degree EAC format videos with a variety of test conditions, including Spotlight (video #1) [31], London on Tower Bridge (video #2) [7], London Park Ducks and Swans (video #3) [7], Doi Suthep (video #4) [7], Amsterdam (video #5) [31], Paris-view6 (video #6) [31], Ping-Chou Half-day Tour (video #7) [30], and Skiing (video #8) [31]. Videos #1 and #8 are captured by moving cameras, and videos #2-#7 are captured by static cameras. Except video #6 that is captured at 60 fps, the frame rate of each test video is 30 fps. Videos #1 and #6 feature illumination variations. Significant geometric deformation of targets occurs in videos #4 and #8 (Figure 12). The target (i.e., a skier) has a complex motion model (e.g., in-plane rotation) in video #8. The target in video #6 is occluded for a period of time. The target moves across different faces in all test videos. Videos are downsampled before tracking, where the spatial resolution of one downsampled face of videos #1–#6 and #8 is 144 × 144 , and that of video #7 is 588 × 588 . The simulation environment is stated as follows. The CPU (Central Processing Unit) and GPU (Graphics Processing Unit) are i7-7700K 4.5GHz and GeForce GTX 1080 Ti, respectively. RAM is DDR4(48G). The OS (Operating System) is Ubuntu 18.04-x64. Tensorflow with Python 3.5 is adopted.

4.1. Performance Analyses of Individual Components of the Proposed Scheme

Table 2 provides comparisons of tracking accuracy in terms of average overlap ratio and average location error among the original SiamFC [12], SiamFC with face stitching (SiamFC + S), and our proposed machine learning-based scheme (SiamFC + S + P). Test results show that face stitching helps existing trackers (e.g., SiamFC) in the face of content discontinuity between several faces of EAC format. The reason is that there is a significant gap between location error and overlap ratio of SiamFC and those of SiamFC + S for most of the test videos, respectively. Moreover, the proposed timing detector of template update (SiamFC + S + P) further improves the tracking accuracy of SiamFC + S on videos #4-#8. Tracking accuracy of the proposed scheme (SiamFC + S+P) is similar to that of SiamFC + S on videos #1 and #2 due to low update frequency of the proposed scheme, where the proposed scheme achieves a higher overlap ratio on videos #1 and #2 while SiamFC + S leads to a smaller location error on these two videos (Table 2). The target in video #3 features with slow motion and slight deformation such that SiamFC + S can track the target well. However, the proposed scheme with template update easily leads to a slight drifting problem and thus SiamFC + S slightly outperforms the proposed scheme (Table 3). Figure 12 provides tracking results of targets located in regions with significant geometric deformation of videos #4 and #8. For the 108th frame of video #4, the proposed scheme (Figure 12c) outperforms SiamFC + S (Figure 12a) and Tai et al. (SiamFC + S + H) (Figure 12b). For the 103th frame of video #8, both the proposed scheme (Figure 12f) and Tai et al. (SiamFC + S + H) (Figure 12e) significantly outperform SiamFC + S (Figure 12d).
By cropping from 1592 score maps that are calculated using SiamFC with face stitching and the proposed timing detector (SiamFC + S + P) on the test dataset, 1592 sub-images are generated. A total of 1496 test samples have negative labels and 96 samples have positive labels. Test results show that accuracy of the proposed timing detector approaches 90%, where there are 7 true positives, 89 false negatives, 75 false positives, and 1421 true negatives (Table 4). By referring to the ground truth, a sample will have a positive label if the tracking result of SiamFC + S + P has an overlap ratio that is greater than 0.8. Thus, 6% (i.e., 96/1592) of predicted bounding boxes (i.e., tracking results) using SiamFC + S + P have high overlap ratio. If a smaller threshold (e.g., 0.6) of overlap ratio is selected, the number of positive samples will increase, and the imbalance problem will be solved. However, template update with an inaccurate new template (lower overlap ratio) easily decreases tracking accuracy. The numbers of template updates of videos #1–#8 are 1, 4, 12, 10, 7, 3, 5, 5, respectively. It is noted that misclassification of a positive sample (i.e., the ground truth has high overlap ratio) causes the right timing to be missed while frequent update leads to the drifting problem. For the proposed timing detector with a true positive rate of 1.5%, low update frequency will avoid a worse one if visual features of the target change slightly over time. For targets with time-varying features, infrequent update with an accurate template still contributes to improvement of tracking accuracy (e.g., videos #7 and #8 (Figure 13 and Figure 14). On the contrary, misclassification of a negative sample (i.e., the ground truth has non-high overlap ratio) leads to the original template being replaced by an inaccurate template and thus the drifting problem arises. However, the proposed timing detector has low false positive rate 3.7% as expected.
Templates selected by the proposed scheme are visually inspected in Figure 13, Figure 14, Figure 15 and Figure 16. It is noticed that SiamFC constructs a template by adding a margin to the scaled bounding box for context [12]. Regardless of the number of updates, most templates selected by the proposed timing detector have a high overlap ratio with the ground truth, except those of video #6. Tracking on video #6 is challenging. It has a cluttered background, and deformation and illumination variation occur after the target is fully occluded for a period of time (Table 5). However, template update with the proposed timing detector still slightly improves tracking accuracy of SiamFC + S on video #6 (Table 5). SiamFC + S on video #7 has an average overlap ratio of 0.236 and an average location error of 90.486. The proposed template update scheme (SiamFC + S + P) significantly improves tracking accuracy in terms of average overlap ratio 0.593 and average location error 9.264 (Table 2). The proposed scheme (SiamFC + S + P) also outperforms SiamFC + S on video #8. The reason is that visual features of the target vary over time notably while the target (i.e., a skier) has a complex motion model. SiamFC+S has an average overlap ratio of 0.235 and an average location error of 140.250 while the proposed scheme with the aid of template update has an average overlap ratio of 0.518 and an average location error of 27.011 (Table 2).

4.2. Comparisons with State-of-the-Art People Trackers for 360-Degree EAC Format Videos

The only existing visual tracker designed for 360-degree EAC format videos is proposed by Tai et al. (SiamFC + S + H) [23], where the heuristics-based timing detector was proposed. Thus, Table 6 provides comparisons of overlap ratio and location error between SiamFC + S + H and our proposed machine learning-based scheme (SiamFC + S + P).
Although SiamFC + S + H heuristically selects thresholds of mean and standard deviation for each test video [23], our proposed machine learning-based method (SiamFC + S + P) still outperforms SiamFC + S + H on videos #2, #3, #4, and #6 without heuristic settings of parameters. For video #1, tracking accuracy of SiamFC + S + P is similar to that of SiamFC + S + H. For video #5, although tracking accuracy of SiamFC + S + P (the proposed scheme) is slightly lower than that of SiamFC + S + H (Tai et al.) [23], tracking results of most frames of these two trackers are similar. Moreover, there is no difference between the predicted bounding box of proposed scheme and that of Tai et al. on the last frame of Video #5 (Table 7). For videos #7 and #8, SiamFC + S + P still significantly improves tracking accuracy of SiamFC + S; even SiamFC + S + P does not outperform SiamFC + S + H on videos #7 and #8. Figure 17 and Figure 18 provide comparisons of tracking results of SiamFC + S, SiamFC + S + H (Tai et al.) [23], and SiamFC + S + P (the proposed scheme) on frames of videos #2 and #4 with combination of non-uniform geometric deformation and occlusion, and fast motion as the target moves into the boundary of a face of EAC format. The face stitching scheme indeed helps the tracker solve the problem of content discontinuity between faces, and the proposed timing detector of template update selects the right timing such that the tracker can successfully adapt to the severe change of the target’s appearance at a short notice.

4.3. Comparisons with State-of-the-Art Score Map Based Timing Detector of Template Update

Only a few schemes have proposed Siamese networks-based timing detectors of template update. The authors implement the timing detector APCE that was proposed by Wang et al. and was adopted by LSSiam (Section 1.3.) [15,21]. In our implementation of the timing detector of LSSiam [10], the template will be updated if both of two conditions are satisfied. (1) The confidence score is greater than six. (2) Frame interval between two updates is not less than 40 frames, as suggested in [15]. Since the focus of our proposed scheme is how to detect the right timing instead of how to generate a new template, fair comparisons between our proposed timing detector and that of LSSiam are made using the same naive new template (i.e., the predicted bounding box at the time instant that is selected). Performance evaluation in terms of precision plot and success plot of the proposed scheme and SiamFC + S + Timing detector of template update of LSSiam on the test dataset is shown in Figure 19a,b, respectively. The success plot is generated using the overlap ratio between the predicted bounding box and the ground truth. If the overlap ratio is greater than a given threshold, the target will be taken as being tracked successfully in one frame. In the success plot, the success rate represents the ratios of successful frames at different thresholds that range between zero and one. The precision plot is generated using the Euclidean distance-based location error between the center of the predicted bounding box and that of the ground truth. The precision rate indicates the ratio of frames with location error that is less than a given threshold [28]. As we can see, the proposed scheme (SiamFC + S + P) outperforms SiamFC with face stitching and the timing detector of LSSiam (SiamFC + S +Timing detector of LSSiam) on both precision plots and success plots. The numbers of template updates of LSSiam + S on videos #1–#8 are 3, 4, 4, 4, 4, 4, 4, 0, 2, respectively.

4.4. Validation of the Effectiveness of the Proposed Template Update Scheme for Siamese Networks Based Tracking on 360-Degree EAC Format Videos

To validate the effectiveness of the proposed template update scheme for tracking on 360-degree EAC format videos, face stitching and a Bayes classifier-based timing detector of template update, referring to a Fisher Linear Discriminant (FLD)-based feature and statistical feature of a sub-image of a score map and face stitching (Section 3), are incorporated into SA-Siam [16]. The training dataset of FLD and the Bayes classifier is constructed using sub-images of overall score maps’ output by SA-Siam. Comparison of tracking accuracy between the SA-Siam with face stitching (SA-Siam + S) and SA-Siam combined with face stitching and the proposed timing detector (SA-Siam + S + P) is shown in Figure 20. The average overlap ratios of the SA-Siam + S and SA-Siam + S+P are 0.4884 and 0.5587, respectively. The average location errors of the SA-Siam + S and SA-Siam +S + P are 36.7006 and 9.8804, respectively. As we can see, the timing detector of template update significantly improves the tracking accuracy of SA-Siam. The number of template updates of SA-Siam + S on videos #1–#8 is 8, 1, 4, 0, 0, 2, 5, 14, respectively.
Comparisons of the tracking speed of considered trackers in Section 4. are shown in Table 8. For videos #1–#6 and #8, the tracking speed of the proposed tracker is 52.9 fps (beyond the acquisition rate). For video #7 which is not downsampled before tracking, the tracking speed of the proposed tracker is 11.2 fps. Although the tracking speed of the proposed scheme is slower than those of SiamFC with face stitching and the timing detector of template update of LSSiam and Tai et al. (SiamFC + S + H), the proposed scheme achieves the highest tracking accuracy among the compared trackers (Section 4.3). The percentage of computational time of components of the proposed scheme is further analyzed in Figure 21. Siamese networks-based prediction occupies the largest portion (49.2%) of the exeuction time while face stitching occupies the smallest portion (12.3%) of the execution time.

5. Conclusions

This paper proposes a Siamese networks-based tracker using template update and face stitching for 360-degree EAC format videos. The proposed scheme overcomes the problem of content discontinuity of the EAC format and enables fast speed (52.9 fps). The reasons why the proposed tracker operates beyond the acquisition rate are stated as follows. (1) The Siamese networks-based tracker enables fast speed since it has a small number of convolutional layers and simple operation of similarity. (2) Face stitching does not apply coordinate conversion between the EAC format and the sphere domain. (3) The Fisher linear discriminant and statistics of a score map are computed with low computation complexity. (4) The Bayes classifier-based timing detector of template update is computed with low computation complexity. Finally, with the aid of the proposed machine learning-based timing detector of template update that refers to the linear discriminant and statistical features of a score map, the proposed tracker can adapt to notable change of the target appearance over time. Future work will focus on design of a deep learning-based timing detector of template update.

Author Contributions

Conceptualization, K.-C.T. and C.-W.T.; methodology, K.-C.T. and C.-W.T.; validation, K.-C.T. and C.-W.T.; formal analysis, K.-C.T. and C.-W.T.; investigation, K.-C.T.; software, K.-C.T.; resources, C.-W.T.; data curation, K.-C.T.; writing—original draft preparation, K.-C.T. and C.-W.T.; writing—review and editing, K.-C.T. and C.-W.T.; visualization, K.-C.T.; supervision, C.-W.T.; project administration, C.-W.T.; funding acquisition, C.-W.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Ministry of Science and Technology of Taiwan under the Grants MOST 107-2221-E-008-058 and MOST-108-2221-E-008-024.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Azevedo, R.G.D.A.; Birkbeck, N.; De Simone, F.; Janatra, I.; Adsumilli, B.; Frossard, P. Visual Distortions in 360° Videos. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 2524–2537. [Google Scholar] [CrossRef]
  2. Zhou, M. AHG8: A study on equi-angular cubemap projection (EAC). In Proceedings of the Joint Video Exploration Team (JVET) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11 7th Meeting: JVET-G0056, Torino, Italy, 13–21 July 2017. [Google Scholar]
  3. ISO/IEC JTC 1/SC29/WG11, “WG11 (MPEG) Press Release”. In Proceedings of the 131st WG 11 (MPEG) Meeting, online. 29 June–3 July 2020.
  4. Liu, K.-C.; Shen, Y.-T.; Chen, L.-G. Simple online and realtime tracking with spherical panoramic camera. In Proceedings of the IEEE International Conference on Consumer Electronics, Las Vegas, NV, USA, 12–15 January 2018; pp. 1–6. [Google Scholar]
  5. Wojke, N.; Bewley, A.; Paulus, D. Simple online and realtime tracking with a deep association metric. In Proceedings of the IEEE International Conference on Image Processing, Beijing, China, 17–20 September 2017; pp. 3645–3649. [Google Scholar]
  6. Zhou, Z.; Niu, B.; Ke, C.; Wu, W. Static object tracking in road panoramic videos. In Proceedings of the IEEE International Symposium on Multimedia, Taichung, Taiwan, 13–15 December 2010; pp. 57–64. [Google Scholar]
  7. Available online: https://www.mettle.com/360vr-master-series-free-360-downloads-page (accessed on 1 June 2018).
  8. Mikolajczyk, K.; Schmid, C. A performance evaluation of local descriptors. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1615–1630. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Wang, N.; Yeung, D.Y. Learning a deep compact image representation for visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 809–817. [Google Scholar]
  10. Bromley, J.; Guyon, I.; LeCun, Y.; Sackinger, E.; Shah, R. Signature verification using a Siamese time delay neural network. Int. J. Pattern Recog. 1993, 7, 669–688. [Google Scholar] [CrossRef] [Green Version]
  11. Held, D.; Thrun, S.; Savarese, S. Learning to track at 100 fps with deep regression networks. In Proceedings of the European Conference on Computer Vision, Amsterdan, The Netherlands, 11–14 October 2016; pp. 749–765. [Google Scholar]
  12. Bertinetto, L.; Valmadre, J.; Henriques, J.F.; Vedaldi, A.; Torr, P.H.A. Fully-convolutional Siamese networks for object tracking. In Proceedings of the European Conference on Computer Vision, Amsterdan, The Netherlands, 11–14 October 2016; pp. 850–865. [Google Scholar]
  13. Li, B.; Yan, J.; Wu, W.; Zhu, Z.; Hu, X. High performance visual tracking with Siamese region proposal network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8971–8980. [Google Scholar]
  14. Li, B.; Wu, W.; Wang, Q.; Zhang, F.; Xing, J.; Yan, J. SiamRPN++: Evolution of Siamese visual tracking with very deep networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 4277–4286. [Google Scholar]
  15. Liang, Z.; Shen, J. Local semantic Siamese networks for fast tracking. IEEE Trans. Image Process. 2020, 20, 3351–3364. [Google Scholar] [CrossRef] [PubMed]
  16. He, A.; Luo, C.; Tian, X.; Zeng, W. A twofold Siamese network for real-time object tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4834–4843. [Google Scholar]
  17. Han, Z.; Wang, P.; Ye, Q. Adaptive discriminative deep correlation filter for visual object tracking. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 155–166. [Google Scholar] [CrossRef]
  18. Valmadre, J.; Bertinetto, L.; Henriques, J.; Vedaldi, A.; Torr, P.H.S. End-to-end representation learning for correlation filter based tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5000–5008. [Google Scholar]
  19. Yang, T.; Chan, A.B. Learning dynamic memory networks for object tracking. arXiv 2018, arXiv:1803.07268. [Google Scholar]
  20. Xu, Z.; Luo, H.; Bin, H.; Chang, Z. Siamese tracking with adaptive template updating strategy. Appl. Sci. 2019, 9, 3725. [Google Scholar] [CrossRef] [Green Version]
  21. Wang, M.; Liu, Y.; Huang, Z. Large margin object tracking with circulant feature maps. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 21–26. [Google Scholar]
  22. Wien, M.; Boyce, J.M.; Stockhammer, T.; Peng, W.-H. Standardization status of immersive video coding. IEEE J. Emerg. Sel. Topics Power Electron. 2019, 9, 5–17. [Google Scholar] [CrossRef]
  23. Tai, K.-C.; Tang, C.-W. Siamese networks based people tracking for 360-degree videos with equi-angular cubemap format. In Proceedings of the IEEE International Conference on Consumer Electronics-Taiwan, Taoyuan, Taiwan, 28–30 September 2020. [Google Scholar]
  24. Woods, R.; Czitrom, D.J.; Gonzales, R.C.; Armitage, S. Digital Image Processing, 3rd ed.; Prentice Hall: New Jersey, NJ, USA, 2008. [Google Scholar]
  25. Fisher, R.A. The use of multiple measurements in taxonomic problems. Ann. Eugen. 1936, 7, 179–188. [Google Scholar] [CrossRef]
  26. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Fei-Fei Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  27. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  28. Wu, Y.; Lim, J.; Yang, M.-H. Object tracking benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1834–1848. [Google Scholar] [CrossRef] [Green Version]
  29. Fukunaga, K. Introduction to Statistical Pattern Recognition, 2nd ed.; Elsevier: Amsterdam, The Netherlands, 2013. [Google Scholar]
  30. Shum sir, 360 VR, Shum sir Rebik’s Cube. 2017. Available online: https://www.youtube.com/watch?v=g5taEwId2wA (accessed on 19 August 2019).
  31. Duanmu, F.; Mao, Y.; Liu, S.; Srinivasan, S.; Wang, Y. A subjective study of viewer navigation behaviors when watching 360-degree videos on computers. In Proceedings of the IEEE International Conference on Multimedia Expo, San Diego, CA, USA, 23–27 July 2018; pp. 1–6. Available online: https://vision.poly.edu/index.html/index.php/HomePage/360-degreeVideoViewPrediction (accessed on 1 June 2019).
  32. Fodor, I.K. A Survey of Dimension Reduction Techniques, No. UCRL-ID-148494; Lawrence Livermore National Lab.: Livermore, CA, USA, 2020.
  33. Cooke, T.; Peake, M. The optimal classification using a linear discriminant for two point classes having known mean and covariance. J. Multivar. Anal. 2002, 82, 379–394. [Google Scholar] [CrossRef] [Green Version]
  34. Bilylee/SiamFC-TensorFlow. Available online: https://github.com/bilylee/SiamFC-TensorFlow (accessed on 27 March 2019).
Figure 1. The 1330th frame of video #3 (London Park Ducks and Swans) [7]. (a) The equi-angular cubemap (EAC) format is a variant of the cubemap projection (CMP) format. The EAC format has six faces (yellow dashed line). The target is in the blue bounding box. The dashed orange line indicates boundary discontinuity between faces. Sphere-to-plane projection projects the target onto distant regions on an EAC image. Non-uniform geometric distortions occur in each face. (b) The ERP format has more severe non-uniform geometric distortions than the EAC format.
Figure 1. The 1330th frame of video #3 (London Park Ducks and Swans) [7]. (a) The equi-angular cubemap (EAC) format is a variant of the cubemap projection (CMP) format. The EAC format has six faces (yellow dashed line). The target is in the blue bounding box. The dashed orange line indicates boundary discontinuity between faces. Sphere-to-plane projection projects the target onto distant regions on an EAC image. Non-uniform geometric distortions occur in each face. (b) The ERP format has more severe non-uniform geometric distortions than the EAC format.
Sensors 21 01682 g001
Figure 2. An example of conversion from the equirectangular projection (ERP) format to the EAC format for the 1330th frame of Video #3 (London Park Ducks and Swans) [7]. The orange dot denotes the location of a sample pixel.
Figure 2. An example of conversion from the equirectangular projection (ERP) format to the EAC format for the 1330th frame of Video #3 (London Park Ducks and Swans) [7]. The orange dot denotes the location of a sample pixel.
Sensors 21 01682 g002
Figure 3. Block diagram of the proposed tracking scheme (tracking phase). The input is the 249th frame of Doi Suthep [7].
Figure 3. Block diagram of the proposed tracking scheme (tracking phase). The input is the 249th frame of Doi Suthep [7].
Sensors 21 01682 g003
Figure 4. Block diagram of face stitching. The yellow face is taken as the central face. (EAC image: The 249th frame of Doi Suthep [7]).
Figure 4. Block diagram of face stitching. The yellow face is taken as the central face. (EAC image: The 249th frame of Doi Suthep [7]).
Sensors 21 01682 g004
Figure 5. Stitched images that center at individual faces of the 1360th frame of London Park Ducks and Swans [7]. The face containing the smiling face indicates the target location. The orientation of the index number corresponds to the north–south axis in the sphere domain. The indices of central faces in the stitched images are (a) 0, (b) 4, (c) 5, (d) 1, (e) 3, and (f) 2, respectively.
Figure 5. Stitched images that center at individual faces of the 1360th frame of London Park Ducks and Swans [7]. The face containing the smiling face indicates the target location. The orientation of the index number corresponds to the north–south axis in the sphere domain. The indices of central faces in the stitched images are (a) 0, (b) 4, (c) 5, (d) 1, (e) 3, and (f) 2, respectively.
Sensors 21 01682 g005
Figure 6. Conversion of the coordinate of the predicted bounding box of the target (blue) in different faces in the stitched image back to the coordinate in the EAC image by the TN (·). (a) Front. (b) Right. (c) Left. (d) Bac€ (e) Bottom. (f) Top.
Figure 6. Conversion of the coordinate of the predicted bounding box of the target (blue) in different faces in the stitched image back to the coordinate in the EAC image by the TN (·). (a) Front. (b) Right. (c) Left. (d) Bac€ (e) Bottom. (f) Top.
Sensors 21 01682 g006
Figure 7. An example of a deformed and occluded target in the stitched image of EAC format of 360-degree videos. (a) An EAC image: the 274th frame of London on Tower Bridge [7]. (b) The stitched image with the center face containing the deformed target (the people in black) in (a). The search window of SiamFC is in orange. The padding pixels (black square) are similar to the occluding pixels on the target.
Figure 7. An example of a deformed and occluded target in the stitched image of EAC format of 360-degree videos. (a) An EAC image: the 274th frame of London on Tower Bridge [7]. (b) The stitched image with the center face containing the deformed target (the people in black) in (a). The search window of SiamFC is in orange. The padding pixels (black square) are similar to the occluding pixels on the target.
Sensors 21 01682 g007
Figure 8. Comparisons among the proposed scheme with various settings of overlap ratio. (a) Success plots. (b) Precision plots.
Figure 8. Comparisons among the proposed scheme with various settings of overlap ratio. (a) Success plots. (b) Precision plots.
Sensors 21 01682 g008
Figure 9. Tracking results (yellow bounding box) of SiamFC [12], the ground truth of the target (blue bounding box), and local window centering at the highest score on the score map. (a) Tracking result and ground truth of the 12th frame of video Ping-chou Halfday Tour [30], captured by a static camera. The overlap ratio is 0.74. (b) The score map of (a). (c) Tracking result and ground truth of the 80th frame of the video Ping-chou Halfday Tour, captured by a static camera. The overlap ratio is 0.14. (d) The score map of (c). (e) Tracking result and ground truth of the 4th frame of video Skiing [31], captured by a moving camera. The overlap ratio is 0.89. (f) The score map of (e). (g) Tracking result and ground truth of the 40th frame of the video Skiing [31], captured by a moving camera. The overlap ratio is 0.45. (h) The score map of (g).
Figure 9. Tracking results (yellow bounding box) of SiamFC [12], the ground truth of the target (blue bounding box), and local window centering at the highest score on the score map. (a) Tracking result and ground truth of the 12th frame of video Ping-chou Halfday Tour [30], captured by a static camera. The overlap ratio is 0.74. (b) The score map of (a). (c) Tracking result and ground truth of the 80th frame of the video Ping-chou Halfday Tour, captured by a static camera. The overlap ratio is 0.14. (d) The score map of (c). (e) Tracking result and ground truth of the 4th frame of video Skiing [31], captured by a moving camera. The overlap ratio is 0.89. (f) The score map of (e). (g) Tracking result and ground truth of the 40th frame of the video Skiing [31], captured by a moving camera. The overlap ratio is 0.45. (h) The score map of (g).
Sensors 21 01682 g009
Figure 10. The likelihood probability distribution of a score map-based feature given high (blue)/non-high (grey) overlap ratio and clustering of training data. (a) Mean. (b) Standard deviation. (c) Fisher linear discriminant (FLD)-based feature (linear discriminant feature).
Figure 10. The likelihood probability distribution of a score map-based feature given high (blue)/non-high (grey) overlap ratio and clustering of training data. (a) Mean. (b) Standard deviation. (c) Fisher linear discriminant (FLD)-based feature (linear discriminant feature).
Sensors 21 01682 g010
Figure 11. A cluster of training samples with high overlap ratio (yellow dot) and that with non-high overlap ratio (purple dot) in the 3-D feature (linear discriminant feature, mean, and standard deviation) space.
Figure 11. A cluster of training samples with high overlap ratio (yellow dot) and that with non-high overlap ratio (purple dot) in the 3-D feature (linear discriminant feature, mean, and standard deviation) space.
Sensors 21 01682 g011
Figure 12. Tracking results of targets located in regions with significant geometric deformation of videos #4 and #8. (Blue: ground truth. Red: SiamFC + face stitching (S). Green: SiamFC + S + template update (P). Yellow: SiamFC + S + H) (a) SiamFC + S on video #4. (b) Tai et al. (SiamFC + S + H) on video #4. (c) Proposed scheme (SiamFC + S + P) on video #4. (d) SiamFC + S on video #8. (e) Tai et al. (SiamFC + S + H) on video #8. (f) Proposed scheme (SiamFC + S + P) on video #8.
Figure 12. Tracking results of targets located in regions with significant geometric deformation of videos #4 and #8. (Blue: ground truth. Red: SiamFC + face stitching (S). Green: SiamFC + S + template update (P). Yellow: SiamFC + S + H) (a) SiamFC + S on video #4. (b) Tai et al. (SiamFC + S + H) on video #4. (c) Proposed scheme (SiamFC + S + P) on video #4. (d) SiamFC + S on video #8. (e) Tai et al. (SiamFC + S + H) on video #8. (f) Proposed scheme (SiamFC + S + P) on video #8.
Sensors 21 01682 g012
Figure 13. Templates selected by the proposed scheme (video #7). (a) The 35th frame. (b) The 58th frame. (c) The 93rd frame. (d) The 177th frame. (e) The 187th frame.
Figure 13. Templates selected by the proposed scheme (video #7). (a) The 35th frame. (b) The 58th frame. (c) The 93rd frame. (d) The 177th frame. (e) The 187th frame.
Sensors 21 01682 g013
Figure 14. Templates selected by the proposed scheme (video #8). (a) The 10th frame. (b) The 25th frame. (c) The 38th frame. (d) The 166th frame. (e) The 180th frame.
Figure 14. Templates selected by the proposed scheme (video #8). (a) The 10th frame. (b) The 25th frame. (c) The 38th frame. (d) The 166th frame. (e) The 180th frame.
Sensors 21 01682 g014
Figure 15. Templates selected by the proposed scheme (video #2). (a) The 43rd frame. (b) The 131st frame. (c) The 186th frame. (d) The 194th frame.
Figure 15. Templates selected by the proposed scheme (video #2). (a) The 43rd frame. (b) The 131st frame. (c) The 186th frame. (d) The 194th frame.
Sensors 21 01682 g015
Figure 16. Templates selected by the proposed scheme (video #4). (a) The 22nd frame. (b) The 33rd frame. (c) The 48th frame. (d) The 93rd frame. (e) The 101st frame. (f) The 109th frame. (g) The 118th frame. (h) The 134th frame. (i) The 158th frame. (j) The 183rd frame.
Figure 16. Templates selected by the proposed scheme (video #4). (a) The 22nd frame. (b) The 33rd frame. (c) The 48th frame. (d) The 93rd frame. (e) The 101st frame. (f) The 109th frame. (g) The 118th frame. (h) The 134th frame. (i) The 158th frame. (j) The 183rd frame.
Sensors 21 01682 g016
Figure 17. Samples of the predicted bounding box in the face of non-uniform deformation and occlusion caused by padded pixels on video #2. (Blue: ground truth) (a) SiamFC + S, the 25th frame. (b) SiamFC + S + H (Tai et al.) [23], the 25th frame. (c) Proposed scheme, the 25th frame. (d) SiamFC + S, the 104th frame. (e) SiamFC + S + H (Tai et al.) [23], the 104th frame. (f) Proposed scheme, the 104th frame.
Figure 17. Samples of the predicted bounding box in the face of non-uniform deformation and occlusion caused by padded pixels on video #2. (Blue: ground truth) (a) SiamFC + S, the 25th frame. (b) SiamFC + S + H (Tai et al.) [23], the 25th frame. (c) Proposed scheme, the 25th frame. (d) SiamFC + S, the 104th frame. (e) SiamFC + S + H (Tai et al.) [23], the 104th frame. (f) Proposed scheme, the 104th frame.
Sensors 21 01682 g017aSensors 21 01682 g017b
Figure 18. Samples of predicted bounding box in the face of non-uniform deformation and occlusion caused by padded pixels on video #4. (Blue: ground truth) (a) SiamFC + S, the 96th frame. (b) SiamFC + S + H (Tai et al.) [23], the 96th frame. (c) Proposed scheme, the 96th frame. (d) SiamFC + S, the 142th frame. (e) SiamFC + S + H (Tai et al.) [23], the 142th frame. (f) Proposed scheme, the 142th frame.
Figure 18. Samples of predicted bounding box in the face of non-uniform deformation and occlusion caused by padded pixels on video #4. (Blue: ground truth) (a) SiamFC + S, the 96th frame. (b) SiamFC + S + H (Tai et al.) [23], the 96th frame. (c) Proposed scheme, the 96th frame. (d) SiamFC + S, the 142th frame. (e) SiamFC + S + H (Tai et al.) [23], the 142th frame. (f) Proposed scheme, the 142th frame.
Sensors 21 01682 g018
Figure 19. Comparisons between SiamFC + S +Timing detector of template update of LSSiam (proposed by Liang et al.) and the proposed scheme (SiamFC + S + P). Eight test videos are tested. (a) Success plots. (b) Precision plots.
Figure 19. Comparisons between SiamFC + S +Timing detector of template update of LSSiam (proposed by Liang et al.) and the proposed scheme (SiamFC + S + P). Eight test videos are tested. (a) Success plots. (b) Precision plots.
Sensors 21 01682 g019
Figure 20. Comparisons between SA-Siam with face stitching (SA-Siam + S) and SA-Siam with face stitching and the timing detector of template update (SA-Siam + S + P). Eight test videos with 1600 frames were tested. (a) Success plots. (b) Precision plots.
Figure 20. Comparisons between SA-Siam with face stitching (SA-Siam + S) and SA-Siam with face stitching and the timing detector of template update (SA-Siam + S + P). Eight test videos with 1600 frames were tested. (a) Success plots. (b) Precision plots.
Sensors 21 01682 g020
Figure 21. The percentage of computational time of individual components of the proposed scheme including face stitching, extraction of feature maps and calculation of the score map, prediction of the coordinate, and template update.
Figure 21. The percentage of computational time of individual components of the proposed scheme including face stitching, extraction of feature maps and calculation of the score map, prediction of the coordinate, and template update.
Sensors 21 01682 g021
Table 1. Summarization of described trackers in Section 1.
Table 1. Summarization of described trackers in Section 1.
ExamplesTemplate UpdateTiming Detector of Template UpdateYearRemark
Trackers for 360-Degree VideosLiu et al. 2018 [4]XX20181. ERP format
2. DeepSort-based [5]
Zhou et al. [6]XX20101. CMP format
2. Mean Shift-Based
The First Deep Learning-Based TrackerDLT [9]XX2013Stacked autoencoder-based
Siamese Networks Based TrackersGoTurn [11]XX2016Pioneer
SiamFC [12]XX2016Pioneer
SiamRPN [13]XX2018Region proposal network
SiamRPN++ [14]XX2019Effective sampling strategy
SA-Siam [16]XX2018Semantic branch and
similarity branch
adaDCF [17]XX2020FDA discriminates between foreground and background
CFNet [18]Aggressive strategyX2017Correlation filter-based
template generation
Yang et al. [19]Aggressive strategyX2018Attentional LSTM controls memory for template generation
Xu et al. [20]Non-aggressive strategyUsing highest score of a score map2019UAVs-Based Tracking
Wang et al. [21]Non-aggressive strategyAPCE of a score map2017
Liang et al. [15]Non-aggressive strategyUpdate interval and
APCE [21]
2020
Tai et al. [23]Non-aggressive strategyStatistics of score map2020EAC format
Table 2. Comparisons of Tracking Accuracy among SiamFC [12], SiamFC + S, and SiamFC + S + P (proposed scheme) (S: Face Stitching. P: Template Update).
Table 2. Comparisons of Tracking Accuracy among SiamFC [12], SiamFC + S, and SiamFC + S + P (proposed scheme) (S: Face Stitching. P: Template Update).
Overlap RatioLocation Error (Units: Pixels)
SiamFCSiamFC + SSiamFC + S + P (Proposed)SiamFCSiamFC + SSiamFC + S + P (Proposed)
Video#10.0460.4810.479106.34310.17110.152
Video#20.0650.5630.547118.24699589730
Video#30.1210.7280.68894.03322702714
Video#40.2470.5180.53284.88075076746
Video#50.6070.6480.653627659364956
Video#60.3030.3630.36625.02411.34110.763
Video#70.1000.2360.593316.71190.4869264
Video#80.1020.2350.518172.300140.25027.011
Average0.1990.4910.547115.47734.74010.167
Table 3. Samples of Tracking Results of SiamFC + S and SiamFC + S + P (proposed scheme) on Video #3 (London Park Ducks and Swans). The last frame index is 200. (S: Face Stitching. P: Template Update. Blue box: Ground Truth).
Table 3. Samples of Tracking Results of SiamFC + S and SiamFC + S + P (proposed scheme) on Video #3 (London Park Ducks and Swans). The last frame index is 200. (S: Face Stitching. P: Template Update. Blue box: Ground Truth).
t = 2t = 120t = 200
SiamFC + S Sensors 21 01682 i001 Sensors 21 01682 i002 Sensors 21 01682 i003
SiamFC + S + P
(Proposed Scheme)
Sensors 21 01682 i004 Sensors 21 01682 i005 Sensors 21 01682 i006
Table 4. Confusion matrix of the proposed timing detector.
Table 4. Confusion matrix of the proposed timing detector.
Ground Truth
TrueFalse
PredictionPositive775
Negative891421
Table 5. Samples of tracking results of SiamFC + S and SiamFC + S + P (proposed scheme) on video #6 (Paris-view 6). The last frame index is 200. (S: Face Stitching. P: Template Update. Blue box: Ground Truth.)
Table 5. Samples of tracking results of SiamFC + S and SiamFC + S + P (proposed scheme) on video #6 (Paris-view 6). The last frame index is 200. (S: Face Stitching. P: Template Update. Blue box: Ground Truth.)
t = 2t = 120t = 200
SiamFC + S Sensors 21 01682 i007 Sensors 21 01682 i008 Sensors 21 01682 i009
SiamFC + S+P
(Proposed Scheme)
Sensors 21 01682 i010 Sensors 21 01682 i011 Sensors 21 01682 i012
Table 6. Comparisons of tracking accuracy between SiamFC + S + H (Tai et al.) and SiamFC + S + P (proposed scheme). (S: Face Stitching. P: Template Update).
Table 6. Comparisons of tracking accuracy between SiamFC + S + H (Tai et al.) and SiamFC + S + P (proposed scheme). (S: Face Stitching. P: Template Update).
Overlap RatioLocation Error (Units: Pixels)
SiamFC + S + H
(Tai et al.) [23]
SiamFC + S + P
(Proposed)
SiamFC + S + H
(Tai et al.) [23]
SiamFC + S + P
(Proposed)
Video#10.4810.47910.16010.152
Video#20.4050.54713.2679730
Video#30.6600.68837182714
Video#40.5310.53277026746
Video#50.6920.65340434956
Video#60.3370.36611.92810.763
Video#70.5990.59357589264
Video#80.5720.51817.33027.011
Average0.5340.547923810.167
Table 7. Samples of tracking results of SiamFC + S + H (Tai et al.) and SiamFC + S + P (proposed Scheme 5. (Amsterdam, The Netherlands). The last frame index is 200. (S: Face Stitching. P: Template Update. Blue box: Ground Truth).
Table 7. Samples of tracking results of SiamFC + S + H (Tai et al.) and SiamFC + S + P (proposed Scheme 5. (Amsterdam, The Netherlands). The last frame index is 200. (S: Face Stitching. P: Template Update. Blue box: Ground Truth).
t = 2t = 85t = 200
SiamFC + S+H
(Tai et al.) [23]
Sensors 21 01682 i013 Sensors 21 01682 i014 Sensors 21 01682 i015
SiamFC + S+P
(Proposed Scheme)
Sensors 21 01682 i016 Sensors 21 01682 i017 Sensors 21 01682 i018
Table 8. Tracking speed of all considered trackers in Section 4. Before tracking, videos #1 and #5 are downsampled from 3432 × 2288 to 438 × 292, and videos #2, #3, #4, #6, and #8 are downsampled from 2800 × 1920 to 444 × 296, respectively. Video #7 with spatial resolution 1764 × 1176 is not downsampled. (S: Face stitching, H: Heuristics-based timing detector, P: Machine learning-based timing detector).
Table 8. Tracking speed of all considered trackers in Section 4. Before tracking, videos #1 and #5 are downsampled from 3432 × 2288 to 438 × 292, and videos #2, #3, #4, #6, and #8 are downsampled from 2800 × 1920 to 444 × 296, respectively. Video #7 with spatial resolution 1764 × 1176 is not downsampled. (S: Face stitching, H: Heuristics-based timing detector, P: Machine learning-based timing detector).
TrackerVideos #1–#6, #8Videos #7
SiamFC [12] +S62.35 fps23.42 fps
Tai et al. [23] (SiamFC + S + H)60 fps20 fps
Proposed Tracker (SiamFC + S + P)52.9 fps11.2 fps
SiamFC [12] +S +Timing Detector of Template Update of LSSiam [15]55.95 fps10.53 fps
SA-Siam [16] +S50.16 fps11.7 fps
SA-Siam [16] +S + P46.04 fps9.78 fps
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tai, K.-C.; Tang, C.-W. Siamese Networks-Based People Tracking Using Template Update for 360-Degree Videos Using EAC Format. Sensors 2021, 21, 1682. https://doi.org/10.3390/s21051682

AMA Style

Tai K-C, Tang C-W. Siamese Networks-Based People Tracking Using Template Update for 360-Degree Videos Using EAC Format. Sensors. 2021; 21(5):1682. https://doi.org/10.3390/s21051682

Chicago/Turabian Style

Tai, Kuan-Chen, and Chih-Wei Tang. 2021. "Siamese Networks-Based People Tracking Using Template Update for 360-Degree Videos Using EAC Format" Sensors 21, no. 5: 1682. https://doi.org/10.3390/s21051682

APA Style

Tai, K. -C., & Tang, C. -W. (2021). Siamese Networks-Based People Tracking Using Template Update for 360-Degree Videos Using EAC Format. Sensors, 21(5), 1682. https://doi.org/10.3390/s21051682

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop