Next Article in Journal
Design and Development of a Near Isotropic Printed Arc Antenna for Direction of Arrival (DoA) Applications
Next Article in Special Issue
UAV Object Tracking Application Based on Patch Color Group Feature on Embedded System
Previous Article in Journal
An Improved Approach for Object Proposals Generation
Previous Article in Special Issue
Backhaul-Aware Resource Allocation and Optimum Placement for UAV-Assisted Wireless Communication Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Robust Quadruplet and Faster Region-Based CNN for UAV Video-Based Multiple Object Tracking in Crowded Environment

Center for Intelligent Bionic, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
*
Author to whom correspondence should be addressed.
Electronics 2021, 10(7), 795; https://doi.org/10.3390/electronics10070795
Submission received: 11 January 2021 / Revised: 3 March 2021 / Accepted: 16 March 2021 / Published: 27 March 2021
(This article belongs to the Special Issue UAV-Femtocell Systems and Applications)

Abstract

:
Multiple object tracking (MOT) from unmanned aerial vehicle (UAV) videos has faced several challenges such as motion capture and appearance, clustering, object variation, high altitudes, and abrupt motion. Consequently, the volume of objects captured by the UAV is usually quite small, and the target object appearance information is not always reliable. To solve these issues, a new technique is presented to track objects based on a deep learning technique that attains state-of-the-art performance on standard datasets, such as Stanford Drone and Unmanned Aerial Vehicle Benchmark: Object Detection and Tracking (UAVDT) datasets. The proposed faster RCNN (region-based convolutional neural network) framework was enhanced by integrating a series of activities, including the proper calibration of key parameters, multi-scale training, hard negative mining, and feature collection to improve the region-based CNN baseline. Furthermore, a deep quadruplet network (DQN) was applied to track the movement of the captured objects from the crowded environment, and it was modelled to utilize new quadruplet loss function in order to study the feature space. A deep 6 Rectified linear units (ReLU) convolution was used in the faster RCNN to mine spatial–spectral features. The experimental results on the standard datasets demonstrated a high performance accuracy. Thus, the proposed method can be used to detect multiple objects and track their trajectories with a high accuracy.

1. Introduction

Object tracking [1] is a significant task in computer vision applications based on unmanned aerial vehicles (UAVs). Compared to single object tracking (SOT), the task of multiple object tracking (MOT) has to develop the trajectories of all the objects in a precise scene of video surveillance [1,2]. Online MOT two-dimensional space is a complex task when there are similar objects [3]. In computer vision, the MOT task mainly includes action recognition [4], behavior analysis [5], and pose estimation [6]. With the enhancement of object detection methods such as the single-shot detector (SSD), the faster region-based convolutional neural network (RCNN), and the deformable part model (DPM), tracking by a prediction framework possesses high performances for MOT because prediction can provide object location and object trajectories. Assignment could be inferred by various problems such as context [7], shape [8], motion and appearance [9], out-of-plane rotations [10], and background clutter [11]. The above problems are more difficult in MOT than in SOT [11].
In recent years, UAVs have been increasingly used in surveillance and traffic monitoring due to their ability to cover great distances while being remotely controlled to travel at a predefined speed [12], combined with the computer vision technology in security monitoring [13,14]. Meanwhile, UAV videos are different than conventional videos. Generally, objects are small and difficult to differentiate due to their appearance. In such a situation, motion is an important piece of data for linking objects. However, due to high movement and altitude of UAVs in videos, the motion of an object is difficult to discriminate, thus resulting in high challenge for MOT, and the motion of UAVs causes image instability and platform motion. In this study, object refers to a car or human. In UAVs, the video motion of an object can be divided into two types of motion models: object motion and view modifications or the appearance feature of the UAV.
Many methods have been proposed to solve these problems, including a simple tracking-by-detection method such as the intersection over union (IoU) tracker that can only achieve a good result when objects are in good viewpoints [15]. A deep association metric that considers the deep appearance feature and information motion of an object while matching was proposed in [16]. The authors of [17] handled detection issues and improved the appearance feature through the collection of samples from an output model. Based on deep learning, image processing algorithms have made it easier to achieve auto-navigation in commercial UAVs. Additionally, deep learning researchers have developed fast and end-to-end trackers such as Cascade RCNN [18], FlowNet [19], DaSiameseRPN (region proposal network) [20], simple online and real-time tracking (SORT) [21] and DeepSORT [22] for MOT. However, all these trackers have at least one limitation, especially in the larger detection of pedestrian or car and at a higher frame rate per seconds (FPS).
To predict the motion of an individual object from the crowded objects, a faster RCNN was designed in this study with numerous important features such as multi-scale training, negative mining, and concatenation to improve the region-based CNN baseline. Similarly for the tracking motion of an individual object, a deep quadruplet network (DQN) was introduced to track the predicted objects from the crowd of objects, and the accuracy of tracking can be improved by using a quadruplet loss function and a deep CNN. We examined our tracking model with the existing algorithms based on motion prediction with certain indexes such as the IDF1-identity F1 score, as well as datasets such as the Unmanned Aerial Vehicle Benchmark: Object Detection and Tracking (UAVDT) and Stanford Drone datasets for the state-of-the-art comparisons. The major contributions of this paper are summarized as:
  • A quadruplet network is proposed by employing a faster region-based convolutional neural network with novel generated image G-16 (VGG-16) network to train the fine-tuned data for the prediction of individual motion from a crowded object.
  • A newly optimized background learning method that uses boundary-based quadruplet with deep networks is introduced to track prediction objects from crowded environments.
  • We applied the proposed quadruplet method in some of the recent state-of-the-art comparison benchmarks, such as the UAVDT and Stanford Drone datasets. The re results shows outstanding performance.

2. Related Work

Based on the architecture of deep learning, many methods have been proposed to solve MOT problems. In [23], a novel DQN was introduced to analyze the potential connections between training samples, with the intention to attain a powerful representation. This was modelled by a shared network with four branches that acquired numerous sample, tuples as inputs which were linked by new loss function containing triplet loss and pair loss. Related to similar metrics, similar and dissimilar samples were selected as the positive and negative inputs, respectively, of triplet loss from each tuple. Additionally, a weight layer was presented to automatically choose appropriate combinations of weight in order to eliminate conflict among pair loss and triplet. The authors of [24] connected some benefits of deep learning with data associated with tracking. They introduced a DAN (deep affinity network) to study the compact features of pre-predicted objects at various stages of abstraction and conducted comprehensive pairing permutations of those features in any two frames to gather object affinities. The DAN accounted for numerous objects disappearing and appearing among video frames in MOT.
A unite re-identification (RE-ID) model and detector within an end-to-end network was proposed in [25] by adding an extra track branch for monitoring a faster RCNN architecture. This network could train the entire model with multi-loss. The RE-ID design in deep simple online real-time tracking allowed for the mining of the feature map from the predicted object images and the region of interest (RoI) feature vector in the faster RCNN baseline to minimize calculations. The triplet loss function was utilized to optimize and track the branch, while the neighboring frame was concatenated in video to build the dataset. The authors of [26] proposed deep quadruplet appearance learning (DQAL), where the quadruplet network was considered as an input and each quadruplet consisted of anchor, negative, positive, and similar vehicles with different IDs. Then a quadruplet network with softmax loss and quadruplet loss was created to study more discriminative features for vehicle Re-ID. Deep learning was utilized to resolve the issue of pedestrian prediction from drone images [27]; 1500 images were collected by an SJRC S30W drone at various places with varying weather conditions, and different time spans were applied for pedestrian detection with high performance.
The authors of [28] presented a UAV dataset that addressed numerous challenges such as large camera motion, small object, high density, and complex scenes to predict and monitor them. These challenges inspired researchers to state benchmarks for the three fundamental tasks of MOT, SOT, and object prediction on a UAV dataset. Additionally, a new technique based on CMSN (context-aware multi-task Siamese network) was introduced to resolve issues in UAV videos by interceding for the degree of consistency among contexts and objects that were utilized for MOT and SOT. In [29], a feature selection based on new intuitionistic fuzzy clustering technique was developed for MOT. The adaptive selection from the visual objects was comprised of several object features using the neighborhood rough set that could measure the distance similarity among observations and objects, proving to be efficient in robust environment. The authors of [30] suggested a spatial–temporal scheme to study MOT in different scenarios. The location and insertion of candidates were addressed by a novel motion model and semantic-feature spatial attention mechanism. Then, the target occlusion estimated by the online-learned CNN of specific target and appearance model were classified by adaptation in [31]. The authors proposed a target re-detection model over the region that could adaptively learn to perform and handle the existing scale estimation challenge. The noisy feature and channel redundancy representation in convolutional features were overcome via the correlation filter learning of channel regularization.
The target detection in a video sequence is the most essential stage in aerial surveillance for subsequent processing, while the performance of surveillance system depends on the detection of motion of any object in video. Therefore, understanding the moving patterns of objects and persons is required to uncover suspicious events. In the literature, various techniques for moving object tracking and detection in video have been proposed. The anomaly detection methods can be divided into two types: machine learning-based and pattern recognition-based methods. The disadvantages of existing methods are that they have a minimum accuracy, are time-consuming, and have difficulties in the processing of noisy data.

3. Materials and Methods

The proposed methodology was based on a faster RCNN of a deep learning framework, with its structure summary divided into two parts: RoI represents the position of the objects in the frame, and the fast RCNN network categorizes the region of image to objects and filters the boundaries of the region. These two parts share the general parameter of a convolutional layer that is utilized for feature extraction and for permitting the architecture to include object prediction tasks at a competitive speed. Here, a faster RCNN was designed to predict objects with a high accuracy and recall rate. The trained proposed methodology is depicted in Figure 1.
Primarily, a faster RCNN was trained with the Stanford Drone and UAVDT datasets. This dataset was further used to test the pre-trained model to enable it find the hard negatives and use them as the given input to the network in the 2nd step of the training process. By training with this kind of hard negative sample, the model developed fewer false positives. Moreover, this model was fine-tuned on the dataset. During the fine-tuning, a multi-scale training process was used and implemented for feature concatenation to boost the performance of the proposed model. For the training process, we utilized a faster RCNN to enhance the performance of object prediction. Finally, the resulting predicted bounding boxes were converted to ellipses as regions of objects. Below is the summary of our methodology.

3.1. Feature Concatenation

Normally, in a conventional fast RCNN network, RoI pooling is done in the final feature map layer to develop the features that are examined by the classification part of the network. The inspired model made the classification process to use features that were measured from the risk priority number and also preserved unwanted calculations. However, such a technique is not optimal at all times, and, in some cases, it could eliminate certain significant features. Thus, in the proposed work, to capture a finely grained parts of the RoI, it was required to enhance the RoI pooling using the integrated feature maps of numerous convolutional layers with high and low level features.
Therefore, we concatenated the pooling result of numerous convolutional feature maps to develop the final pooling features for an accurate prediction tasks. This means that we used few intermediate results with a risk priority number final feature map by integrating them to provide the pooling features. In particular, the features that were from many lower-level convolution layers were mainly L2-normalized and RoI-pooled. Then the resultant features were combined to obtain the matching feature scale. Afterwards, a 1 × 1 convolution was used to match the channels of the original network.

3.2. Hard Negative Mining

Hard negative mining is an efficient technique to boost deep learning performance, mainly for object prediction [32]. The concept of this technique implies that a hard negative is the region where a network fails to make right prediction. Therefore, a hard negatives are given as inputs to the network as reinforcements to enhance the trained model with a minimized number of false positives and the best classification performance.
In our design, hard negatives were gathered from the pre-trained model of the training process. Then, a region was considered as a hard negative if its IOU (insertion over union) and over a region ground truth was less than 0.5. During the training process, the hard negatives were explicitly added to the RoI to tune the model and balance the ratio of the background and foreground to 1:3.

3.3. Model Pre-Training

To adjust the faster RCNN for object prediction, we chose to fine-tune the pre-trained model from the ImageNet dataset. Our model was pre-trained on the datasets with numerous complex instances, which might have affected the convergence of the training process. To overcome these difficulties, certain training data had to be removed. Additionally, for a pre-training on this dataset, hard negative mining was a vital way to minimize false positives.

3.4. Multi-Scale Training

A faster RCNN was used for a fixed scale in the whole training of images. Then, the image was resized to a random scale, and a predictor is utilized to study the features over a broad range of sizes, thus enhancing its performance of scale invariance. Here, among the three scales, one scale was assigned to an image before it was inputted to network.

3.5. Number of Anchors

The various key hyper-parameters in the faster RCNN had to be turned, and the number of anchors was crucial; each was found within the region proposal network, where the conventional faster RCNN utilized nine anchors, which made it fail to recall smaller objects. For object prediction, small objects are the most common, mainly for unconstrained prediction. Hence, instead of utilizing the default setting, a size of 64 ∗ 64 was added to increase the number of anchors to 12.

3.6. Boundary-Based Quadruplet Network

The DQN is trained to study the feature space. Testing data were moved from learning feature space to mine features. Classification was achieved with the help of the CNN classifier: Є ( w ) = 1 2 y i = 1 I B   x i c i 22 i = 1 I pmci22, where pm denotes penalization matrix and c i represents the dth channel of the correlation filter.

3.7. Deep Network

The deep CNN block contained 6 convolutional layers. “Conv” is the convolutional operation with 3 × 3 × 3 kernels, while “2 Conv” the denotes convolutional layer with 2 kernels. Normally, a CNN block of six layers consists of 6 connections. However, as the network becomes deeper, issues with the normal CNN arise, i.e., the features in the input vanish once it travels through several layers until it comes to the end. Thus, we preferred to use a deep CNN with 6 layer connections, where 3 connections were between the 1st, 3rd, 4th, 5th, and 6th layers, and 2 connections were between the 2nd and 6th layers. Figure 2 illustrates the connection points in the dense CNN, and ⨁ indicates the sum of all imported connected lines. The dense layer resolved issues that were related to information vanishing while passing over multiple layers, and it also utilized the features mined by the entire layers.

3.8. Quadruplet Learning

Metric learning means to transfer the input data to new feature space S(C) (e_θ: S(E)→ S(C) from the original space, where E denotes the dimension of original the space, C denotes the dimension the of new space, and θ is the learning parameter. In the new feature space S(C), the samples from the same class were assumed to be nearer than those from the diverse classes so that the classification could be completed in S(C) with the NN. A network such as a quadruplet network [33] could be created.

3.9. Quadruplet Loss (QL)

Here, the quadruplet loss (QL) included 4 various samples: y p   ( i ) , y q   ( i ) , y n 1   ( i ) , and y n 2   ( i ) , where y p   ( i )   and y q   ( i ) are samples from the same class and y n 1   ( i )   and y n 2   ( i ) are sample of another class. All samples were moved to the feature space by e θ : S E   S C . QL is given by
LF quad   =   1 NUM quad i = 1 NUM quad ( e d ( y p ( i ) ,   y q ( i ) ) e d ( y p ( i ) ,   y n 1 ( i ) )   +   Є   ) + ( e d ( y p ( i ) ,   y q ( i ) ) e d ( y n 1 ( i ) ,   y n 2 ( i ) ) +   Ω   ) + )
where   Ω and Є are margins of 2 terms, and NUM quad denotes the number of quadruplets. The 2nd term of the QL limits the intra-class distance and causes it to be smaller than the inter-class distances. However, the loss function in Equation (1) performed poorly since the quadruplet pairs and the number of quadruplet grew faster while the dataset got broader. In certain cases, all the samples were not sufficient to train the network, which led to poor performance [33]. To overcome all the issue, a new QL function was designed:
LF nquad   =   1 NUM nquad i = 1 NUM nquad ( e d ( y p ( i ) ,   y q ( i ) ) e d ( y m ( i ) ,   y n ( i ) )   + Є ) +
where y q   ( i ) denotes the farthest sample to the reference y p   ( i ) in the same class, y m   ( i )   and   y n   ( i ) are the nearby negative pairs in the entire batch, NUM nquad is the number of quadruplet in the new loss function, and Є is the value of the margin. Each sample in Equation (2) was moved by e θ : S E   S C . The pseudocode for the batch training with the proposed loss function is expressed in Algorithm 1, where TD = {td1, td2, td3,…, tds} is the training dataset for this batch and s is the number of the labelled samples. As said in the algorithm, NUM nquad = r and t d j , t d k indicate the sample in the dataset TD. Meanwhile, t d j and t d k indicate the pair of samples, C(tdj) is a class of label of sample t d j , and represents the learning rate. tdp, tdq, tdm, and tdn are the quadruplets before the deep network, while yp, yq, ym, and yn are the corresponding quadruplets after the deep network.
Algorithm 1. Quadruplet Network
  • Input: The training dataset for this batch TD = {td1, td2, td3,…, tds}
  • The initialized or updated learnable parameter θ.
  • For all pairs of samples ( t d j , t d k ) in TD, DO
  • Calculate the Euclidean distance ed ( f θ ( t d j ), f θ ( t d k ))
  • End For
  • Set the loss LFquad = 0
  • Set (m, n) = argmined(j,k) (fθ(tdj), fθ(tdk)), under the condition C(tdj) ≠ C(tdk)
  • ym = fθ (tdm), yn = fθ (tdn). (Transfer tdm, tdn to ym, yn by the network fθ: EG EH)
  • For tdp in the dataset TD, DO
  • yp = fθ (tdp). (Transfer tdp to yp by the network fθ: EG EH)
  • Set p = argmaxed(j) (fθ(tdj), yp), under the condition CL(tdj) = C(tdp)
  • yq = fθ (tdp). (Transfer tdp to yq by the network fθ: EG EH)
  • Update: LFquad = LFquad + 1 NUM nquad (d(yp, yq) − d(ym, yn) + Є )+
  • End For
  • Update: θ = θ − Δ θ LFnquad
  • Output: θ, LFnquad

4. Results

The proposed tracking technique quadruplet-based faster RCNN was tested with existing algorithms, such as Bayesian multi-object tracking (BMOT) [34], intersection-over-union tracker (IOUT) [15], global optimal greedy (GOG) algorithm [35], continuous energy minimization (CEM) [36], social long short term memory (SLSTM) [37], simple online and real-time tracking (SORT) [21], relative long short term memory (RLSTM) [38], and relative motion online tracking (RMOT) [39]. To examine the performance of the MOT techniques, we utilized numerous metrics such as identification precision (IDP), identification recall (IDR), and F1 score, which are together referred as IDF1. IDF1 is stated as
IDF 1   =   2 IDTP 2 IDPP + IDP + IDN
where IDPP represents the ID of the true positives, IDP represents the ID of the false positive, and IDN represents the ID of the false negative. Additionally, the accuracy, precision of MOT, mostly tracked target, mostly lost target, count of false positives, count of false negatives, count of ID switches, and count of time trajectory are fragmented.

4.1. Datasets

The proposed method was executed on the public Stanford Drone and UAVDT datasets [7,33]. The UAVDT dataset provided a sequence of 50 records of the traffic state from UAVs. There are four types of prediction results for MOT, faster RCNN [40], reverse connection with objectiveness prior network (RON) [41], Single Shot Multi Box Detector (SSD) [42], and region-based fully convolutional network (R-FCN) [42]. The Stanford Drone dataset consists of sequences of 60 videos that were recorded from a high altitude looking down by a flying UAV on a campus scene. The gathered videos consisted of several kinds of objects, as seen in Figure 3.

4.2. Evaluation

We utilized four types of object prediction results as inputs to the quadruplet-based faster RCNN on the UAVDT datasets. It was noticed that our quadruplet-based faster RCNN attained good results for the basic metric of IDP and IDF1, while SORT attained an IDF1 of 37.1 and SDD achieved an IDF1 of 54.6. Compared to other prediction technique, our proposed technique had great improvements on most of the metrics and could also handle most difficult trajectories. This means that with the proposed approach, the prediction of motion of object trajectories was stronger. The highest value of the IDF1 score illustrates that the proposed technique was more effective.
The UAVDT dataset consists of four types of attributes (camera view, weather condition, flying altitude, and time attribute such as duration). The proposed approach was examined for those four attributes. Regarding the camera view attribute, our technique performed very well for the inside, front, and bird views, as the images captured from all these views had more object information that helped in the object prediction and tracking. Our proposed technique attained good result because the scene information of the side and front views was needed for global motion analysis. Regarding the weather condition attribute, most of the techniques performed well in night and day but failed to perform well in fog, as the view was altered in the night and day. However, our proposed approach did well in all conditions. Regarding the flying altitude attribute, most other MOT technique were reduced with increasing altitude. However, our proposed approach performed very well at the medium and low altitudes, where there were drastic changes in view due to UAV movements, which caused wrong predictions for the other techniques. Regarding the duration attribute, the proposed approach performance is stable for short and long term with robustness of the proposed method.
We also executed our proposed approach on the Stanford Drone dataset. As it was difficult to predict objects in this type of dataset, the ground truth position of each object was used for MOT to examine the performance of MOT techniques regarding the position of real objects. It was noticed that our proposed approach outperformed the other MOT techniques. This was actualized by comparing all the positive and negative indexes for each prediction, as seen below.

4.2.1. Faster Region-Convolutional Network (Faster RCNN)

Table 1 shows the detection input of the faster RCNN evaluation with respect to all the performance metrics. Figure 4 represents the of performance comparison measures of IDF1, IDP, IDR and multiple object tracking precision (MOTP). The proposed method contained 55% of IDF1, 67% of IDP, 43% of IDR, 39% of multiple object tracking accuracy (MOTA), and 75% of MOTP. These measures were compared with the existing methods of SMOT, IOUT, GOG, CEM, SLSTM, SORT, RLSTM, RMOT, and the previously existing faster RCNN. This statistical analysis proved that the proposed method had higher efficiencies than the existing methods.
Figure 5 shows the proposed method’s results in regards to the performance measures of FP, FN, IDS and FM. The proposed method contained 3000 FP, 13,000 FN, 2 IDR, and 5 FM values. These measures were compared with the existing methods of BMOT, IOUT, GOG, CEM, SLSTM, SORT, RLSTM, and RMOT in the faster RCNN. This statistical analysis proved that the proposed method of the minimum false rate had higher efficiencies compared to the existing methods.

4.2.2. The Faster Region-Based Fully Convolutional Network (Faster R-FCN)

Table 2 shows the detection input of the faster R-FCN with respect to all the performance metrics. Figure 6 shows the performance comparison measures of IDF1, IDP, IDR, and MOTP. The proposed method contained 50% IDF1, 64% IDP, 44% IDR, 40% MOTA, and 80% MOTP values. These measures were compared with the existing methods of SMOT, IOUT, GOG, CEM, SLSTM, SORT, RLSTM, and RMOT in the existing faster R-FCN. This statistical analysis proved that the proposed method had a higher accuracy and precision rate compared to the existing methods.
Figure 7 shows the proposed method’s results regarding the performance measures of FP, FN, IDS, and FM. The proposed method contained 42,065 FP, 13, 5617 FN, 1191 IDR, and 2057 FM values. These measures were compared with the existing methods of SMOT, IOUT, GOG, CEM, SLSTM, SORT, RLSTM, and RMOT in the existing faster SSD. This statistical analysis proved that the proposed method had a higher accuracy and precision rate than the existing methods.

4.2.3. Single Shot Multi Box Detector (SSD)

Table 3 shows the detection input of SSD evaluation with respect to all the performance metrics.
Figure 8 shows the proposed method’s results regarding the performance measures of IDF1, IDP, IDR, and MOTP. The proposed method contained 45% IDF1, 51% IDP, 42% IDR, 36% MOTA, and 79% MOTP values.
These measures were compared with the existing methods of SMOT, IOUT, GOG, CEM, SLSTM, SORT, RLSTM, and RMOT in the existing faster SSD. This statistical analysis proved that the proposed method had a higher accuracy and precision rate than the existing methods.
Figure 9 shows the proposed method’s results regarding the performance measures of FP, FN, IDS, and FM. The proposed method contained 54,065 FP, 12, 9617 FN, 1267 IDR, and 1890 FM values. These measures were compared with the existing methods of SMOT, IOUT, GOG, CEM, SLSTM, SORT, RLSTM, and RMOT in the existing SSD. This statistical analysis proved that the proposed method of the minimum false rate had higher efficiencies than the existing methods.

4.2.4. Reverse Connection with Objectiveness Prior Network (RON)

Table 4 shows the detection input of RON evaluation with respect to all the performance metrics.
Figure 10 shows the proposed method’s results regarding the performance measures of IDF1, IDP, IDR, and MOTP. The proposed method contained 58% IDF1, 67% IDP, 47% IDR, and 39% MOTA values. These measures were then compared with the existing methods of GOG, CEM, SLSTM, SORT, RLSTM, and RMOT in the previously existing faster RON.
Figure 11 shows the proposed method’s results regarding the performance measures of FP, FN, IDS, and FM. The proposed method contained 52,065 FP, 150,917 FN, 1476 IDR, and 2539 FM values. These measures were compared with the existing methods of SMOT, IOUT, GOG, CEM, SLSTM, SORT, RLSTM, and RMOT in RON. This statistical analysis proved that the proposed method of the minimum false rate had higher efficiencies than the existing methods. We also evaluated our tracker with one-pass evaluation (OPE), and our tracker was found to attain the best expected average overlap (EAO) while maintaining beyond real-time speed.

5. Conclusions

For MOT, motion is generally utilized as a cue for linking objects and predictions. However, the conventional MOT technique addresses a performance drop in UAV videos due to UAVs’ autonomous motion. Therefore, we proposed a new quadruplet-based RCNN framework to find every objects’ motion and view modifications of UAVs. Experiments were conducted on two public datasets (Stanford Drone and UAVDT), and the results show the efficiency of the proposed approach. Our future research work will include a combination of other modules such as a triplet network for abnormal event detection and tracking.

Author Contributions

Conceptualization, H.U.D.; methodology, H.U.D.; software, H.U.D.; val-idation, H.U.D. and Y.Z.; formal analysis, H.U.D. and Y.Z.; investigation, H.U.D. and Y.Z.; re-sources, H.U.D. and Y.Z.; data curation, H.U.D.; writing—original draft preparation, H.U.D.; writing—review and editing, H.U.D. and Y.Z.; visualization, H.U.D. and Y.Z.; supervision, Y.Z.; project administration, Y.Z.; funding acquisition, Y.Z. Both authors have read and agreed to the published version of the manuscript.

Funding

This work was supported under the National Key Research and Development Program of China (2018YFB1305505), National Natural Science Foundation of China (NSFC) (61973296), Shenzhen Basic Research Program Refs. JCYJ20180507182619669 and Science and Technology Planning Project of Guangdong Province, Ref. 2017B010117009.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhong, B.; Bai, B.; Li, J.; Zhang, Y.; Fu, Y. Hierarchical Tracking by Reinforcement Learning-Based Searching and Coarse-to-Fine Verifying. IEEE Trans. Image Process. 2019, 28, 2331–2341. [Google Scholar] [CrossRef]
  2. Wang, X. Intelligent multi-camera video surveillance: A review. Pattern Recognit. Lett. 2013, 34, 3–19. [Google Scholar] [CrossRef]
  3. Leal-Taixe, L.; Canton-Ferrer, C.; Schindler, K. Learning by Tracking: Siamese CNN for Robust Target Association. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), IEEE, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 418–425. [Google Scholar]
  4. Milan, A.; Leal-Taixé, L.; Reid, I.; Roth, S.; Schindler, K. MOT16: A benchmark for multi-object tracking. Comput. Vis. Pattern Recognit. 2016, arXiv:1603.00831. [Google Scholar]
  5. Milan, A.; Rezatofighi, S.H.; Dick, A.; Reid, I.; Schindler, K. Online multi-target tracking using recurrent neural networks. AAAI Tech. Track Vis. 2016, arXiv:1604.03635. [Google Scholar]
  6. Zhou, Q.; Zhong, B.; Zhang, Y.; Li, J.; Fu, Y. Deep Alignment Network Based Multi-Person Tracking with Occlusion and Motion Reasoning. IEEE Trans. Multimed. 2019, 21, 1183–1194. [Google Scholar] [CrossRef]
  7. Yu, H.; Qin, L.; Huang, Q.; Yao, H. Online multiple object tracking via exchanging object context. Neurocomputing 2018, 292, 28–37. [Google Scholar] [CrossRef]
  8. Luo, W.; Xing, J.; Milan, A.; Zhang, X.; Liu, W.; Kim, T.-K. Multiple object tracking: A literature review. Artif. Intell. 2021, 293, 103448. [Google Scholar] [CrossRef]
  9. Comaniciu, D.; Ramesh, V.; Meer, P. Kernel-based object tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 564–577. [Google Scholar] [CrossRef] [Green Version]
  10. Hu, W.; Tan, T.; Wang, L.; Maybank, S. A Survey on Visual Surveillance of Object Motion and Behaviors. IEEE Trans. Syst. ManCybern. Part. C Appl. Rev. 2004, 34, 334–352. [Google Scholar] [CrossRef]
  11. Choi, W.; Savarese, S. A Unified Framework for Multi-target Tracking and Collective Activity Recognition. In Proceedings of the Lecture Notes in Computer Science, Florence, Italy, 7–13 October 2012; pp. 215–230. [Google Scholar]
  12. Hong, I.; Kuby, M.; Murray, A. A deviation ow refueling location model for continuous space: A commercial drone de-livery system for urban areas. In Proceedings of the 13th International Conference on Advances in Geocomputation, Geocomputation, Dallas, TX, USA, 20–23 May 2017; pp. 125–132. [Google Scholar]
  13. Paul, M.; Haque, S.M.E.; Chakraborty, S. Human detection in surveillance videos and its applications-a review. Eurasip J. Adv. Signal. Process. 2013, 2013, 176. [Google Scholar] [CrossRef]
  14. Lee, K.S.; Ovinis, M.; Nagarajan, T. Autonomous patrol and surveillance system using unmanned aerial vehicles. In Proceedings of the International Conference on Environment and Electrical Engineering (EEEIC), Rome, Italy, 10–13 June 2015; pp. 1291–1297. [Google Scholar]
  15. Bochinski, E.; Eiselein., V.; Sikora., T. High-speed tracking-by-detection without using image information. In Proceedings of the 14th IEEE International Conference on Advanced Video and Signal Based Surveillance (ICAVSS), Lecce, Italy, 29 August–1 September 2017; pp. 1–6. [Google Scholar] [CrossRef] [Green Version]
  16. Wojke, N.; Bewley, A.; Paulus, D. Simple online and realtime tracking with a deep association metric. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2017; pp. 3645–3649. [Google Scholar]
  17. Chen, L.; Ai, H.; Zhuang, Z.; Shang, C. Real-time multiple people tracking with deeply learned candidate selection and person re-identification. In Proceedings of the 2018 IEEE International Conference on Multimedia and Expo (ICME), San Diego, CA, USA, 23–27 July 2018; Volume 5, p. 8. [Google Scholar]
  18. Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6154–6162. [Google Scholar]
  19. Sun, D.; Yang, X.; Liu, M.Y.; Kautz, J. Pwc-net: Cnns for optical flow using pyramid, warping, and cost volume. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8934–8943. [Google Scholar]
  20. Zhu, Z.; Wang, Q.; Li, B.Q.; Wu, W.; Yan, J.; Hu, W. Distractor-Aware Siamese Networks for Visual Object Tracking. In Proceedings of the ECCV, Munich, Germany, 8–14 September 2018; pp. 101–117. [Google Scholar]
  21. Bewley, A.; Zongyuan, G.; Ott, L.; Ramos, F.; Upcroft, B. Simple online and realtime tracking. In Proceedings of the IIEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 3464–3468. [Google Scholar]
  22. Son, J.; Baek, M.; Cho, M.; Han, B. Multi-object Tracking with Quadruplet Convolutional Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3786–3795. [Google Scholar]
  23. Dong, X.; Shen, J.; Wu, D.; Guo, K.; Jin, X.; Porikli, F. Quadruplet Network with One-Shot Learning for Fast Visual Object Tracking. IEEE Trans. Image Process. 2019, 28, 3516–3527. [Google Scholar] [CrossRef] [Green Version]
  24. Sun, S.; Akhtar, N.; Song, H.; Mian, A.S.; Shah, M. Deep Affinity Network for Multiple Object Tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 1. [Google Scholar] [CrossRef] [Green Version]
  25. Xu, Y.; Wang, J. A unified neural network for object detection, multiple object tracking and vehicle re-identification. Comput. Vis. Pattern Recognit. 2019, arXiv:1907.03465. [Google Scholar]
  26. Hou, J.; Zeng, H.; Zhu, J.; Hou, J.; Chen, J.; Ma, K.-K. Deep quadruplet appearance learning for vehicle re-identification. IEEE Trans. Vehi. Tech. 2019, 1, 1–9. [Google Scholar] [CrossRef]
  27. Hung, G.L.; Sahimi, M.S.B.; Samma, H.; Almohamad, T.A.; Lahasan, B. Faster R-CNN Deep Learning Model for Pe-destrian Detection from Drone Images. SN Comp. Sci. 2020, 1, 1–9. [Google Scholar]
  28. Yu, H.; Li, G.; Zhang, W.; Huang, Q.; Du, D.; Tian, Q.; Sebe, N. The Unmanned Aerial Vehicle Benchmark: Object Detection, Tracking and Baseline. Int. J. Comput. Vis. 2019, 128, 1141–1159. [Google Scholar] [CrossRef]
  29. Li, L.-Q.; Wang, X.-L.; Liu, Z.-X.; Xie, W.-X. A Novel Intuitionistic Fuzzy Clustering Algorithm Based on Feature Selection for Multiple Object Tracking. Int. J. Fuzzy Syst. 2019, 21, 1613–1628. [Google Scholar] [CrossRef]
  30. Meng, F.; Wang, X.; Wang, D.; Shao, F.; Fu, L. Spatial–Semantic and Temporal Attention Mechanism-Based Online Multi-Object Tracking. Sensors 2020, 20, 1653. [Google Scholar] [CrossRef] [Green Version]
  31. Lu, X.; Ma, C.; Ni, B.; Yang, X. Adaptive Region Proposal with Channel Regularization for Robust Object Tracking. IEEE Trans. Circuits Syst. Video Technol. 2019, 1. [Google Scholar] [CrossRef]
  32. Wan, S.; Chen, Z.; Zhang, T.; Zhang, B.; Wong, K.-K. Bootstrapping face detection with hard negative examples. Comput. Vis. Pattern Recognit. 2016, arXiv:1608.02236. [Google Scholar]
  33. Xiao, Q.; Luo, H.; Zhang, C. Margin sample mining loss: A deep learning based method for person re-identification. Comput. Vis. Pattern Recognit. 2017, arXiv:1710.00478. [Google Scholar]
  34. Dicle, C.; Camps, O.I.; Sznaier, M. The Way They Move: Tracking Multiple Targets with Similar Appearance. In Proceedings of the 2013 IEEE International Conference on Computer Vision; Institute of Electrical and Electronics Engineers (IEEE), Sydney, Australia, 1–8 December 2013; pp. 2304–2311. [Google Scholar] [CrossRef] [Green Version]
  35. Pirsiavash, P.; Ramanan, D.; Fowlkes, C.C. Globally-optimal greedy algorithms for tracking a variable number of objects. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Colorado Springs, CO, USA, 20–25 June 2011; pp. 1201–1208. [Google Scholar]
  36. Milan, A.; Roth, S.; Schindler, K. Continuous Energy Minimization for Multitarget Tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 58–72. [Google Scholar] [CrossRef] [PubMed]
  37. Alahi, A.; Goel, K.; Ramanathan, V.; Robicquet, A.; Fei-Fei, L.; Savarese, S. Social LSTM: Human Trajectory Prediction in Crowded Spaces. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Institute of Electrical and Electronics Engineers (IEEE), Las Vegas, NV, USA, 27–30 June 2016; pp. 961–971. [Google Scholar] [CrossRef] [Green Version]
  38. Yoon, J.H.; Yang, M.-H.; Lim, J.; Yoon, K.-J. Bayesian Multi-object Tracking Using Motion Context from Multiple Objects. In Proceedings of the 2015 IEEE Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 5–9 January 2015; pp. 33–40. [Google Scholar] [CrossRef] [Green Version]
  39. Robicquet, A.; Sadeghian, A.; Alahi, A.; Savarese, S. Learning Social Etiquette: Human Trajectory Understanding IN Crowded Scenes. Lect. Notes Comput. Sci. 2016, 549–565. [Google Scholar] [CrossRef] [Green Version]
  40. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Patt. Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Kong, T.; Sun, F.; Yao, A.; Liu, H.; Lu, M.; Chen, Y. RON: Reverse Connection with Objectness Prior Networks for Object Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, Honolulu, HI, USA, 21–26 July 2017; pp. 5244–5252. [Google Scholar] [CrossRef] [Green Version]
  42. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y. Ssd: Single shot multibox detector. In Proceedings of the 14th European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 21–37. [Google Scholar]
Figure 1. Quadruplet network-based faster region-based convolutional neural network (RCNN) architecture. The unmanned aerial vehicle (UAV) videos are fed into the novel generated image G-16 (VGG-16) network model. At this stage, the hard negative mining were fine-tuned and moved to the quadruplet network based Deep CNN. Finally, the resulting output is shown in sequence M1303 benchmark.
Figure 1. Quadruplet network-based faster region-based convolutional neural network (RCNN) architecture. The unmanned aerial vehicle (UAV) videos are fed into the novel generated image G-16 (VGG-16) network model. At this stage, the hard negative mining were fine-tuned and moved to the quadruplet network based Deep CNN. Finally, the resulting output is shown in sequence M1303 benchmark.
Electronics 10 00795 g001
Figure 2. Deep CNN architecture.
Figure 2. Deep CNN architecture.
Electronics 10 00795 g002
Figure 3. Five sequences of videos in different benchmarks. (a) M0204 sequence with quadruplet-based faster RCNN, (b) M0301 sequence with quadruplet-based faster RCNN, (c) M0704 sequence with quadruplet-based faster RCNN, (d) M1301 sequence with quadruplet-based faster RCNN, and (e) M1303 sequence with quadruplet-based faster RCNN.
Figure 3. Five sequences of videos in different benchmarks. (a) M0204 sequence with quadruplet-based faster RCNN, (b) M0301 sequence with quadruplet-based faster RCNN, (c) M0704 sequence with quadruplet-based faster RCNN, (d) M1301 sequence with quadruplet-based faster RCNN, and (e) M1303 sequence with quadruplet-based faster RCNN.
Electronics 10 00795 g003aElectronics 10 00795 g003b
Figure 4. Comparative analysis of the proposed method with existing methods. CEM: continuous energy minimization, GOG: global optimal greedy, IOUT: intersection-over-union tracker, RMOT: relative motion online tracking, RLSTM: relative long short term memory, Bayesian multi-object tracking (BMOT), SORT: simple online and real-time tracking, SLSTM: social long short term memory.
Figure 4. Comparative analysis of the proposed method with existing methods. CEM: continuous energy minimization, GOG: global optimal greedy, IOUT: intersection-over-union tracker, RMOT: relative motion online tracking, RLSTM: relative long short term memory, Bayesian multi-object tracking (BMOT), SORT: simple online and real-time tracking, SLSTM: social long short term memory.
Electronics 10 00795 g004
Figure 5. Comparative analysis of proposed method with existing methods based on the minimum false rate. CEM: continuous energy minimization, GOG: global optimal greedy, IOUT: intersection-over-union tracker, RMOT: relative motion online tracking, RLSTM: relative long short term memory, Bayesian multi-object tracking (BMOT), SORT: simple online and real-time tracking, SLSTM: social long short term memory.
Figure 5. Comparative analysis of proposed method with existing methods based on the minimum false rate. CEM: continuous energy minimization, GOG: global optimal greedy, IOUT: intersection-over-union tracker, RMOT: relative motion online tracking, RLSTM: relative long short term memory, Bayesian multi-object tracking (BMOT), SORT: simple online and real-time tracking, SLSTM: social long short term memory.
Electronics 10 00795 g005
Figure 6. Comparative analysis of proposed method with existing methods. CEM: continuous energy minimization, GOG: global optimal greedy, IOUT: intersection-over-union tracker, RMOT: relative motion online tracking, RLSTM: relative long short term memory, Bayesian multi-object tracking (BMOT), SORT: simple online and real-time tracking, SLSTM: social long short-term memory.
Figure 6. Comparative analysis of proposed method with existing methods. CEM: continuous energy minimization, GOG: global optimal greedy, IOUT: intersection-over-union tracker, RMOT: relative motion online tracking, RLSTM: relative long short term memory, Bayesian multi-object tracking (BMOT), SORT: simple online and real-time tracking, SLSTM: social long short-term memory.
Electronics 10 00795 g006
Figure 7. Comparative analysis of proposed method with existing methods. CEM: continuous energy minimization, GOG: global optimal greedy, IOUT: intersection-over-union tracker, RMOT: relative motion online tracking, RLSTM: relative long short term memory, Bayesian multi-object tracking (BMOT), SORT: simple online and real-time tracking, SLSTM: social long short term memory.
Figure 7. Comparative analysis of proposed method with existing methods. CEM: continuous energy minimization, GOG: global optimal greedy, IOUT: intersection-over-union tracker, RMOT: relative motion online tracking, RLSTM: relative long short term memory, Bayesian multi-object tracking (BMOT), SORT: simple online and real-time tracking, SLSTM: social long short term memory.
Electronics 10 00795 g007
Figure 8. Comparative analysis of proposed method with existing methods. CEM: continuous energy minimization, GOG: global optimal greedy, IOUT: intersection-over-union tracker, RMOT: relative motion online tracking, RLSTM: relative long short term memory, Bayesian multi-object tracking (BMOT), SORT: simple online and real-time tracking, SLSTM: social long short term memory.
Figure 8. Comparative analysis of proposed method with existing methods. CEM: continuous energy minimization, GOG: global optimal greedy, IOUT: intersection-over-union tracker, RMOT: relative motion online tracking, RLSTM: relative long short term memory, Bayesian multi-object tracking (BMOT), SORT: simple online and real-time tracking, SLSTM: social long short term memory.
Electronics 10 00795 g008
Figure 9. Comparative analysis of proposed method with existing methods. CEM: continuous energy minimization, GOG: global optimal greedy, IOUT: intersection-over-union tracker, RMOT: relative motion online tracking, RLSTM: relative long short term memory, Bayesian multi-object tracking (BMOT), SORT: simple online and real-time tracking, SLSTM: social long short term memory.
Figure 9. Comparative analysis of proposed method with existing methods. CEM: continuous energy minimization, GOG: global optimal greedy, IOUT: intersection-over-union tracker, RMOT: relative motion online tracking, RLSTM: relative long short term memory, Bayesian multi-object tracking (BMOT), SORT: simple online and real-time tracking, SLSTM: social long short term memory.
Electronics 10 00795 g009
Figure 10. Comparative analysis of proposed method with existing methods. CEM: continuous energy minimization, GOG: global optimal greedy, IOUT: intersection-over-union tracker, RMOT: relative motion online tracking, RLSTM: relative long short term memory, Bayesian multi-object tracking (BMOT), SORT: simple online and real-time tracking, SLSTM: social long short term memory.
Figure 10. Comparative analysis of proposed method with existing methods. CEM: continuous energy minimization, GOG: global optimal greedy, IOUT: intersection-over-union tracker, RMOT: relative motion online tracking, RLSTM: relative long short term memory, Bayesian multi-object tracking (BMOT), SORT: simple online and real-time tracking, SLSTM: social long short term memory.
Electronics 10 00795 g010
Figure 11. Comparative analysis of proposed method with existing methods. CEM: continuous energy minimization, GOG: global optimal greedy, IOUT: intersection-over-union tracker, RMOT: relative motion online tracking, RLSTM: relative long short term memory, Bayesian multi-object tracking (BMOT), SORT: simple online and real-time tracking, SLSTM: social long short term memory.
Figure 11. Comparative analysis of proposed method with existing methods. CEM: continuous energy minimization, GOG: global optimal greedy, IOUT: intersection-over-union tracker, RMOT: relative motion online tracking, RLSTM: relative long short term memory, Bayesian multi-object tracking (BMOT), SORT: simple online and real-time tracking, SLSTM: social long short term memory.
Electronics 10 00795 g011
Table 1. The detection input of the faster RCNN. MOT: multiple object tracking; IDP: identification precision; IDR: identification recall. MOTA: multiple object tracking accuracy. MOTP: multiple object tracking precision. FP: false positive. FN: false negative. IDS: identification switches, and the total number of times a trajectory is fragmented (FM).
Table 1. The detection input of the faster RCNN. MOT: multiple object tracking; IDP: identification precision; IDR: identification recall. MOTA: multiple object tracking accuracy. MOTP: multiple object tracking precision. FP: false positive. FN: false negative. IDS: identification switches, and the total number of times a trajectory is fragmented (FM).
MOT MethodsIDF1IDPIDRMOTAMOTPFPFNIDSFM
CEM 10.219.477.369.672,378290,96224884248
GOG 1823.314.634.472.241,126168,19414,30112,516
IOUT 23.730.319.536.672.142,245163,881993810,463
BMOT 33.327.841.439.872.3319,008151,48559735897
RLSTM 31.338.626.325.669.171,955180,461133313,088
SMOT 4555.737.833.972.257,112166,52817529577
SORT 43.758.934.83974.333,037172,62823505787
SLSTM 37.246.830.837.97244,783161,009604812,051
IPGAT 49.463.240.63972.242,135163,837209110,057
PROPOSED55674540.37430,065150,83710913057
Table 2. R-FCN (region-based fully convolutional network) evaluation with respect to performance metrics. MOT: multiple object tracking; IDP: identification precision; IDR: identification recall. MOTA: multiple object tracking accuracy. MOTP: multiple object tracking precision. FP: false positive. FN: false negative. IDS: identification switches, and the total number of times a trajectory is fragmented (FM).
Table 2. R-FCN (region-based fully convolutional network) evaluation with respect to performance metrics. MOT: multiple object tracking; IDP: identification precision; IDR: identification recall. MOTA: multiple object tracking accuracy. MOTP: multiple object tracking precision. FP: false positive. FN: false negative. IDS: identification switches, and the total number of times a trajectory is fragmented (FM).
MOT MethodsIDF1IDPIDRMOTAMOTPFPFNIDSFM
CEM10.318.47.29.670.481,617289,68322013789
GOG30.938.825.628.577.160,511176,25669356823
IOUT4447.540.926.975.998,789145,61749036129
RMOT34.428.74343.876.3328,677158,76029492286
RLSTM29.334.825.214.770.697,670191,72013959953
SMOT4453.537.324.577.276,544179,60913705142
SORT42.658.733.530.278.544,612190,99922484378
SLSTM35.643.630.127.576.966,980172,94273559791
IPGAT47.560.139.230.277.158,875177,30417996705
PROPOSED49.56444.240.379.642,065135,61711912057
Table 3. SSD (single-shot detector) evaluation with respect to performance metrics. MOT: multiple object tracking; IDP: identification precision; IDR: identification recall. MOTA: multiple object tracking accuracy. MOTP: multiple object tracking precision. FP: false positive. FN: false negative. IDS: identification switches, and the total number of times a trajectory is fragmented (FM).
Table 3. SSD (single-shot detector) evaluation with respect to performance metrics. MOT: multiple object tracking; IDP: identification precision; IDR: identification recall. MOTA: multiple object tracking accuracy. MOTP: multiple object tracking precision. FP: false positive. FN: false negative. IDS: identification switches, and the total number of times a trajectory is fragmented (FM).
MOT MethodsIDF1IDPIDRMOTAMOTPFPFNIDSFM
CEM10.121.16.66.870.464,373298,09015302835
GOG29.233.625.933.676.470,080148,369796410,023
IOUT29.434.525.633.576.665,549154,04269938793
RMOT26.118.941.81075.1544,591131,38292527211
RLSTM26.528.924.47.669129,660182,82834712,654
SMOT41.945.938.627.276.595,737149,77727389605
SORT37.145.831.133.276.757,440166,49339187898
SLSTM39.444.235.53375.877,706144,617601913,332
IPGAT43.349.638.534.17671,519148,248473911,128
PROPOSED44.55142.235.578.654,065129,61712671890
Table 4. RON evaluation with respect to performance metrics. MOT: multiple object tracking; IDP: identification precision; IDR: identification recall. MOTA: multiple object tracking accuracy. MOTP: multiple object tracking precision. FP: false positive. FN: false negative. IDS: identification switches, and the total number of times a trajectory is fragmented (FM).
Table 4. RON evaluation with respect to performance metrics. MOT: multiple object tracking; IDP: identification precision; IDR: identification recall. MOTA: multiple object tracking accuracy. MOTP: multiple object tracking precision. FP: false positive. FN: false negative. IDS: identification switches, and the total number of times a trajectory is fragmented (FM).
MOT MethodsIDF1IDPIDRMOTAMOTPFPFNIDSFM
CEM10.118.86.99.768.878,265293,57620863526
GOG5160.244.335.77262,929153,33631045130
IOUT50.159.143.435.67263,086153,34829915103
RMOT28.822.5403470.9418,222153,48079027007
RLSTM36.542.432.124.168.787,589169,866115612,657
SMOT52.660.846.332.87273,226154,69611574643
SORT54.666.946.137.272.253,435159,34713693661
SLSTM48.156.441.93571.965,093152,48140136059
IPGAT54.164.346.835.97262,038154,87116795062
PROPOSED57.56747.339.273.952,065150,91714762539
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dike, H.U.; Zhou, Y. A Robust Quadruplet and Faster Region-Based CNN for UAV Video-Based Multiple Object Tracking in Crowded Environment. Electronics 2021, 10, 795. https://doi.org/10.3390/electronics10070795

AMA Style

Dike HU, Zhou Y. A Robust Quadruplet and Faster Region-Based CNN for UAV Video-Based Multiple Object Tracking in Crowded Environment. Electronics. 2021; 10(7):795. https://doi.org/10.3390/electronics10070795

Chicago/Turabian Style

Dike, Happiness Ugochi, and Yimin Zhou. 2021. "A Robust Quadruplet and Faster Region-Based CNN for UAV Video-Based Multiple Object Tracking in Crowded Environment" Electronics 10, no. 7: 795. https://doi.org/10.3390/electronics10070795

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop