Next Article in Journal
Trust–Region Nonlinear Optimization Algorithm for Orientation Estimator and Visual Measurement of Inertial–Magnetic Sensor
Previous Article in Journal
Robust Flight-Path Angle Consensus Tracking Control for Non-Minimum Phase Unmanned Fixed-Wing Aircraft Formation in the Presence of Measurement Errors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Robust Pedestrian Re-Identification and Out-Of-Distribution Detection Framework

by
Abdelhamid Bouzid
1,*,
Daniel Sierra-Sosa
2 and
Adel Elmaghraby
1
1
Department of Computer Science and Engineering, University of Louisville, Louisville, KY 40208, USA
2
Department of Computer Science and Information Technology, Hood College, Frederick, MD 21701, USA
*
Author to whom correspondence should be addressed.
Drones 2023, 7(6), 352; https://doi.org/10.3390/drones7060352
Submission received: 18 April 2023 / Revised: 15 May 2023 / Accepted: 25 May 2023 / Published: 27 May 2023
(This article belongs to the Topic Artificial Intelligence in Sensors, 2nd Volume)

Abstract

:
Pedestrian re-identification is an important field due to its applications in security and safety. Most current solutions for this problem use CNN-based feature extraction and assume that only the identities that are in the training data can be recognized. On the one hand, the pedestrians in the training data are called In-Distribution (ID). On the other hand, in real-world scenarios, new pedestrians and objects can appear in the scene, and the model should detect them as Out-Of-Distribution (OOD). In our previous study, we proposed a pedestrian re-identification based on von Mises–Fisher (vMF) distribution. Each identity is embedded in the unit sphere as a compact vMF distribution far from other identity distributions. Recently, a framework called Virtual Outlier Synthetic (VOS) was proposed, which detects OOD based on synthesizing virtual outliers in the embedding space in an online manner. Their approach assumes that the samples from the same object map to a compact space, which aligns with the vMF-based approach. Therefore, in this paper, we revisited the vMF approach and merged it with VOS to detect OOD data points. Experiment results showed that our framework was able to detect new pedestrians that do not exist in the training data in the inference phase. Furthermore, this framework improved the re-identification performance and holds a significant potential in real-world scenarios.

1. Introduction

Pedestrian tracking and re-identification systems based on machine learning have emerged as a significant solution for various safety and security applications [1,2]. These systems utilize a mapping function that is trained to embed images into a compact Euclidean space, such as a unit sphere [3,4]. The primary objective of this embedding is to ensure that images depicting the same person are mapped to nearby feature points, while images depicting different people are mapped to distant feature points. However, in real-world scenarios, there may be situational changes such as differences in pedestrian position, orientation, and occlusion within a single scene, which can adversely affect the effectiveness of the embedding approach. To overcome these challenges, it is essential to develop robust pedestrian re-identification systems that can handle such variations. Additionally, the embedding approach should not rely on clothing appearance, given that individuals may wear different clothing over time, spanning days or weeks.
Pedestrian re-identification is a challenging task in computer vision, where the goal is to recognize a pedestrian across multiple camera views. It has been a topic of intense research over the last decade due to its importance in various applications, such as surveillance and forensics [5,6,7]. Traditional approaches rely on hand-crafted features and metrics to match individuals across cameras [8,9]. However, with the recent advancements in deep learning, machine learning-based pedestrian re-identification systems have gained significant popularity, outperforming traditional methods. These systems use deep neural networks to learn discriminative feature representations from images and use them for matching individuals across camera views. Despite the considerable progress made, Pedestrian re-identification remains a challenging task, particularly in real-world settings, due to variations in lighting, pose, viewpoint, and occlusion.
Recent advancements in deep learning have revolutionized the field of re-ID [10,11]. Deep learning-based approaches, particularly those based on Convolutional Neural Networks (CNNs), have demonstrated remarkable performance on public benchmarks. The core idea behind these approaches is to learn a discriminative embedding that maps images of the same pedestrian to nearby points in the embedding space, while images of different pedestrian are mapped to distant points. The embedding is learned in an end-to-end manner, by jointly optimizing a classification loss and a triplet loss, which encourages the embeddings to preserve inter-class and intra-class distances. However, deep learning-based approaches are still prone to overfitting and suffer from the problem of detecting Out-Of-Distribution (OOD) data, which is data that differs significantly from what the model was trained on. To address these issues, there is a need to develop more robust deep learning-based pedestrian re-identification systems. The detection of OOD is crucial for the robustness of trained models in real-world scenarios.
The performance of many machine learning-based tracking and re-identification applications is significantly influenced by the image acquisition systems utilized, such as static cameras, as well as the associated costs of data collection. Unmanned Aerial Vehicles (UAVs) have recently emerged as a viable alternative for monitoring public spaces, offering a low-cost means of data collection while covering broad and challenging-to-access regions [12,13,14,15]. The advancements in UAV technology have greatly benefited multi-object tracking (MOT), particularly pedestrian tracking and re-identification, by providing a practical solution to various challenges, such as occlusion, moving cameras, and difficult-to-reach locations. Compared to static cameras, UAVs are considerably more flexible, allowing for adaptability in emplacement location and direction in the Three-Dimensional (3D) space.
In a previous study [16], we presented a solution to the problem of pedestrian tracking and re-identification from aerial devices. Our approach involved modeling each identity as a von Mises–Fisher (vMF) distribution, which was inspired by a methodology proposed by [17] for image classification and retrieval. Specifically, we learned a compact embedding for each identity in a unit sphere using a base Convolutional Neural Network (CNN) encoder.
Figure 1 is an illustration of one of the deep learning classification model’s biggest limitation. This limitation happens when it comes to detecting Out-Of-Distribution data. Because the embeddings are learned from only In-Distribution data, they may not be able to accurately represent or detect data that is significantly different from what they were trained on. This means that deep learning models can struggle to identify and handle novel or anomalous data, which is a significant challenge in many real-world applications.
In an open-world environment, it is highly likely that a deployed model will encounter new pedestrians that it has not learned to recognize, which falls under the category of OOD. Additionally, the model is expected to encounter objects that differ from humans, which are also referred to as OOD data points. The OOD data points can be classified into two types of shifts: non-semantic shifts, such as new pedestrians, and semantic shifts, such as objects that differ from humans. The definition of semantic shifts is drawn from [18]. Therefore, the detection of OOD is crucial in ensuring the robustness of trained models in real-world scenarios.
In order to tackle the challenge of OOD detection, one possible approach is to train a deep learning classifier to distinguish between ID and OOD using real-world OOD data points. However, obtaining or generating sufficient real-world OOD data points in high-dimensional pixel space is a challenging and intractable task. To address this issue, a recent study by [19] proposed a framework called Virtual Outlier Synthetic (VOS), which synthesizes virtual outliers in the embedding space in an online manner. This framework utilizes the compact embedding of samples from the same object to generate virtual outliers, using a class-conditional multivariate Gaussian distribution, which is consistent with the objective of the vMF method.
The recognition of pedestrians and detection of Out-Of-Distribution (OOD) objects is of great importance for developing robust and efficient surveillance systems. In light of this, the combination of the von Mises–Fisher (vMF) method and Virtual Outlier Synthetic (VOS) framework presents a promising solution for improving the accuracy of pedestrian re-identification and OOD detection. Building upon our previous research, we propose an integrated end-to-end learning framework that leverages the vMF method for modeling each ID, enabling simultaneous recognition of individuals and detection of OOD objects. Our proposed approach builds on the strengths of both the vMF and VOS methods, with the vMF method providing a compact embedding for pedestrian recognition and the VOS framework synthesizing virtual outliers for effective OOD detection.
To address this challenge, the paper proposes an integrated end-to-end learning framework that leverages the vMF method for pedestrian recognition and a Virtual Outlier Synthetic (VOS) framework for effective Out-Of-Distribution (OOD) detection. The proposed approach builds on the strengths of both the vMF and VOS methods to provide a promising solution for improving the accuracy of pedestrian re-identification and OOD detection. This paper aims to contribute to the development of robust and efficient surveillance systems by providing a practical solution to various challenges faced by machine learning-based person re-identification systems, particularly when using UAVs for data collection.
The present paper makes the following key contributions:
  • A novel end-to-end framework for re-identifying pedestrians and detecting OOD instances from aerial devices.
  • The first method, to the best of our knowledge, that leverages online virtual outlier synthetic to address OOD in pedestrian re-identification.
The structure of this paper is as follows. Section 2 provides a background and related work overview. Section 3 presents a review of the preliminaries related to the vMF method and VOS, providing a basic understanding of directional statistics, along with a review of the VOS framework. Section 4 describes the dataset case study. In Section 5, we propose our online pedestrian re-identification and OOD detection framework. In Section 6, we present the experimental results. Finally, we conclude with a summary of our contributions and potential future work in Section 8.

2. Related Work

2.1. Pedestrian Re-Identification

Research on pedestrian re-identification encompasses various aspects, ranging from feature-based [20] to metric-based [21] approaches, as well as from hand-crafted features to deeply learned features [22,23]. In this paper, we focus on three recent and relevant sub-areas within the pedestrian re-identification topic.
Open-world person re-identification is a specific instance of set matching, where the goal is to match one-to-one between two pedestrian sets, namely the probe and gallery sets. This task assumes that each person appears in both sets and aims to identify matching pairs. The open-world setting means that the identity of the pedestrians in the probe set may not be known in advance, which adds an additional layer of complexity to the problem. This is in contrast to the closed-world setting, where all pedestrian identities are known beforehand. In this work, we focus on the open-world person re-identification problem, where the gallery set is assumed to be known.
The generalized-view re-identification problem involves learning discriminative features from two different views obtained by two distinct, stationary cameras [24,25]. However, collecting, annotating, and matching data from two separate cameras can be expensive in practical scenarios.
In recent years, there has been a growing interest in pedestrian re-identification from drones, leading to the development of new benchmark datasets [14,26]. Drones offer a novel tool for data acquisition, particularly in the field of video surveillance and analysis. This presents new opportunities and challenges for pedestrian detection, tracking, and re-identification, as it helps to overcome some of the limitations associated with static cameras.

2.2. Out-Of-Distribution Detection

The Out-Of-Distribution (OOD) problem may be treated as a classification task, wherein D i n = ( X i , y i ) i N denotes the ID, where X i R k , and y i 1 . . C for C classes. A distribution p i n is utilized to produce D i n . To predict class probabilities, many classification models, including deep neural networks, are trained on D i n datasets. When the model is used in production in the open world, it encounters data derived from a different distribution p o u t during inference, where p i n p o u t . These two distributions are not identical. A direct approach involves sampling from the p o u t distribution. However, sampling from the high-dimensional pixel space is complex and impractical.
The disparity between distributions p i n and p o u t can be classified into two primary categories, namely semantic and non-semantic shifts. A semantic shift arises when a novel class appears in p o u t , while a non-semantic shift arises when instances of objects from the same class that are present in p i n appear differently in comparison to those observed during the training phase. The latter type is akin to an anomaly detection configuration. To enable the model to function effectively in the production setting, it must detect these types of shifts in the data.
In recent years, many studies have addressed the issue of detecting Out-Of-Distribution (OOD) samples in a classification task setup. A binary scoring function, S ( x ) , is commonly used in these studies, where a high score is assigned to data points from the In-Distribution (ID) and a low score is assigned to OOD samples. This scoring function can be learned using energy models, linear transformations with deep neural networks, or a combination of both techniques.

2.3. Anomaly Detection

Anomaly detection refers to the identification of unusual events, items, or observations that deviate significantly from expected behaviors or patterns. In the context of computer vision, outliers, noise, and novel objects are detected as anomalies when compared to the distribution of known objects. This problem is often encountered in various industrial applications where acquiring images of normal samples is easy, but specifying expected variations in defects is difficult and costly. Such scenarios are often referred to as Out-Of-Distribution (OOD) detection problems, where a model is required to distinguish between samples drawn from the training data distribution and those lying outside its support. Existing work on anomaly detection is predominantly based on learning compact visual representations in a latent space using auto-encoders and GANs. Unsupervised methods using pre-trained CNNs, such as PatchCore and SPADE, as well as PaDIM, are widely used in industrial applications.
Pedestrian re-identification can be framed as an anomaly detection problem, where the task is to identify pedestrians from the known training data distribution and to detect pedestrians that do not belong to this distribution as outliers or anomalies. These anomalous pedestrians could arise due to a variety of reasons such as novel viewpoints, changes in illumination, or occlusions.

2.4. Deep Metric Learning

In numerous machine learning applications, such as multi-object tracking (MOT), it is crucial to establish a measure of similarity among data objects. Metric learning aims to learn a mapping function that quantifies this similarity. Specifically, the objective of metric learning is to minimize the distance between data points belonging to the same category while maximizing the distance between data points from different categories.
In recent years, deep learning has demonstrated remarkable performance in various machine learning tasks, including image classification, image embedding, and multiple object tracking (MOT). The superior representational power of deep learning in extracting highly abstract non-linear features has resulted in the emergence of a new research area known as Deep Metric Learning (DML) [17,27,28,29]. This field aims to learn a mapping function that quantifies the similarity between data points, with the objective of minimizing the similarity between data points from the same category and maximizing the distance between data points from different categories.
Multiple Object Tracking (MOT) has been enhanced by the success of Deep Metric Learning (DML), which involves training a neural network to extract features and learn a similarity measure between object instance patches.

3. Preliminaries

3.1. Directional Statistics in Machine Learning

Directional data refers to points in an Euclidean space whose norm is constrained to be one, denoted by | x | 2 = 1 , where | . | 2 represents the Euclidean norm of order two. In other words, these points lie on the surface of a unit sphere. The statistical analysis of such data is referred to as directional statistics.
The topic of directional statistics has gained a significant amount of attention due to high demands from fields such as machine learning or the availability of big datasets that necessitate adaptive statistical methodologies, as well as technical improvements. Recently, the directional statistics method has led to tremendous success in many computer vision tasks, such as image classification and retrieval [17], pose estimation [30], and face verification [31]. It has also been introduced to other machine learning fields such as text mining [32].

3.1.1. Von Mises–Fisher Distribution

Von Mises–Fisher Distribution (vMF) is a probability distribution function for directional data. It can be seen as a Gaussian distribution because they have very similar properties. In a directional data space, S P 1 , the probability distribution density function is defined as:
f p ( x ; μ , κ ) = Z p ( κ ) exp ( κ μ T x ) .
where μ is the mean direction of the distribution, κ 0 is a concentration parameter, which can be seen as the standard deviation for Gaussian distribution, p is the space dimension, Z p ( κ ) = κ p / 2 1 ( 2 π ) p / 2 I p / 2 1 ( κ ) is a normalization term, and I v is the modified Bessel function of the first kind with order v.
Given N samples from a vMF distribution, we can estimate its parameters as follows:
μ ^ = i = 1 N x i i = 1 N x i 2 ,
and
κ ^ = R ¯ ( p R ¯ 2 ) 1 R ¯ 2 .
In (16), R ¯ = i = 1 N x i 2 N .

3.1.2. Learning von Mises–Fisher Distribution

The learning problem is defined as follows. Given C identities, the goal is to learn a vMF distribution for every ID parameterized by { κ i , μ i } , where i = 1 C .
Given a point, x, in the mapping space, the normalized probability of x belonging to a chosen class, c, is defined as:
P ( c | x , { κ i , μ i } i = 1 C ) = Z p ( κ c ) exp ( κ c μ c T x ) i = 1 C Z p ( κ i ) exp ( κ i T μ i x )
Equation (4) can be used to increase the likelihood that the sample belongs to the correct class while decreasing the likelihood that it belongs to other classes. Given a mini-batch with N samples and for a C identity, we can maximize the following objective function:
P ( Y | X , Θ , , κ ) = n = 1 N P ( c | x , { κ i , μ i } i = 1 C )
= n = 1 N Z p ( κ c ) exp ( κ c μ c T x ) i = 1 C Z p ( κ i ) exp ( κ i μ i T x ) ,
where X and Y represent the data points in the mini-batch and their ID labels, Θ contains the deep model parameters, and = { μ i } i = 1 C , κ = { κ i } i = 1 C . For simplification purposes, we assumed κ to be a constant for all IDs, and by applying the negative likelihood, Equation (6) can be simplified to:
arg min Θ , L = n = 1 N log exp ( κ μ c T x ) i = 1 C exp ( κ μ i T x )
In [17], a learning algorithm (Algorithm 1) based on alternative learning is proposed to overcome the difficulty of optimizing both the neural network parameters, Θ , and the von Mises–Fisher (vMF) mean direction distributions, κ . In this algorithm, the mean directions are first fixed, and then the neural network parameters are trained for several iterations before updating them using the training dataset. The mean direction update is based on the estimation using all training data points. The algorithm converges when the mean directions and loss are stagnant. Given a class, i, with N training data points, x n denotes the mapping of the n th sample using the current mapping function, where n = 1 N . The mean direction of class i can be updated using the following formula:
μ ^ i = n = 1 N x n n = 1 N x n 2 ,
Algorithm 1 VMF learning algorithm.
  • Initialize CNN parameters Θ .
  • Repeat:
    (a)
    Estimate mean directions using (8) and all the training data.
    (b)
    Train CNN for several iterations and update Θ .
  • Until convergence.
In the inference phase, the identification (ID) label of a given object can be predicted by computing the cosine similarity between the object’s feature vector and the learned mean directions. The object will be labeled with the ID of the mean vector that is closest in terms of cosine similarity.

3.2. Visual Outlier Synthetic

The synthesis of outliers is a potent method for generating synthetic data points, and various approaches have been developed to synthesise images in computer vision. Among these approaches, Generative Adversarial Networks (GANs) [33] are the most widely used and straightforward technique. However, synthesising images in the high-dimensional pixel space is challenging to optimize and track. To overcome this challenge, ref. [19] proposed the “Visual Outlier Synthetic” (VOS) framework, which synthesizes virtual outliers in the embedding space in an online manner. Their method depends on the model learning the embedding of the ID objects to generate hard virtual outliers.
The Visual Outlier Synthetic (VOS) framework is predicated on the assumption that the feature representation of object instances in the embedding space conforms to a class-conditional multi-variate Gaussian distribution. Specifically, objects belonging to the same class form a multi-variate Gaussian distribution within the latent representational space.

3.2.1. Learning of Class-Conditional Multi-Variate Gaussian Distribution

Given a coreset that represents the embeddings of the objects, it is possible to learn a class-conditional Gaussian distribution by estimating its parameters. Specifically, from the coreset, the empirical class mean μ ^ and covariance Σ ^ of the training samples ( x n , y i ) i = 1 N can be computed as follows:
μ k ^ = 1 N k n : y n = k x n .
Σ k ^ = 1 N k n : y n = k ( x n μ k ^ ) ( μ k ^ x n ) T .
where N k is the number of objects in class k, and N is the total number of objects.

3.2.2. Sampling from the Features Representational Space

The authors propose to generate virtual outliers by sampling from the feature representation space using the class-conditional multivariate distributions described above. The virtual outliers are generated in an online manner, with the learning progress resulting in increasingly compact embeddings for each class. Sampling virtual outliers from the learned class-conditional distribution aligns with this objective and helps to achieve a more compact embedding.
f ( x n ) = 1 ( 2 π ) m 2 | Σ ^ | exp 1 2 ( x n μ k ^ ) T | Σ ^ | 1 ( μ k ^ x n ) .
k = { ν k | f ( ν k ) < ϵ } .
The sampling of virtual outliers for a given class is achieved by drawing samples from the class-conditional Gaussian distribution using the expression ν k N ( ν k , Σ ^ ) , where Σ ^ is the estimated covariance matrix of the coreset for that class. The sampled virtual outliers are restricted to the sublevel set based on the likelihood, ensuring that they align with the underlying distribution of the class. Additionally, the magnitude of ϵ is set to be sufficiently small so that the generated outliers are situated in the vicinity of the class boundary.

3.2.3. Out-Of-Distribution Detection

The VOS framework utilized a linear transformation to differentiate between virtual outliers. Specifically, they employed a Fully Connected Pooling (FCP) layer that learned to delimit the precise boundaries surrounding the ID. Additional information regarding the classification process between In-Domain (ID) and Out-of-Domain (OOD) instances, as well as the learning algorithm employed, can be found in the original publication [19].

4. Dataset Description

In recent years, pedestrian tracking and re-identification have become a topic of significant interest owing to their wide-ranging applications, including but not limited to surveillance systems and traffic control. Nevertheless, tracking and re-identification of individuals present a substantial challenge due to the limitations of the acquisition system, particularly when using stationary cameras. In recent years, Unmanned Aerial Vehicles (UAVs) have emerged as a promising alternative for monitoring public areas, as they offer an inexpensive means of data collection while effectively covering large and remote areas that may otherwise be inaccessible.
The P-DESTRE dataset, developed by researchers from the University of Beira Interior (Portugal) and the JSS Science and Technology University (India), is a fully annotated dataset for detecting, tracking, re-identificating, and searching pedestrians from aerial devices [26]. The dataset was collected using “DJI Phantom 4” drones piloted by humans to fly at altitudes ranging from 5.5 to 6.7 m and collect data from a volunteer audience walking. An illustration of how the data was acquired is showed in Figure 2, along with a statistic summary detailed in Table 1. The dataset comprises 75 videos recorded at a rate of 30 frames Per Second (FPS), containing a total of 318,745 annotated instances of 269 different IDs. The image resolution is 3840 × 2160 pixels.
The primary distinguishing feature of the P-DESTRE dataset is the pedestrian search challenge, where data are collected over long periods with constant ID labels across observations. This characteristic distinguishes it from comparable datasets, making it an excellent case study for training and evaluating frameworks for pedestrian tracking and re-identification from aerial devices. In this context, re-identification techniques cannot rely on clothing appearance-based features, which is a key property that distinguishes search from the less challenging re-identification problem.
The P-DESTRE dataset was used for all experiments, and it provides a unique and valuable resource for research on pedestrian tracking and re-identification from aerial devices. In the future, the researchers plan to explore more aerial datasets for further investigation.

5. Methodology

In our prior publication [16], we introduced a method founded on directional statistics that enables the learning of a condensed representation for each identification (ID) within a unit spherical space. The ID data obtained was a collection of von Mises–Fisher (vMF) distributions that were parameterized by κ i , μ i , where i = 1 . . C . The learning procedure for this method was detailed in Section 2. The aim of this method was to track and re-identify a set of pre-defined pedestrians. Nonetheless, in an open-world scenario such as security environments, the prospect of encountering new pedestrians not present in the dataset is highly probable. Consequently, the model must be capable of detecting such pedestrians as Out-Of-Distribution (OOD).
The proposed scoring functions presented for discriminating between ID and OOD are chiefly derived from the compressed representation of the objects within the ID. These functions enforce the embedding of each class within a compact cluster by the Convolutional Neural Networks (CNNs).
Motivated by this observation, the potential synergy between the vMF-based model and OOD scoring functions warrants exploration. Inspired by the two works presented in the previous section, we propose a framework that merges both approaches.

5.1. Out-Of-Distribution Pedestrian Based on von Mises–Fisher Distribution

5.1.1. Learning

We propose an end-to-end learning framework for pedestrian re-identification and novelty detection that is as simple to train as the traditional training method with soft-max loss. The framework consists of three main components: a representational visual feature based on a Convolutional Neural Network (CNN), the adoption of the VOS method to detect Out-Of-Distribution (OOD) pedestrians, and the generation of hard virtual outliers by sampling from the embedding space. We posit that sampling from the embedding space cannot only aid in detecting novelty but also help build a robust model. By assuming that the score function will help to learn a more compact embedding for each identification while simultaneously detecting OOD pedestrians as non-semantic shifts, we generate hard virtual outliers. Figure 3 illustrates the proposed framework during the training phase.
Once the hard virtual outliers are generated, two parallel heads are computed. The first head computes the von Mises–Fisher (vMF) loss using the embeddings of the identification. The second head computes the novelty loss over the binary output of the linear transformation. This loss aims to distinguish between virtual outliers and identification embeddings. The objective loss is then computed by taking a weighted sum of the two losses.
In the VOS framework, the uncertainty loss (novelty loss) is defined using the binary sigmoid loss.
L N o v e l t y = log 1 1 + e x p Θ u . E ( v , Θ ) log e x p Θ u . E ( x , Θ ) 1 + e x p Θ u . E ( x , Θ ) .
where Θ u represents the weights of the classification head for novelty detection, E ( . ) is the energy score function, and Θ is the weights of the CNNs base encoder.
To train this framework in an end-to-end manner, we combined the two losses to form one objective training loss.
L l o s s = L v m f + γ L N o v e l t y .
This is a weighted loss, where γ is the weighted of the uncertainty loss. The Algorithm  2 illustrates the different steps of the learning framework.
Algorithm 2 The learning algorithm.
Input: ID data D i n , queue size | Q | for Gaussian density estimation, γ weight for uncertainty regularization, and ϵ .
Output: pedestrian re-identification parameterized by Θ , and novelty detector S parameterized by Θ u
While train do:
  • Initialize CNN parameters Θ .
  • Repeat:
    (a)
    Estimate mean directions using (8) and all the training data.
    (b)
    For several iteration:
    i.
    update the ID queue Q k with the embeds of training inputs.
    ii.
    estimate the multi-variate distribution using the Q k .
    iii.
    sample virtual outlier
    iv.
    compute the objective loss using Equation (14).
    (c)
    Train CNN for several iterations and update Θ .
  • Until convergence.

5.1.2. Inference

In the inference process for a given pedestrian input, two main parts are involved. The first part involves utilizing the OOD scoring function, S , to determine whether the pedestrian is OOD or not. This decision can be made based on a chosen threshold, which is determined by the experimental settings.
I D , if S ( x ) > T O O D , otherwise .
Once a pedestrian input is determined to be an ID, the proposed framework uses the vMF framework as the second step to re-identify the pedestrian. The re-identification process involves measuring the similarity between the input embedding in the unit sphere and the learned mean direction, μ i i = 1 N . This comparison is achieved by calculating the cosine similarity between the input pedestrian and the mean direction associated with each pedestrian ID. The input pedestrian is then assigned to the ID of the pedestrian with the highest cosine similarity to its mean direction.

6. Experiments

In this section, we present a comprehensive summary of the experiments conducted to assess the performance of our proposed framework. We followed a standard procedure in all experiments, which involved training the proposed framework on the training set of the In-Distribution (ID) dataset, denoted as D i n . The evaluation was performed on the union of two sets: the validation set of the ID, denoted as D i n v a l , and the Out-Of-Distribution (OOD) dataset, denoted as D o u t . It is crucial to note that the ID and OOD datasets should not have any overlapping person identities.
The setting of the ID and OOD datasets can be performed in two distinct ways. Firstly, two separate datasets with no common person identities can be selected and set as the ID and OOD datasets, respectively. Secondly, the same dataset can be divided into ID and OOD by splitting each set into ID and OOD with a predetermined ratio, for the training, validation, and testing sets.
It is important to emphasize that during both the training and validation phases, the framework does not have access to the OOD dataset. The performance of the model was monitored based on the validation set of the ID and the online generated virtual outliers. This involved selecting the appropriate model checkpoint and optimizing the hyper-parameters of the framework. During the testing phase, we used both the test set of D i n and D o u t to evaluate the model’s long-term re-identification and OOD detection capabilities.

6.1. Two Different Dataset Settings

6.1.1. ID Dataset

The P-DESTRE dataset http://http://p-destre.di.ubi.pt/download.html (accessed on 5 March 2023) was utilized as the (ID) data in the pedestrian re-identification experiments. A thorough description of the P-DESTRE dataset can be found in Section 3. To perform the experiments, the data were randomly divided into 5 folds, with each fold consisting of a 50% learning set, a 10% gallery set, and a 40% query set. This division of data into folds helps to ensure a fair evaluation of the models and provides a comprehensive understanding of the performance of the models across different data splits. Additionally, this also helps to mitigate the risk of overfitting and provides a more robust evaluation of the models, given that the models are tested on unseen data. By using a randomly divided dataset, the results of the experiments are more representative of the models’ general performance and can be better compared to other models and their results. The detailed information regarding this split is provided in http://p-destre.di.ubi.pt/pedestrian_detection_splits.zip (accessed on 5 March 2023).

6.1.2. OOD Dataset

As OOD, we used CUHK03-NP, which is a widely used dataset in the field of pedestrian re-identification. It contains 14,097 images of 1467 identities [22] https://github.com/zhunzhong07/person-re-ranking/blob/master/CUHK03-NP/README.md (accessed on 5 March 2023). All the splits were used in the evaluation as an D o u t . The CUHK03-NP dataset is a commonly used benchmark dataset in the field of computer vision and machine learning, particularly for evaluating person re-identification algorithms. The dataset was collected from two camera views at the Chinese University of Hong Kong and consists of 1467 identities captured in a multi-shot manner. Each identity is captured with both color and depth images, providing a rich source of information for developing and testing algorithms. The dataset has become a popular choice for researchers due to its large scale and the presence of challenging conditions, such as occlusions and viewpoint changes, which pose difficulties for re-identification algorithms. The CUHK03-NP dataset has been used in many recent studies, demonstrating its utility as a benchmark for evaluating the performance of re-identification algorithms under various conditions. Table 2 summarises the dataset properties.

6.1.3. Training Details

From each image, the bounding boxes are cropped and scaled to patches of dimension ( 48 , 64 , 3 ) pixels. For the prepossessing, the input patches are normalized using the mean and standard deviation learned from imagenet dataset. The feature extractor architecture is made up of two parts: a base model and a header. As a base model, we used Wide ReseNet-50 (WRN) [34], with a header consisting of two Fully connected Pooling Layers (FPL) of sizes [ 4096 , 128 ] neurons. The feature extractor was trained using 50 epochs with a batch size of 64; 128 is the embedding space dimension. The used optimizer is Adam, with a learning rate starting with 0.2 and decreasing by a factor of 0.5 every 25 of the training epochs. We set the concentration parameter, κ , to 15 for the learning algorithm hyper-parameter. This number produced the best outcomes experimentally. We update the mean directions after every epoch.
For hyper-parameters related to VOS, we used 500 samples per identity to estimate the class-conditional Gaussians. We set ϵ = 0.0001 . We sampled 1000 virtual outliers from the embedding space. We also set γ = 0.1 . The linear transformation consists of two layers of [ 269 , 1 ] neurons. The first layer has a number of nodes equal to the number of identities so that the energy based can be computed per identity as designed in VOS.

6.1.4. Two Datasets Settings

In evaluating the framework’s capability in detecting Out-Of-Distribution (OOD) samples, we employed the Area Under the Curve (AUC) based on Precision/Recall curve metric. This metric was chosen for a number of compelling reasons. Firstly, it is a more accurate metric when dealing with imbalanced data, which is the case in the ratio of ID vs. OOD samples in the open world scenario. Secondly, it is preferred when the positive class, in this case ID, is of utmost importance. The results in Table 3 summarize the performance of a pedestrian tracking and re-identification framework based on the Area Under the Curve based on the Precision/Recall curve (AUC-PR) metric. The framework was tested using two datasets, the ID dataset P-DESTRE and the OOD dataset CUHK03-NP. The results show that the framework had a performance of 63.10% ± 1.64% AUC-PR using the Wide Residual Network (WRN) as its backbone. This suggests that the framework performed well in detecting ODD samples; however, the results should be interpreted with caution as the standard deviation of the results is not provided, which can give an indication of the variance in the performance of the model. Additionally, it would be beneficial to compare the performance with other existing methods to put the results into perspective.

6.2. One Dataset Setting

Another way to evaluate the framework is on the same pedestrian dataset but divided into ID and OOD by identities. Detecting OOD data points in a different dataset setting is expected to be easier comparing to the one dataset setting regarding the image acquisition setups. The differences between the two image acquisitions, such as the type of cameras, the angle, and the lighting, can play a role in obtaining a separable embedding between ID and OOD. Although it is important to detect OOD from a different setup, it is also important to test the model on ID and OOD from the same image acquisition setup. We believe that the one dataset setup is more likely to happen in the open world environment. In addition, to detect the non-semantic shifts (new pedestrians) in the same setup is more challenging, and the model has to rely on features such as face and body characteristics rather than clothing appearances.
Because the P-DESTRE dataset is randomly divided into 5 folds, we divided each fold into two datasets, ID and OOD, based on the identities. The division ration is 70 / 30 % for ID and OOD, respectively. It is worth mentioning that this division is preformed for each fold.
For the training details, everything is almost the same as in the two different dataset settings, except we lowered the value of ϵ to 0.001 . This can be explained by the fact that distinguishing the ID from the OOD in this setting is more challenging and requires a harder virtual outlier to learn a better score function, S .
We evaluated the model the same way we evaluated the two different dataset settings. The results presented in Table 4 show the performance of a framework using the WRN backbone on the ID dataset. The performance is evaluated using the Area Under the Curve based on Precision-Recall (AUC-PR) metric, and the results are reported as the mean ± standard deviation over multiple runs. The results show that the framework has an AUC-PR of 55.19% ± 3.02%. This indicates that the framework has moderate performance in identifying instances of the positive class (ID) in the P-DESTRE dataset, with some variance in performance between runs.

6.3. Long-Term Pedestrian Re-Identification

The effectiveness of our re-identification method can be evaluated using two methods. Firstly, we calculate the nearest mean direction of the IDs using Equation ( 9 ) . Secondly, we assess the top-N recall performance using commonly used metrics in the field. The results, summarized in Table 5, demonstrate a significant improvement over other state-of-the-art methods in different metrics. This confirms that the vMF-based feature extractor is able to learn robust features that help in recognizing a person, rather than their clothing appearance. Furthermore, our results show that the integration of the vMF method with the VOS framework leads to even better re-identification performance, as it creates a synergy that pushes the embedding of each identity to be more compact and distinct from other identities and Out-Of-Distribution samples. The comparison between the method ArcFace with COSAM and the proposed vMF identifier with and without the VOS framework was performed in terms of Mean Average Precision (mAP), Rank-1 accuracy, Rank-20 accuracy, and Mean Direction. The results indicate that the vMF identifier, both with and without VOS, significantly outperforms the ArcFace with COSAM method. With a mAP of 40.85% ± 3.42%, the vMF identifier achieved a higher performance compared to the ArcFace with COSAM, which had a mAP of 34.90% ± 6.43%. Similarly, the Rank-1 accuracy and Rank-20 accuracy were higher for the vMF identifier, with values of 63.81% ± 4.50% and 88.61% ± 8.50%, respectively, compared to the 49.88% ± 8.01% and 70.10% ± 11.25% achieved by the ArcFace with COSAM. Furthermore, the vMF identifier also outperformed the ArcFace with COSAM in terms of the Mean Direction with a value of 64.45% ± 3.90%. These results demonstrate the effectiveness of the proposed vMF identifier in pedestrian re-identification.

6.4. Analysis

Figure 4 presents a binary confusion matrix that delineates instances of ID and OOD, whereby the model prediction is compared to the ground truth. In delving deeper into examples of wrong predictions, we observed that these instances were frequently mapped in close proximity to the decision boundary where low energy function is learned.
In the case of the example where OOD was predicted as ID, we noted that when an individual’s image was captured from the back, it was often predicted as ID. Conversely, when the same individual’s image was captured from the front, the model predicted it as ID with greater accuracy.
In the instance where ID was predicted as OOD, our analysis suggested that this was a limitation of the trained model. Further improvements in the model are needed to better distinguish between similar-looking identities.

7. Discussion

7.1. Limitations

Despite the success of pedestrian tracking and re-identification using deep learning, there are still some limitations that need to be addressed. One of the main limitations is the limited size and diversity of datasets, which limits the algorithms’ ability to generalize to real-world scenarios. In fact, there are only a handful of publicly available datasets that have tracking and re-identification labels. In addition, the quality of the training data significantly affects the algorithms’ performance, which emphasizes the importance of collecting large and diverse datasets for training. Another limitation is the sensitivity of deep learning algorithms to occlusions and their computational intensity, making it challenging to implement them in real-time scenarios. Lastly, the training process is complex and time-consuming, requiring specialized hardware, such as GPUs, and taking several hours to several days to complete, making it an expensive and challenging task for researchers and practitioners.

7.2. Future Work

The field of pedestrian tracking and re-identification using deep learning is still in its early stages, and several limitations and challenges need to be addressed to realize its full potential. Future work needs to focus on developing improved datasets, real-time processing algorithms, robustness to changes in the environment, transfer learning algorithms, and multi-modal data approaches. The development of large-scale datasets with high-quality annotations will be essential to drive future advances in the field. The focus on developing more efficient algorithms that can run in real-time will increase their reliability in real-world applications. Furthermore, algorithms need to be robust to changes in the environment and incorporate other modalities such as audio, depth, and motion to improve the performance of pedestrian tracking and re-identification algorithms. We anticipate as a follow up that convergence analysis will be performed and consideration will be made to use other neural network architectures. By addressing the current challenges and limitations, the field has the potential to make significant contributions to the development of real-world applications.

8. Conclusions

In this study, we have presented an extension of our previous work on pedestrian re-identification using vMF distribution. We have revisited this method and combined it with the Visual Object Segmentation (VOS) framework to propose a new approach that we believe is worth exploring for the pedestrian re-identification problem. Our proposed framework is evaluated on a pedestrian dataset acquired from aerial devices. The results demonstrate that our approach improves long-term re-identification performance, not only over previously applied methods but also over the same method without the VOS. Our goal was to detect non-semantic shifts as Out-Of-Distribution (OOD) data. However, the experiments also revealed that it is more challenging to detect non-semantic shifts when the OOD data comes from the same acquisition setup.
As a future direction, we plan to extend our study to a more diverse range of datasets.

Author Contributions

All authors contributed equally to the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Informed Consent Statement

The objective of this study is to augment the dependability, security, and safety of machine learning pedestrian tracking and re-identification models. The research conducted herein has the potential to generate direct societal benefits and impact safety and surveillance applications, including public area monitoring. Regulatory compliance has been upheld throughout the course of our study. By undertaking this research, we seek to advance the understanding of the problem of pedestrian monitoring and re-identification in practical settings, thus heightening public awareness of safety concerns. We anticipate no deleterious effects from the pursuit of this investigation.

Data Availability Statement

The data used is publicly available at http://p-destre.di.ubi.pt/ (accessed on 20 August 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ming, Z.; Zhu, M.; Wang, X.; Zhu, J.; Cheng, J.; Gao, C.; Yang, Y.; Wei, X. Deep learning-based person re-identification methods: A survey and outlook of recent works. Image Vis. Comput. 2022, 119, 104394. [Google Scholar] [CrossRef]
  2. Singh, N.K.; Khare, M.; Jethva, H.B. A comprehensive survey on person re-identification approaches: Various aspects. Multimed. Tools Appl. 2022, 81, 15747–15791. [Google Scholar] [CrossRef]
  3. Wang, B.H.; Wang, Y.; Weinberger, K.Q.; Campbell, M. Deep Person Re-identification for Probabilistic Data Association in Multiple Pedestrian Tracking. arXiv 2018, arXiv:1810.08565. [Google Scholar]
  4. Jiang, Y.F.; Shin, H.; Ju, J.; Ko, H. Online pedestrian tracking with multi-stage re-identification. In Proceedings of the 2017 14th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Lecce, Italy, 29 August–1 September 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–6. [Google Scholar]
  5. Simonnet, D.; Lewandowski, M.; Velastin, S.A.; Orwell, J.; Turkbeyler, E. Re-identification of pedestrians in crowds using dynamic time warping. In Proceedings of the Computer Vision–ECCV 2012. Workshops and Demonstrations: Florence, Italy, 7–13 October 2012, Proceedings, Part I 12; Springer: Berlin/Heidelberg, Germany, 2012; pp. 423–432. [Google Scholar]
  6. Varior, R.R.; Haloi, M.; Wang, G. Gated siamese convolutional neural network architecture for human re-identification. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016, Proceedings, Part VIII 14; Springer: Berlin/Heidelberg, Germany, 2016; pp. 791–808. [Google Scholar]
  7. Zhao, R.; Ouyang, W.; Wang, X. Person re-identification by salience matching. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2528–2535. [Google Scholar]
  8. Cheng, D.S.; Cristani, M.; Stoppa, M.; Bazzani, L.; Murino, V. Custom pictorial structures for re-identification. In Proceedings of the Bmvc, Dundee, UK, 29 August–2 September 2011; Citeseer: Princeton, NJ, USA, 2011; Volume 1, p. 6. [Google Scholar]
  9. Farenzena, M.; Bazzani, L.; Perina, A.; Murino, V.; Cristani, M. Person re-identification by symmetry-driven accumulation of local features. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 2360–2367. [Google Scholar]
  10. Barbosa, I.B.; Cristani, M.; Del Bue, A.; Bazzani, L.; Murino, V. Re-identification with rgb-d sensors. In Proceedings of the Computer Vision–ECCV 2012. Workshops and Demonstrations: Florence, Italy, 7–13 October 2012, Proceedings, Part I 12; Springer: Berlin/Heidelberg, Germany, 2012; pp. 433–442. [Google Scholar]
  11. Kim, S.; Zimmermann, T.; Kim, M.; Hassan, A.; Mockus, A.; Girba, T.; Pinzger, M.; Whitehead Jr, E.J.; Zeller, A. TA-RE: An exchange language for mining software repositories. In Proceedings of the 2006 International Workshop on Mining Software Repositories, Shanghai, China, 22–23 May 2006; pp. 22–25. [Google Scholar]
  12. Bonetto, M.; Korshunov, P.; Ramponi, G.; Ebrahimi, T. Privacy in mini-drone based video surveillance. In Proceedings of the 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Ljubljana, Slovenia, 4–8 May 2015; IEEE: Piscataway, NJ, USA, 2015; Volume 4, pp. 1–6. [Google Scholar]
  13. Hirzer, M.; Beleznai, C.; Roth, P.M.; Bischof, H. Person re-identification by descriptive and discriminative classification. In Proceedings of the Scandinavian Conference on Image Analysis, Ystad, Sweden, 23–27 May 2011; Springer: Berlin/Heidelberg, Germany, 2011; pp. 91–102. [Google Scholar]
  14. Layne, R.; Hospedales, T.M.; Gong, S. Investigating open-world person re-identification using a drone. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; Springer: Berlin/Heidelberg, Germany, 2014; pp. 225–240. [Google Scholar]
  15. Singh, A.; Patil, D.; Omkar, S. Eye in the sky: Real-time Drone Surveillance System (DSS) for violent individuals identification using ScatterNet Hybrid Deep Learning network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1629–1637. [Google Scholar]
  16. Bouzid, A.; Sierra-Sosa, D.; Elmaghraby, A. Directional Statistics-Based Deep Metric Learning for Pedestrian Tracking and Re-Identification. Drones 2022, 6, 328. [Google Scholar] [CrossRef]
  17. Zhe, X.; Chen, S.; Yan, H. Directional statistics-based deep metric learning for image classification and retrieval. Pattern Recognit. 2019, 93, 113–123. [Google Scholar] [CrossRef]
  18. Hsu, Y.C.; Shen, Y.; Jin, H.; Kira, Z. Generalized odin: Detecting out-of-distribution image without learning from out-of-distribution data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 10951–10960. [Google Scholar]
  19. Du, X.; Wang, Z.; Cai, M.; Li, Y. VOS: Learning What You Don’t Know by Virtual Outlier Synthesis. arXiv 2022, arXiv:2202.01197. [Google Scholar]
  20. Zheng, W.S.; Gong, S.; Xiang, T. Person re-identification by probabilistic relative distance comparison. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 649–656. [Google Scholar]
  21. Dikmen, M.; Akbas, E.; Huang, T.S.; Ahuja, N. Pedestrian recognition with a learned metric. In Proceedings of the Asian Conference on Computer Vision, Queenstown, New Zealand, 8–12 November 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 501–512. [Google Scholar]
  22. Li, W.; Zhao, R.; Xiao, T.; Wang, X. Deepreid: Deep filter pairing neural network for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 152–159. [Google Scholar]
  23. Yi, D.; Lei, Z.; Liao, S.; Li, S.Z. Deep metric learning for person re-identification. In Proceedings of the 2014 22nd International Conference on Pattern Recognition, Stockholm, Sweden, 24–28 August 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 34–39. [Google Scholar]
  24. Avraham, T.; Gurvich, I.; Lindenbaum, M.; Markovitch, S. Learning implicit transfer for person re-identification. In Proceedings of the European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; Springer: Berlin/Heidelberg, Germany, 2012; pp. 381–390. [Google Scholar]
  25. Koestinger, M.; Hirzer, M.; Wohlhart, P.; Roth, P.M.; Bischof, H. Large scale metric learning from equivalence constraints. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 2288–2295. [Google Scholar]
  26. Kumar, S.A.; Yaghoubi, E.; Das, A.; Harish, B.; Proença, H. The p-destre: A fully annotated dataset for pedestrian detection, tracking, and short/long-term re-identification from aerial devices. IEEE Trans. Inf. Forensics Secur. 2020, 16, 1696–1708. [Google Scholar] [CrossRef]
  27. Bouzid, A. Automatic Target Recognition with Deep Metric Learning. Master’s Thesis, University of Louisville, Louisville, KY, USA, 2020. [Google Scholar]
  28. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
  29. Rippel, O.; Paluri, M.; Dollar, P.; Bourdev, L. Metric learning with adaptive density discrimination. arXiv 2015, arXiv:1511.05939. [Google Scholar]
  30. Prokudin, S.; Gehler, P.; Nowozin, S. Deep directional statistics: Pose estimation with uncertainty quantification. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 534–551. [Google Scholar]
  31. Hasnat, M.; Bohné, J.; Milgram, J.; Gentric, S.; Chen, L. von mises-fisher mixture model-based deep learning: Application to face verification. arXiv 2017, arXiv:1706.04264. [Google Scholar]
  32. Straub, J.; Chang, J.; Freifeld, O.; Fisher, J., III. A Dirichlet process mixture model for spherical data. In Proceedings of the Artificial Intelligence and Statistics. PMLR, San Diego, CA, USA, 9–12 May 2015; pp. 930–938. [Google Scholar]
  33. Salimans, T.; Goodfellow, I.; Zaremba, W.; Cheung, V.; Radford, A.; Chen, X. Improved techniques for training gans. Adv. Neural Inf. Process. Syst. 2016, 29. [Google Scholar]
  34. Zagoruyko, S.; Komodakis, N. Wide residual networks. arXiv 2016, arXiv:1605.07146. [Google Scholar]
  35. Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. Arcface: Additive angular margin loss for deep face recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4690–4699. [Google Scholar]
  36. Subramaniam, A.; Nambiar, A.; Mittal, A. Co-segmentation inspired attention networks for video-based person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 562–572. [Google Scholar]
Figure 1. The figure depicted herein provides a representation of the Convolutional Neural Network’s (CNN) learning process in the embedding space: (a) presents a visual demonstration of how deep learning algorithms are trained to differentiate between three classes; (b) shows the OOD area that the trained model will classify as ID.
Figure 1. The figure depicted herein provides a representation of the Convolutional Neural Network’s (CNN) learning process in the embedding space: (a) presents a visual demonstration of how deep learning algorithms are trained to differentiate between three classes; (b) shows the OOD area that the trained model will classify as ID.
Drones 07 00352 g001
Figure 2. The P-DESTRE datasets were obtained using a consistent data gathering technique. Human operators flew “DJI Phantom 4” aircraft at altitudes ranging from 5.5 to 6.7 m to mimic autonomous surveillance of urban scenes. The gimbal pitch angle ranged from 45 to 90 degrees [26].
Figure 2. The P-DESTRE datasets were obtained using a consistent data gathering technique. Human operators flew “DJI Phantom 4” aircraft at altitudes ranging from 5.5 to 6.7 m to mimic autonomous surveillance of urban scenes. The gimbal pitch angle ranged from 45 to 90 degrees [26].
Drones 07 00352 g002
Figure 3. Pedestrian re-identification and novelty detection framework. Inputs are mapped using the CNN backbone. Then, virtual outliers are generated using the VOS framework. Vmf loss is computed using only the features of the inputs. Novelty loss is computed over the generated virtual outlier features and the the inputs features.
Figure 3. Pedestrian re-identification and novelty detection framework. Inputs are mapped using the CNN backbone. Then, virtual outliers are generated using the VOS framework. Vmf loss is computed using only the features of the inputs. Novelty loss is computed over the generated virtual outlier features and the the inputs features.
Drones 07 00352 g003
Figure 4. We presented a binary confusion matrix illustrating instances of ID and OOD, where the ground truth was compared against the model prediction.
Figure 4. We presented a binary confusion matrix illustrating instances of ID and OOD, where the ground truth was compared against the model prediction.
Drones 07 00352 g004
Table 1. P-DESTRE Dataset Statistics Summary.
Table 1. P-DESTRE Dataset Statistics Summary.
Total number of videos75
Frames Per Second (FPS)30
Total number of identities269
Total number of annotated instances 318 , 745
Camera range distance(5.5–6.7) m
Table 2. Properties of the CUHK03-NP Dataset.
Table 2. Properties of the CUHK03-NP Dataset.
PropertyValue
Dataset NameCUHK03-NP
Number of Images14,097
Number of Identities1467
Resolution640 × 480
AnnotationsPedestrian bounding boxes
SourceCUHK Person Re-Identification Dataset
Table 3. The Area Under the Curve (AUC)-based Precision/Recall curve (PR) results obtained by applying our framework to the two different datasets setting. The ID is P-DESTRE, and the OOD is CUHK03-NP.
Table 3. The Area Under the Curve (AUC)-based Precision/Recall curve (PR) results obtained by applying our framework to the two different datasets setting. The ID is P-DESTRE, and the OOD is CUHK03-NP.
ID DatasetOOD DatasetBackboneAUC-PR
P-DESTRECUHK03-NPWRN63.10% ± 1.64%
Table 4. The Area Under the Curve (AUC)-based Precision/Recall curve (PR) results obtained by applying our framework to the two different datasets setting. The ID is P-DESTRE, and the OOD is CUHK03-NP.
Table 4. The Area Under the Curve (AUC)-based Precision/Recall curve (PR) results obtained by applying our framework to the two different datasets setting. The ID is P-DESTRE, and the OOD is CUHK03-NP.
DatasetBackboneAUC-PR
ID datasetP-DESTREWRN55.19% ± 3.02%
Table 5. Comparison between the re-identification performance attained by the state-of-the-art methods and ours based on vMF on the P-DESTRE dataset [26]. ArcFace + COSAM taken from [26].
Table 5. Comparison between the re-identification performance attained by the state-of-the-art methods and ours based on vMF on the P-DESTRE dataset [26]. ArcFace + COSAM taken from [26].
MethodmAPRank-1Rank-20Mean Direction
ArcFace [35] + COSAM [36]34.9% ± 6.43%49.88% ± 8.01%70.10% ± 11.25%
vMF identifier37.85% ± 3.42%53.81% ± 4.5%74.61% ± 8.5%64.45% ± 3.9%
[vMF + VOS] identifier39.15% ± 2.41%56.18% ± 3.2%78.59% ± 7.3%66.5% ± 2.9%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bouzid, A.; Sierra-Sosa, D.; Elmaghraby, A. A Robust Pedestrian Re-Identification and Out-Of-Distribution Detection Framework. Drones 2023, 7, 352. https://doi.org/10.3390/drones7060352

AMA Style

Bouzid A, Sierra-Sosa D, Elmaghraby A. A Robust Pedestrian Re-Identification and Out-Of-Distribution Detection Framework. Drones. 2023; 7(6):352. https://doi.org/10.3390/drones7060352

Chicago/Turabian Style

Bouzid, Abdelhamid, Daniel Sierra-Sosa, and Adel Elmaghraby. 2023. "A Robust Pedestrian Re-Identification and Out-Of-Distribution Detection Framework" Drones 7, no. 6: 352. https://doi.org/10.3390/drones7060352

Article Metrics

Back to TopTop