Next Article in Journal
Application of Artificial Intelligence and Sensor Fusion for Soil Organic Matter Prediction
Previous Article in Journal
Analysis of Heart Rate Variability in Individuals Affected by Amyotrophic Lateral Sclerosis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Toward Better Pedestrian Trajectory Predictions: The Role of Density and Time-to-Collision in Hybrid Deep-Learning Algorithms

by
Raphael Korbmacher
* and
Antoine Tordeux
Department for Traffic Safety and Reliability, University of Wuppertal, 42119 Wuppertal, Germany
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(7), 2356; https://doi.org/10.3390/s24072356
Submission received: 26 February 2024 / Revised: 27 March 2024 / Accepted: 6 April 2024 / Published: 8 April 2024
(This article belongs to the Section Physical Sensors)

Abstract

:
Predicting human trajectories poses a significant challenge due to the complex interplay of pedestrian behavior, which is influenced by environmental layout and interpersonal dynamics. This complexity is further compounded by variations in scene density. To address this, we introduce a novel dataset from the Festival of Lights in Lyon 2022, characterized by a wide range of densities (0.2–2.2 ped/ m 2 ). Our analysis demonstrates that density-based classification of data can significantly enhance the accuracy of predictive algorithms. We propose an innovative two-stage processing approach, surpassing current state-of-the-art methods in performance. Additionally, we utilize a collision-based error metric to better account for collisions in trajectory predictions. Our findings indicate that the effectiveness of this error metric is density-dependent, offering prediction insights. This study not only advances our understanding of human trajectory prediction in dense environments, but also presents a methodological framework for integrating density considerations into predictive modeling, thereby improving algorithmic performance and collision avoidance.

1. Introduction

The challenge of predicting pedestrian trajectories has emerged as a pivotal challenge in recent years. This surge in interest is largely attributed to the profound implications it holds for autonomous vehicle navigation [1], service robot deployment [2], and the strategic planning of infrastructure and mass gatherings [3]. In addressing these intricate challenges, researchers have traditionally employed physics-based (PB) models to simulate and understand pedestrian behavior. These models have been instrumental in dissecting collective phenomena and enhancing our understanding of pedestrian dynamics, particularly in high-density contexts relevant to crowd management and evacuation strategies [3]. However, the landscape of pedestrian trajectory prediction has witnessed a paradigm shift over the last decade with the advent and integration of deep learning (DL) algorithms [4]. Despite the opaqueness of these models in terms of interpretability, their superiority in mirroring observed trajectories has been markedly pronounced, especially when juxtaposed with their PB counterparts [5]. Nonetheless, it is important to acknowledge that the domains of applicability for PB and DL models do not entirely overlap. While PB models excel in the realm of high-density simulations, providing insights into collective behavior, DL algorithms predominantly thrive in low-density environments where individual pedestrian movements are characterized by a greater degree of freedom and intricate long-range interactions [4]. This paper introduces a novel, real-world pedestrian trajectory dataset, gathered during the Festival of Lights in Lyon. Field pedestrian trajectory datasets are typically gathered from low-density situations. In contrast, this dataset captures the nuanced dynamics of pedestrian movements across a large spectrum of density levels. It ranges from sparse crowds observed during show moments to the densely packed throngs seen after the event. Utilizing this dataset, we train DL algorithms, including Long Short-Term Memory (LSTM) networks and Generative Adversarial Networks (GAN). Our methodology is underscored by a novel approach: we harness situational classification predicated on crowd density to refine our models’ learning process. This two-stage process, which initially classifies the scene based on density and then predicts the trajectories, not only bolsters the efficiency of our models but also substantially elevates the precision of trajectory predictions. This improvement is demonstrated by comparative analyses with traditional DL algorithms and PB models.
Another challenge of trajectory prediction that we face in this paper is overlapping and colliding of predicted trajectories [6]. To tackle that problem, we integrate a time-to-collision (TTC) [7] term into the loss function of the algorithms. A parameter, λ , is utilized to modulate the TTC’s influence on the training. Our empirical research uncovers a significant relationship between the optimal λ values and the density levels, highlighting the intricacies of pedestrian behavior across different densities.
The remainder of this paper is organized as follows: first, we review related studies in Section 2. Then, our novel dataset is presented in Section 3 and the methodology for the empirical work is proposed in Section 4. The results are shown in Section 5. Finally, Section 6 includes a discussion of the results and an outlook on future works.

2. Related Work

The domain of pedestrian trajectory prediction is multifaceted, drawing insights from various disciplines and methodologies. The two main stream are the physics-based models and data-based algorithms [4]. PB models have been the cornerstone of understanding pedestrian dynamics, especially in high-density scenarios. The Social Force model (SF), introduced by Helbing and Molnar [8], exemplifies this approach, simulating pedestrian movement by balancing attractive and repulsive forces. Other famous PB models are the Optimal Reciprocal Collision Avoidance (ORCA) from Van den Berg et al. [9] or the cellular automata model from Burstedde et al. [10]. However, these models are not without their challenges, particularly when it comes to encapsulating the full range of crowd behavior [4]. For more PB models, see reviews such as [11,12,13].
In pursuit of addressing these limitations, the research frontier has gradually shifted towards data-driven methodologies. Notably, the past decade has witnessed a burgeoning interest in DL approaches. Pioneering works such as the Social LSTM from Alahi et al. [5] introduce the use of Recurrent Neural Networks (RNN), specifically LSTM networks, in conjunction with a novel concept known as Social Pooling. This innovative approach incorporates neighbouring information, thereby enriching the model’s contextual understanding. This social concept was further enhanced by Gupta et al. [14] through Social GAN, where the generative adversarial framework allowed for the generation of multiple plausible future paths, addressing the inherent uncertainty in human movement. Alternative methods employed for social predictions include attention mechanisms [15], graph-based approaches [16], and the utilization of relative coordinates [17]. Additionally, deep learning architectures such as Convolutional Neural Networks [18,19] and Transformers [20] have been applied in trajectory prediction tasks.
A pivotal aspect of this paper is the innovative classification of trajectory scenes based on crowd density prior to prediction. To the best of our knowledge, this approach is a novel paradigm, potentially owing to the scarcity of high-density, real-world pedestrian trajectory datasets. Xue et al. [21] predicted pedestrian destinations using bidirectional LSTM classification. This involves an additional classification stage to distinguish between possible destinations of pedestrians. They classify the route manually into four distinct categories. In another paper from the authors [22], the classification was based on a clustering algorithm. Kothari et al. [23] categorized pedestrian trajectories based on the nature of interactions observed, identifying behaviors such as collision avoidance, leader–follower dynamics, and grouping behavior. An alternative methodology involves classifying trajectories based on individual pedestrian characteristics. Papathanasopoulou et al. [24] concentrated on attributes such as age, gender, height, and speed to inform their classification. A second cornerstone of this work is the seamless integration of a PB concept, TTC, into the loss function of DL algorithms. This synthesis of PB principles and DL models is not an isolated endeavor. Alahi et al. [5] and Khadka et al. [25] utilized simulated data from PB models for training DL algorithms. Antonucci et al. [26] embedded a PB model directly into the DL architecture. Furthermore, the works of Silvestri et al. [27] and Kothari et al. [28] stand out for their use of PB principles within the loss function to eliminate unrealistic predictions.

3. The Dataset

With a growing interest in data-based methods, the significance of pedestrian trajectory data has been elevated in recent research. This area has seen a proliferation of datasets published by researchers, which can be categorized into field data and experimental data obtained in laboratory conditions. In the field studies, real-world settings are employed where individuals, unaware of their participation in a study, navigate through various scenarios. Famous field datasets are the ETH [29] and UCY [30] datasets, which are widely used in the machine learning community. Originating from surveillance videos, these datasets capture pedestrians scene of low density (0.1–0.5 ped/ m 2 ). The GLOW dataset from Eindhoven [31], a dataset used for route choice analysis, contains trajectory scenes of higher densities, but only for short length trajectories. Other field datasets are the Stanford Drone Dataset [32], the Grand Central Station Dataset [33], and the Edinburgh Informatics Forum Dataset [34]. None of these have densities above 0.2 ped/ m 2 . In the following, we will present a field dataset with pedestrian densities between 0.2 and 2.2 ped/ m 2 .
The data were collected at Lyon’s Festival of Lights. The event runs for four days from 7 p.m. to 11 p.m., and attracts millions (2 million in 2022) of visitors each year. Key attractions are light shows at Place des Terreaux and Place Saint-Jean. We have installed cameras at the Place des Terreaux to film the area, which is represented by the red rectangle in Figure 1.
In Figure 1a, the entirety of Place des Terreaux is depicted. The red box on the right-hand side delineates our designated tracking region. Figure 1b offers an aerial perspective of this same area. This designated zone measures 9 m in length and 6.5 m in width. On average, we concurrently tracked 55 pedestrians, resulting in a mean density of 0.95 ped/ m 2 . The distribution of pedestrian density exhibited significant variability. During the light show, the majority of pedestrians congregated in the central area of the square, remaining largely stationary. Consequently, the pedestrian density within our tracking zone was relatively low. However, when the show concluded—which consistently lasts approximately 9 min—the crowd dynamics shifted dramatically, as most individuals sought to exit towards another event. In this transition phase, the density within our tracking corridor surged, often exceeding 120 pedestrians moving simultaneously. For video calibration, we meticulously established nine calibration points, ensuring the precise tracking of pedestrian trajectories using the PeTrack software [35]. Throughout our study, we recorded 5195 individual trajectories, which averaged a duration of 12.38 s and a mean velocity of 0.62 m/s. The pedestrian flow changes between mostly unidirectional flow, after the light show and bidirectional flow during the show. More informations about the data can be found in the Appendix A. For the training and testing of the algorithms, we need trajectories of a minimal length of 7 s (see Section 4.1). Because many trajectories are more than 14 s long, they can be used more than once. In total, we obtained 7450 trajectories for training and testing.

4. Methodology

4.1. Overview

For predicting pedestrian trajectories, we have i t h pedestrian in a scene represented by image coordinates ( x t i , y t i ) for each time instant t = k · d t , with k N and a time step of d t = 1 / 3 s. The observed positions from t = T 1 to t = T o b s is taken as input and the aim is to predict future trajectories from t = T o b s + d t to t = T p r e d . Every scene involves a primary pedestrian and their neigbours over the timespan T 1 to T p r e d . A neigbours is a pedestrian whose position at T 1 is closer to the position of the primary pedestrian than a radius r = 5 m. Our dataset has a framerate of three observation for each second. We choose input trajectories of 9 observations (3 s) and want to predict 12 timesteps (4 s).
The predicted trajectory of all primary pedestrians are evaluated on two commonly utilized Euclidean distance metric and a collision metric. In the first distance-based metrics, called average displacement error (ADE) [29], the distance between the predicted trajectory and the ground truth trajectory is measured at any time step t:
ADE = 1 N T i = 1 N t = 1 T x ^ i ( t ) x i ( t ) .
where x i ( t ) is the actual position of the i t h pedestrian at time t and x ^ i ( t ) is the predicted position. The Euclidean distance is denoted as · . The second distance-based metric, called the final displacement error (FDE) [30], displays the distance between the final point t = T p r e d of the predicted trajectory and the ground truth trajectory:
FDE = 1 N i = 1 N x ^ i ( T ) x i ( T ) .
These distance-based metrics are widely used in pedestrian trajectory predictions for their effectiveness in quantifying the goodness-of-fit. However, repulsive forces, which are pivotal in shaping interactions between pedestrians, are not taken into account [36]. Consequently, these metrics do not account for potential overlaps or collisions between pedestrians. Therefore, the collision metric is used to enhance the evaluating process:
COL = 1 | S | X ^ S C O L ( X ^ ) ,
with:
C O L ( X ^ ) = min 1 , t = 1 T i = 1 N j > i N | | x ^ i ( t ) x ^ j ( t ) | | 2 R .
where S includes all scenes in the test set, X ^ represents a scene prediction containing N agents, and x ^ i is the prediction of the position of the agent i over the prediction time of T, while [ · ] is the Iverson bracket.
P = 1 if P is true , 0 otherwise .
This metric counts a prediction as a collision when a predicted pedestrian trajectory intersects with neighboring trajectories, thus indicating the proportion of predictions where collisions occur. A vital factor in this calculation is the chosen pedestrian size R. An increase in R will likewise increase the number of collisions. For calculating the collision metrics, we use a radius R = 0.2 m.

4.2. Prediction Approaches

In the subsequent trajectory predictions, various trajectory prediction approaches, ranging from traditional PB models to modern DL algorithms, are chosen as benchmarks for comparison with our two-stage approach. We present the results of the Constant Velocity model (CV) and SF model [8] as well as the results of a Vanilla LSTM, the Social LSTM (SLSTM) [5], and the Social GAN [14]. These approaches, characterized by their diverse features, are commonly selected for comparison and serve as benchmarks that must be surpassed. The position and velocity of the i t h pedestrian are denoted as x i R 2 and v i R 2 , respectively. For a system of N pedestrians, the position and velocity vectors, x = ( x 1 , , x N ) and v = ( v 1 , , v N ) , have dimensions of 2 N . All variables, including x ( t ) and x i ( t ) , are functions of time t.

4.2.1. Constant Velocity Model

The CV model assumes pedestrian velocities remain unchanged over time. It serves as a baseline for more complex models. The future position of a pedestrian is predicted as:
x i ( t + t p ) = x i ( t ) + t p v i ( t ) , t p [ 0 , T p ] .

4.2.2. Social Force Model

Introduced by Helbing and Molnar [8], the SF model treats pedestrians as particles influenced by forces. Within this framework, the model calculates acceleration based on the cumulative effect of three distinct forces, as delineated in Equation (7):
m i d v i d t = m i v i 0 v i τ + j i U ( x j x i ) + W V ( x W x i )
where m i , v i , and v i 0 signify the mass, current velocity, and desired velocity of pedestrian i, respectively. The term U ( x j x i ) represents the repulsive force from other pedestrians, while V ( x W x i ) indicates the repulsive force from obstacles. The potential functions U ( d ) and V ( d ) are given by:
U ( d ) = A B e | d | / B , A , B > 0 and V ( d ) = A B e | d | / B , A , B > 0
where A , A , B , B > 0 are interaction parameters of the social force model.
The first term of Equation (7) signifies the driving force experienced by the i t h pedestrian. This force propels the individual towards their desired speed and direction within a relaxation time τ > 0 . The second term encapsulates the summation of social forces, originating from the repulsive effects as pedestrians endeavour to maintain a comfortable distance from one another. The third term accounts for the aggregate interaction forces between pedestrian i and various obstacles. While the CV model has no parameter, the SF model has three parameters, i.e., preferred velocity, interaction potential, and reaction time, which can be optimized to get accurate predictions.

4.2.3. Vanilla LSTM

LSTM networks, a class of RNN designed to learn long-term dependencies, have proven effective in handling sequential data, particularly for time series prediction tasks. Introduced by Hochreiter and Schmidhuber [37], LSTMs address the vanishing and exploding gradient problems common in traditional RNNs, making them suitable for complex sequence modeling tasks such as trajectory prediction. The vanilla LSTM model considers historical trajectories to predict future positions.
x i ( t + t p ) = x i ( t ) + LSTM t p , x i ( t t o ) , t o [ 0 , T o ] , t p [ 0 , T p ] .
where x i ( t + t p ) predicts the future trajectory of a pedestrian i at time t + t p , based on its past positions x i ( t t o ) , over an observation window t o [ 0 , T o ] .

4.2.4. Social LSTM

LSTM networks have demonstrated effective performance in sequence learning tasks. One such task, the prediction of pedestrian trajectories, presents the challenge that the trajectory of a pedestrian can be significantly influenced by the trajectories of surrounding pedestrians. The number of these neighboring influences can fluctuate widely, especially in densely crowded environments [38].
Enhancing the LSTM framework, the SLSTM by Alahi et al. [5] incorporates a social pooling layer, enabling the model to consider the influence of neighboring pedestrians explicitly. This is a key distinction from the Vanilla LSTM, reflecting the model’s capacity to capture social interactions:
x i ( t + t p ) = x i ( t ) + SLSTM i , t p , x ( t t o ) , t o [ 0 , T o ] , t p [ 0 , T p ] .
where x is the vector of positions of the neighboring pedestrians. In this formulation, the inclusion of the index i and the collective pedestrian state x emphasizes the model’s attention to the surrounding pedestrians’ trajectories, making it adept at handling complex social behaviors in dense scenarios.

4.2.5. Social GAN

Another approach we take into account is the Social GAN (SGAN) introduced by Gupta et al. [14]. This model extends traditional approaches by incorporating GANs to predict future trajectories. GANs, conceptualized by Goodfellow et al. [39], consist of two competing networks: a Generator, which generates data samples, and a Discriminator, which evaluates the authenticity of the samples against real data. SGAN leverages this architecture to generate plausible future trajectories of pedestrians, addressing the complex dynamics of pedestrian movement in crowded spaces. A key feature of the SGAN model is its pooling mechanism, which processes the relative positions of pedestrians to each other. This mechanism is crucial for understanding the social interactions and dependencies among individuals in crowded environments
x i ( t + t p ) = x i ( t ) + SGAN i , t p , x ( t t o ) , t o [ 0 , T o ] , t p [ 0 , T p ] .

4.3. Two-Stage Process

The foundation of our innovative classification framework lies in its capacity to predict trajectories across varying density levels, marking a departure from traditional models that typically utilize a single algorithm to process a wide array of scenarios within a dataset. Our strategy entails segmenting the dataset according to the density of each scene, thereby generating distinct subsets. At the inception of our methodology, we establish well-defined criteria for classification. This process is underpinned by two distinct methodologies: a statistical analysis and a review of existing literature. The results of this clustering process are depicted in Figure 2. These figures visually represent each dataset item as a point measurement each second, with the left side of Figure 2 illustrating points based on their average density:
ρ ( t ) = N ( t ) A ,
and average velocity:
v ¯ ( t ) = 1 N ( t ) i S ( t ) v i ( t ) ,
for the K-Means clustering. On the right side, the clustering is carried out by using the Agglomerative Hierarchical Clustering algorithm (AHC).
We can see clear vertical colour switch’s of the points at densities around 0.7, 1.1, and 1.6 ped/ m 2 for both cluster algorithms. Remarkably, without presetting the number of clusters or explicitly focusing on density levels, the K-Means and AHC algorithms autonomously reveal density-dependent clustering. The delineation of clusters and their boundary values align closely with those identified in the literature. Stefan Holl [40] delineated critical density thresholds for various infrastructures, signifying points at which pedestrian behavior undergoes significant changes. According to Holl, densities below 0.7 ped/ m 2 indicate a free flow state, densities below 1.3 ped/ m 2 represent a bound flow, and values above 1.3 ped/ m 2 are indicative of congested flow. In our model, we refine these categories by slightly narrowing the bound flow range and subdividing the congested flow category into two distinct segments.
Figure 3 illustrates the procedural steps undertaken to evaluate our proposed methodology and within the sizes of the clusters, which are partly taken from the classification Figure 2.
New trajectory scenes are given to our framework, where the initial step involves calculating the scene’s density using Equation (12) in individuals per square meter (ped/ m 2 ), N represents the total number of pedestrians observed within the scene, and A is the scene’s total area in square meters.
The density categorization is as follows: scenes with a density below 0.7 ped/ m 2 are labeled as lowD; densities ranging from 0.7 to 1.2 ped/ m 2 are classified as mediumD; densities between 1.2 and 1.6 ped/ m 2 are designated as highD; and densities exceeding 1.6 ped/ m 2 are identified as veryHD. Following classification, the scene is bifurcated into two segments: the initial segment spans nine timesteps and serves as input for one of the four specialized Sub-LSTMs, while the subsequent segment, encompassing twelve timesteps, is utilized to appraise the LSTMs’ performance through the computation of error metrics ADE, FDE, and COL. Figure 4a–d provide illustrative examples of each density level encountered in our dataset.
In Figure 4a, the scene exhibits very low density, with pedestrian movement primarily from two directions, leading to numerous interactions and avoidance behaviors. This is characteristic of our lowD data. Conversely, Figure 4b showcases a moderately higher density, yet still affords space for interactions, avoidance, and bidirectional pedestrian flow. In Figure 4c, representing highD data, the dynamics of pedestrian movement markedly differ from those observed in lowD and mediumD scenes, with movement predominantly unidirectional from the top, indicating a tendency to follow the pedestrian ahead. This pattern is even more pronounced in Figure 4d, where the flow from the top is so dense that passage from the bottom becomes challenging, leading pedestrians to follow the leader with limited freedom of movement and space. These observed behavioral differences underpin our classification rationale.

4.4. Collision Weight

Predictions of pedestrian trajectories presents the challenge to predict trajectory paths that do not collide with neighbours. Accurately measuring these collisions is challenging due to the shapes of pedestrians, which can vary from person to person. Traditionally, collisions are defined by the overlap of the radii of two pedestrians, as delineated in Equations (3)–(5). However, this method proves suboptimal for inclusion as a penalty function within the loss function of DL algorithms. Analysis of pedestrian trajectory data frequently reveals instances where collisions are not genuine but rather instances of grouping behavior, with individuals walking closely together, sometimes shoulder-to-shoulder. It is not these interactions we aim to deter, but rather scenarios in which individuals move directly towards one another without any attempt to avoid collision—behaviors that are unrealistic and undesirable.
To address this, we adopt the TTC concept, a widely recognized principle in the study of pedestrian dynamics [7]. Implementing this variable in the loss function of an DL algorithm would reduce predicted situations, where pedestrians walk straight towards each other without avoidance mechanism. Integrating TTC into a DL algorithm’s loss function significantly mitigates predictions where pedestrians are on a direct collision course without any avoidance mechanisms. The TTC term calculates the time until two pedestrians would collide if they continue moving at their current velocities, a concept validated by Karamouzas et al. [7]. The relative position and velocity between the pedestrian i and j can be denoted by x i j = x i x j R 2 and v i j = v i v j R 2 , respectively. A collision between pedestrian i and pedestrian j occurs if a ray, originating from ( x i , y i ) and extending in the direction of v i j , intersects the circle centred at ( x j , y j ) with a radius of R i + R j at some time τ i j in the future. This condition can be mathematically represented as | | x i j + v i j . t | | 2 < ( R i + R j ) 2 , where | | . | | denotes Euclidean norm. Solving this quadratic inequality for t yields τ i j as the smallest positive root:
τ i j = x i j · v i j ( x i j · v i j ) 2 | | v i j | | 2 ( | | x i j | | 2 ( R i + R j ) 2 ) | | v i j | | 2
A collision is imminent when τ i j = 0 , whereas a large positive value for τ i j indicates no collision risk. To implement τ i j into the loss function, we have to use an sigmoid function f that has high values, if τ i j is low and vice versa:
f ( τ ) = 1 1 + e s ( τ δ ) ,
where s = 10 and δ = 0.4 are slope and threshold parameters, respectively. This function is then integrated into the loss function, traditionally focused solely on minimizing the ADE. The revised loss function combines ADE with TTC loss, optimized through minimization:
L i = t = 1 T x i ( t ) x ^ i ( t ) + λ t = 1 T f ( min j i { τ i j } ) ,
where λ > 0 modulates the influence of the TTC component in their loss function. The calculation of τ i j considers all nearby pedestrians to the primary pedestrian, employing the minimum τ i j to identify and mitigate the most critical potential collision scenario in the model.

4.5. Implementation Details

The algorithms are implemented in the commonly accepted configurations of related contributions [23]. All computations are performed using the PyTorch framework. The learning rate is set to 0.001 and an ADAM optimizer is utilizied. The batch size is set to 8 and training is carried out for 15 epochs, if not the early stop mechanism interrupt. This is the case, when the validations error starts to rise for three epochs. For validation and testing, a hold-out validation strategy is adopted by allocating 15% of the dataset for each validation and testing, while the remaining data serves as the training set. For capturing pedestrian interactions, we choose a circles with a radius of r = 5 m surrounding the primary pedestrian.

5. Results

We will unveil the predictive outcomes of our dataset using two distinct yet synergistic methods. Initially, we will showcase the performance of our two-stage prediction framework, comparing it with contemporary state-of-the-art methodologies. Subsequently, we will demonstrate the seamless integration of our two-stage process with the incorporation of the TTC term into the loss function, illustrating its efficacy in mitigating collision instances.

5.1. Two-Stage Predictions

The results of our predictions will be presented in Table 1. As described in Section 4.3, we evaluated the predictions on the different density levels lowD, mediumD, highD, and veryHD. For every approach, we measure ADE, FDE, and COL metrics.
The initial insight gleaned from Table 1 reveals a clear trend: as density increases, the COL metric rises while the ADE/FDE diminish. This pattern emerges because higher densities naturally lead to reduced distances between individuals, consequently resulting in increased overlaps among agents. Additionally, it is observed that velocities decrease as density intensifies, leading to trajectories that are shorter in spatial extent. This reduction in travel distance directly contributes to the observed decrease in both ADE and FDE metrics at higher densities. Furthermore, it is clear that the DB algorithm outperform the traditional models CV and SF in terms of ADE/FDE. In terms of the COL metric, SF performs very well. In the last three rows of Table 1, we present the effectiveness of our two-stage approach and its combination with the TTC term. The results clearly show a significant improvement in the algorithm’s precision, attributed to the strategy of classification before prediction. Our enhanced two-stage SLSTM model consistently outperforms the traditional SLSTM across all evaluated datasets, demonstrating superior performance in terms of ADE, FDE, and COL metrics. Similarly, our adapted SGAN model shows marked improvements over the standard SGAN in three out of four datasets with respect to ADE. Integrating the TTC term further enhances the SLSTM results, notably in reducing collisions. A more detailed discussion on this enhancement is provided in the subsequent Section 5.2.

5.2. Collision Weight

In this study, we propose to integrate TTC in the loss function with the two-stage approach outlined in Section 4.3. As described in Equation (16), the collision part in the loss function can be adjusted by a parameter λ [36]. If λ is high, the impact of the TTC term is high compared to the impact of ADE and vice versa. In Figure 5, the impact of different values of λ on the prediction accuracy of the SLSTM algorithm is displayed: Figure 5a for lowD data, Figure 5b for mediumD data, Figure 5c for highD data, and Figure 5d for veryHD data. The first observation in each Figure is for λ = 0 , which means that it is equivalent to the value of our two-stage SLSTM in Table 1.
Across all figures, a consistent pattern emerges: increasing the collision weight λ generally results in fewer collisions. Optimal predictions occur at a specific λ value, where ADE and FDE are equivalent to or lower than those at λ = 0 . Each dataset exhibits a maximum effective λ value, beyond which ADE sharply increases. In the lowD dataset (Figure 5a), the ideal λ is 0.08, reducing ADE by 19% and collisions by 6%. Values slightly higher than 0.08 are still beneficial, yielding fewer collisions and enhanced avoidance behavior, but λ values exceeding 0.16 lead to a significant increase in ADE. In the mediumD dataset (Figure 5b), the optimal λ is 0.04, reducing ADE and collisions by 3% and 40%, respectively. Here, a λ value above 0.06 results in increased ADE, although a λ of 0.1 reduces the collision metric by 75%. In the highD dataset, improvements in ADE are marginal too, only notable at a λ of 0.02. However, the collision metric significantly decreases, by up to 48%, at a λ of 0.08. Conversely, in the very high-density (veryHD) dataset, increasing collision weight initially results in a rise in ADE, with no subsequent decrease. While the collision metric decreases by 37% at λ = 0.08 values, there are no improvements for ADE. These empirical observations lead to the insight, that pedestrian behavior at different densities is very different and need different parameter configurations. At lower densities, our TTC term can improve overall accuracy (ADE and COL), while at higher densities, we can only reduce COL, by taking higher ADE into account.

6. Conclusions

Pedestrian behavior is inherently complex, exhibiting a wide variety of patterns across different contexts. This paper introduced a novel pedestrian trajectory dataset, characterized by its diversity in situational contexts, including varying densities and motivations. Our analysis of the data reveals variations in pedestrian behaviors correlating with the density of the scene. To address these variations, we propose a novel two-stage classification and prediction process. This approach first classifies scenes based on density and then applies the suitable model for predicting behavior within that specific density context. Implementing this framework enhanced the prediction accuracy of two famous DL algorithms, Social LSTM and Social GAN.
Furthermore, we integrated a TTC-based term into the loss function of the SLSTM to improve avoidance behaviors, consequently reducing potential collisions. Our empirical studies indicate that the effectiveness of the TTC-based term varies with density; it significantly benefits scenarios of low density by correlating higher TTC values with reduced collision incidents. However, the outcomes in high-density situations were more ambiguous, suggesting a nuanced impact of density on the efficacy of this approach. This observation could be attributed to the nuanced dynamics of pedestrian behavior across different densities. Specifically, in environments with lower densities, pedestrians tend to navigate more through avoidance and interactions, making TTC particularly relevant. Conversely, in higher density settings, pedestrian movement is more characterized by forced leader–follower dynamics, diminishing the prominence of TTC in explaining behavior. This study underscores the complexity of pedestrian behavior, which varies significantly under different environmental conditions. It highlights the necessity of adopting a flexible modeling approach to accurately predict pedestrian trajectories in diverse settings.
This research opens several avenues for future investigation in the field of pedestrian trajectory prediction, especially concerning heterogeneous datasets characterized by variable densities. Current methodologies typically rely on a one-size-fits-all model for behavior prediction across all conditions. We advocate for the development and application of multiple specialized models, each tailored to different scene characteristics, with scene density being a pivotal factor. While focusing on density has proven to be a successful strategy, exploring additional factors and individual pedestrian behavior characteristics could yield further improvements. For instance, our methodology utilized an estimate of overall scene density. However, pedestrians do not take global densities for their decision making into account, but rather local densities. Wirth et al. [41] demonstrated that pedestrian decisions are primarily influenced by their visual neighborhood. Future studies should investigate the impact of assessing local density variations within a scene, which could be particularly beneficial in environments exhibiting a wide range of density levels. This direction could unlock new dimensions of accuracy and reliability in trajectory prediction models.
Moreover, incorporating the TTC concept into the loss function has shown promise in enhancing prediction accuracy at lower density levels. Future research should explore alternative loss functions, particularly for high-density scenarios, where the traditional ADE-based approaches may not suffice. Investigating other metrics that could more effectively capture the complexities of high-density pedestrian behavior is crucial for advancing the field.

Author Contributions

Conceptualization, R.K. and A.T.; methodology, R.K. and A.T.; software, R.K.; validation, R.K. and A.T.; formal analysis, R.K.; investigation, R.K.; resources, R.K.; data curation, R.K.; writing—original draft preparation, R.K.; writing—review and editing, R.K. and A.T.; visualization, R.K.; supervision, A.T.; project administration, A.T.; funding acquisition, A.T. All authors have read and agreed to the published version of the manuscript.

Funding

The authors acknowledge the Franco-German research project MADRAS funded in France by the Agence Nationale de la Recherche (ANR, French National Research Agency), grant number ANR-20-CE92-0033, and in Germany by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation), grant number 446168800.

Informed Consent Statement

The pedestrian trajectories collected are completely anonymous, in agreement and cooperation with the Technical Organisation Commitee of the Festival of Lights as well as the Project Management Department of the city of Lyon.

Data Availability Statement

The data are available on the following website: https://madras-data-app.streamlit.app/ (accessed on 1 April 2024). For more information see: https://www.madras-crowds.eu/.

Acknowledgments

We would like to express our gratitude to Armin Seyfried for their engaging and insightful discussions, which have borne fruit in the form of valuable ideas.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
DLDeep learning
PBPhysics-based
LSTMLong Short-Term Memory
SLSTMSocial Long Short-Term Memory
GANGenerative Adversarial Network
SGANSocial Generative Adversarial Network
TTCTime-to-collision
SFSocial Force model
CVConstant Velocity model
ORCAOptimal Reciprocal Collision Avoidance
RNNRecurrent Neural Network
ADEAverage displacement error
FDEFinal displacement error
COLCollision metric
lowDLow density
mediumDMedium density
highDHigh density
VeryHDVery high density

Appendix A

In this appendix we present some of the pedestrian trajectories from the Festival of Lights in Lyon that we used to train and test the algorithms, showing unidirectional dynamics including avoidance of standing pedestrians (Figure A1 and Figure A2) and lane formation for counterflows Figure A3. For illustration, we also show some predictions obtained with our 2-stage TTC-SLSTM in Figure A4.
The legends for the trajectory plots: in Figure A1, Figure A2 and Figure A3 are the following:
  • grey trajectories correspond to pedestrians moving from top to bottom;
  • green trajectories for bottom to top;
  • blue trajectories for right to left;
  • red trajectories for left to right.
The pedestrian speeds are calculated as follows
v i ( t ) = x i ( t + δ t ) x i ( t ) δ t ,
where · is the Euclidean distance and δ t = 1 s is the time step, while the pedestrian’s accelerations are given by
a i ( t ) = v i ( t + δ t ) v i ( t ) δ t .
The instantaneous mean speed reads
v ¯ ( t ) = 1 | P ( t ) | i P ( t ) v i ( t ) ,
where P ( t ) is the set of pedestrians in the scene at time t, while the density is given by
ρ ( t ) = | P ( t ) | | S | ,
where | P ( t ) | is the number of pedestrians in the scene at time t and | S | is the area of the scene.
Figure A1. Extract of the trajectory dataset (recording Topview_2B). On the (left), trajectories are plotted in x–y coordinates. The (center) section details density, mean speed, individual speed, and individual acceleration sequences. On the (right), speed histograms in the different directions of motion are shown. The pedestrian dynamics are predominantly unidirectional with few obstacles.
Figure A1. Extract of the trajectory dataset (recording Topview_2B). On the (left), trajectories are plotted in x–y coordinates. The (center) section details density, mean speed, individual speed, and individual acceleration sequences. On the (right), speed histograms in the different directions of motion are shown. The pedestrian dynamics are predominantly unidirectional with few obstacles.
Sensors 24 02356 g0a1
Figure A2. Extract of the trajectory dataset (recording Topview_2c). On the (left), trajectories are plotted in x–y coordinates. The (center) section details density, mean speed, individual speed, and individual acceleration sequences. On the (right), speed histograms in the different direction of motion are shown. The pedestrian dynamics are predominantly unidirectional, while the scene includes standing pedestrians in the upper left corner, initiating avoidance behavior and queuing.
Figure A2. Extract of the trajectory dataset (recording Topview_2c). On the (left), trajectories are plotted in x–y coordinates. The (center) section details density, mean speed, individual speed, and individual acceleration sequences. On the (right), speed histograms in the different direction of motion are shown. The pedestrian dynamics are predominantly unidirectional, while the scene includes standing pedestrians in the upper left corner, initiating avoidance behavior and queuing.
Sensors 24 02356 g0a2
Figure A3. Extract of the trajectory dataset (recording Topview_2D). On the (left), trajectories are plotted in x–y coordinates. The (center) section details density, mean speed, individual speed and individual acceleration sequences. On the (right), speed histograms in the different direction of motion are shown. This recording shows counter-flow pedestrian dynamics with lane formation.
Figure A3. Extract of the trajectory dataset (recording Topview_2D). On the (left), trajectories are plotted in x–y coordinates. The (center) section details density, mean speed, individual speed and individual acceleration sequences. On the (right), speed histograms in the different direction of motion are shown. This recording shows counter-flow pedestrian dynamics with lane formation.
Sensors 24 02356 g0a3
Figure A4. Example of two pedestrian trajectory prediction scenes. In each scene, the predicted trajectories and the ground truth are displayed. On the left side, predictions are based on our 2stg. TTC-SLSTM, and, on the right side, predictions are based on SLSTM.
Figure A4. Example of two pedestrian trajectory prediction scenes. In each scene, the predicted trajectories and the ground truth are displayed. On the left side, predictions are based on our 2stg. TTC-SLSTM, and, on the right side, predictions are based on SLSTM.
Sensors 24 02356 g0a4

References

  1. Poibrenski, A.; Klusch, M.; Vozniak, I.; Müller, C. M2p3: Multimodal multi-pedestrian path prediction by self-driving cars with egocentric vision. In Proceedings of the 35th Annual ACM Symposium on Applied Computing, Brno, Czech Republic, 30 March–3 April 2020; pp. 190–197. [Google Scholar]
  2. Scheggi, S.; Aggravi, M.; Morbidi, F.; Prattichizzo, D. Cooperative human–robot haptic navigation. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014; pp. 2693–2698. [Google Scholar]
  3. Boltes, M.; Zhang, J.; Tordeux, A.; Schadschneider, A.; Seyfried, A. Empirical results of pedestrian and evacuation dynamics. Encycl. Complex. Syst. Sci. 2018, 16, 1–29. [Google Scholar]
  4. Korbmacher, R.; Tordeux, A. Review of pedestrian trajectory prediction methods: Comparing deep learning and knowledge-based approaches. IEEE Trans. Intell. Transp. Syst. 2022, 23, 24126–24144. [Google Scholar] [CrossRef]
  5. Alahi, A.; Goel, K.; Ramanathan, V.; Robicquet, A.; Li, F.-F.; Savarese, S. Social LSTM: Human trajectory prediction in crowded spaces. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 961–971. [Google Scholar]
  6. Dang, H.T.; Korbmacher, R.; Tordeux, A.; Gaudou, B.; Verstaevel, N. TTC-SLSTM: Human trajectory prediction using time-to-collision interaction energy. In Proceedings of the 2023 15th International Conference on Knowledge and Systems Engineering (KSE), Hanoi, Vietnam, 18–20 October 2023; pp. 1–6. [Google Scholar]
  7. Karamouzas, I.; Skinner, B.; Guy, S.J. Universal power law governing pedestrian interactions. Phys. Rev. Lett. 2014, 113, 238701. [Google Scholar] [CrossRef] [PubMed]
  8. Helbing, D.; Molnar, P. Social force model for pedestrian dynamics. Phys. Rev. E 1995, 51, 4282. [Google Scholar] [CrossRef] [PubMed]
  9. Van Den Berg, J.; Guy, S.J.; Lin, M.; Manocha, D. Reciprocal n-body collision avoidance. In Robotics Research: The 14th International Symposium ISRR, Lucerne, Switzerland, 31 August–3 September 2009; Springer: Berlin/Heidelberg, Germany, 2011; pp. 3–19. [Google Scholar]
  10. Burstedde, C.; Klauck, K.; Schadschneider, A.; Zittartz, J. Simulation of pedestrian dynamics using a two-dimensional cellular automaton. Phys. A Stat. Mech. Its Appl. 2001, 295, 507–525. [Google Scholar] [CrossRef]
  11. Bellomo, N.; Dogbe, C. On the modeling of traffic and crowds: A survey of models, speculations, and perspectives. SIAM Rev. 2011, 53, 409–463. [Google Scholar] [CrossRef]
  12. Chraibi, M.; Tordeux, A.; Schadschneider, A.; Seyfried, A. Modelling of pedestrian and evacuation dynamics. In Encyclopedia of Complexity and Systems Science; Springer: Berlin/Heidelberg, Germany, 2018; pp. 1–22. [Google Scholar]
  13. Duives, D.C.; Daamen, W.; Hoogendoorn, S.P. State-of-the-art crowd motion simulation models. Transp. Res. Part C Emerg. Technol. 2013, 37, 193–209. [Google Scholar] [CrossRef]
  14. Gupta, A.; Johnson, J.; Li, F.-F.; Savarese, S.; Alahi, A. Social GAN: Socially acceptable trajectories with generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2255–2264. [Google Scholar]
  15. Vemula, A.; Muelling, K.; Oh, J. Social attention: Modeling attention in human crowds. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, QLD, Australia, 21–25 May 2018; pp. 4601–4607. [Google Scholar]
  16. Monti, A.; Bertugli, A.; Calderara, S.; Cucchiara, R. Dag-net: Double attentive graph neural network for trajectory forecasting. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 2551–2558. [Google Scholar]
  17. Shi, X.; Shao, X.; Guo, Z.; Wu, G.; Zhang, H.; Shibasaki, R. Pedestrian trajectory prediction in extremely crowded scenarios. Sensors 2019, 19, 1223. [Google Scholar] [CrossRef] [PubMed]
  18. Yi, S.; Li, H.; Wang, X. Pedestrian behavior understanding and prediction with deep neural networks. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 263–279. [Google Scholar]
  19. Li, Y.; Xu, H.; Bian, M.; Xiao, J. Attention based CNN-ConvLSTM for pedestrian attribute recognition. Sensors 2020, 20, 811. [Google Scholar] [CrossRef]
  20. Yao, H.Y.; Wan, W.G.; Li, X. End-to-end pedestrian trajectory forecasting with transformer network. ISPRS Int. J. Geo-Inf. 2022, 11, 44. [Google Scholar] [CrossRef]
  21. Xue, H.; Huynh, D.Q.; Reynolds, M. Bi-prediction: Pedestrian trajectory prediction based on bidirectional LSTM classification. In Proceedings of the 2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA), Sydney, NSW, Australia, 29 November–1 December 2017; pp. 1–8. [Google Scholar]
  22. Xue, H.; Huynh, D.Q.; Reynolds, M. PoPPL: Pedestrian trajectory prediction by LSTM with automatic route class clustering. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 77–90. [Google Scholar] [CrossRef]
  23. Kothari, P.; Kreiss, S.; Alahi, A. Human trajectory forecasting in crowds: A deep learning perspective. IEEE Trans. Intell. Transp. Syst. 2021, 23, 7386–7400. [Google Scholar] [CrossRef]
  24. Papathanasopoulou, V.; Spyropoulou, I.; Perakis, H.; Gikas, V.; Andrikopoulou, E. A data-driven model for pedestrian behavior classification and trajectory prediction. IEEE Open J. Intell. Transp. Syst. 2022, 3, 328–339. [Google Scholar] [CrossRef]
  25. Khadka, A.; Remagnino, P.; Argyriou, V. Synthetic crowd and pedestrian generator for deep learning problems. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 4052–4056. [Google Scholar]
  26. Antonucci, A.; Papini, G.P.R.; Bevilacqua, P.; Palopoli, L.; Fontanelli, D. Efficient prediction of human motion for real-time robotics applications with physics-inspired neural networks. IEEE Access 2021, 10, 144–157. [Google Scholar] [CrossRef]
  27. Silvestri, M.; Lombardi, M.; Milano, M. Injecting domain knowledge in neural networks: A controlled experiment on a constrained problem. In Proceedings of the Integration of Constraint Programming, Artificial Intelligence, and Operations Research: 18th International Conference, CPAIOR 2021, Vienna, Austria, 5–8 July 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 266–282. [Google Scholar]
  28. Kothari, P.; Alahi, A. Safety-compliant generative adversarial networks for human trajectory forecasting. IEEE Trans. Intell. Transp. Syst. 2023, 24, 4251–4261. [Google Scholar] [CrossRef]
  29. Pellegrini, S.; Ess, A.; Schindler, K.; Van Gool, L. You’ll never walk alone: Modeling social behavior for multi-target tracking. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 27 September–4 October 2009; pp. 261–268. [Google Scholar]
  30. Lerner, A.; Chrysanthou, Y.; Lischinski, D. Crowds by example. In Computer Graphics Forum; Wiley Online Library: Hoboken, NJ, USA, 2007; Volume 26, pp. 655–664. [Google Scholar]
  31. Gabbana, A.; Toschi, F.; Ross, P.; Haans, A.; Corbetta, A. Fluctuations in pedestrian dynamics routing choices. PNAS Nexus 2022, 1, pgac169. [Google Scholar] [CrossRef] [PubMed]
  32. Robicquet, A.; Sadeghian, A.; Alahi, A.; Savarese, S. Learning social etiquette: Human trajectory understanding in crowded scenes. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 549–565. [Google Scholar]
  33. Zhou, B.; Wang, X.; Tang, X. Understanding collective crowd behaviors: Learning a mixture model of dynamic pedestrian-agents. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2871–2878. [Google Scholar]
  34. Majecka, B. Statistical Models of Pedestrian Behaviour in the Forum. Master’s Thesis, University of Edinburgh, Edinburgh, UK, 2009. [Google Scholar]
  35. Boltes, M.; Seyfried, A. Collecting pedestrian trajectories. Neurocomputing 2013, 100, 127–133. [Google Scholar] [CrossRef]
  36. Korbmacher, R.; Dang, H.T.; Tordeux, A. Predicting pedestrian trajectories at different densities: A multi-criteria empirical analysis. Phys. A Stat. Mech. Its Appl. 2024, 634, 129440. [Google Scholar] [CrossRef]
  37. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  38. Korbmacher, R.; Dang-Huu, T.; Tordeux, A.; Verstaevel, N.; Gaudou, B. Differences in pedestrian trajectory predictions for high-and low-density situations. In Proceedings of the 14th International Conference on Traffic and Granular Flow (TGF) 2022, New Delhi, India, 15–17 October 2022. [Google Scholar]
  39. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 2672–2680. [Google Scholar]
  40. Holl, S. Methoden für die Bemessung der Leistungsfähigkeit Multidirektional Genutzter Fußverkehrsanlagen; FZJ-2017-00069; Jülich Supercomputing Center: Jülich, Germany, 2016. [Google Scholar]
  41. Wirth, T.D.; Dachner, G.C.; Rio, K.W.; Warren, W.H. Is the neighborhood of interaction in human crowds metric, topological, or visual? PNAS Nexus 2023, 2, pgad118. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Area for the trajectory tracking at Lyon’s Festival of Lights 2022.
Figure 1. Area for the trajectory tracking at Lyon’s Festival of Lights 2022.
Sensors 24 02356 g001
Figure 2. Results of the K-Means and the agglomerative hierarchical clustering. Trajectory scenes are clustered as shown by the different colors of the points.
Figure 2. Results of the K-Means and the agglomerative hierarchical clustering. Trajectory scenes are clustered as shown by the different colors of the points.
Sensors 24 02356 g002
Figure 3. Schemata of our two-stage prediction approach.
Figure 3. Schemata of our two-stage prediction approach.
Sensors 24 02356 g003
Figure 4. Examples for each of the four density levels.
Figure 4. Examples for each of the four density levels.
Sensors 24 02356 g004
Figure 5. ADE and COL metrics for the two-stage SLSTM algorithm according to the collision weight λ for the four density levels.
Figure 5. ADE and COL metrics for the two-stage SLSTM algorithm according to the collision weight λ for the four density levels.
Sensors 24 02356 g005
Table 1. Quantitative comparison of ADE, FDE and COL prediction metrics for classical algorithms borrowed from the literature and our 2-stage approaches on the Festival of Lights dataset.
Table 1. Quantitative comparison of ADE, FDE and COL prediction metrics for classical algorithms borrowed from the literature and our 2-stage approaches on the Festival of Lights dataset.
 ModelLowDMediumDHighDVeryHD
ADE/FDECOLADE/FDE COLADE/FDECOLADE/FDECOL
CV0.71/0.9754.760.85/0.9845.730.53/0.862.350.44/0.6781.74
Social Force [8]0.78/1.3324.40.55/0.8931.160.5/0.8236.430.36/0.6354.78
Vanilla LSTM0.5/0.9931.550.33/0.6337.690.29/0.5236.430.24/0.4163.8
Social LSTM [5]0.53/1.0257.740.37/0.7359.30.41/0.7864.260.35/0.6675.37
Social GAN [14]0.53/0.9931.360.39/0.7232.160.36/0.6132.330.25/0.4155.94
Our 2stg. SLSTM0.48/0.9330.950.3/0.6336.180.26/0.442.020.24/0.4152.23
Our 2stg. SGAN0.44/0.8332.740.27/0.5240.20.28/0.535.330.26/0.4358.6
Our 2stg. TTC-SLSTM0.39/0.7329.170.3/0.6222.610.23/0.3636.290.24/0.4152.23
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Korbmacher, R.; Tordeux, A. Toward Better Pedestrian Trajectory Predictions: The Role of Density and Time-to-Collision in Hybrid Deep-Learning Algorithms. Sensors 2024, 24, 2356. https://doi.org/10.3390/s24072356

AMA Style

Korbmacher R, Tordeux A. Toward Better Pedestrian Trajectory Predictions: The Role of Density and Time-to-Collision in Hybrid Deep-Learning Algorithms. Sensors. 2024; 24(7):2356. https://doi.org/10.3390/s24072356

Chicago/Turabian Style

Korbmacher, Raphael, and Antoine Tordeux. 2024. "Toward Better Pedestrian Trajectory Predictions: The Role of Density and Time-to-Collision in Hybrid Deep-Learning Algorithms" Sensors 24, no. 7: 2356. https://doi.org/10.3390/s24072356

APA Style

Korbmacher, R., & Tordeux, A. (2024). Toward Better Pedestrian Trajectory Predictions: The Role of Density and Time-to-Collision in Hybrid Deep-Learning Algorithms. Sensors, 24(7), 2356. https://doi.org/10.3390/s24072356

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop