Next Article in Journal
Identifying the Regions of a Space with the Self-Parameterized Recursively Assessed Decomposition Algorithm (SPRADA)
Next Article in Special Issue
Gradient-Based Neural Architecture Search: A Comprehensive Evaluation
Previous Article in Journal
Alternative Formulations of Decision Rule Learning from Neural Networks
Previous Article in Special Issue
Autoencoder Feature Residuals for Network Intrusion Detection: One-Class Pretraining for Improved Performance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Behavior-Aware Pedestrian Trajectory Prediction in Ego-Centric Camera Views with Spatio-Temporal Ego-Motion Estimation †

1
Perception & Maps Department, Mercedes-Benz AG, 71063 Sindelfingen, Germany
2
Institute of Signal Processing and System Theory, University of Stuttgart, 70550 Stuttgart, Germany
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published in the Proceedings of the 21st IEEE International Conference on Machine Learning and Applications (ICMLA), Nassau, Bahamas, 12–14 December 2022.
Mach. Learn. Knowl. Extr. 2023, 5(3), 957-978; https://doi.org/10.3390/make5030050
Submission received: 14 June 2023 / Revised: 12 July 2023 / Accepted: 1 August 2023 / Published: 3 August 2023
(This article belongs to the Special Issue Deep Learning and Applications)

Abstract

:
With the ongoing development of automated driving systems, the crucial task of predicting pedestrian behavior is attracting growing attention. The prediction of future pedestrian trajectories from the ego-vehicle camera perspective is particularly challenging due to the dynamically changing scene. Therefore, we present Behavior-Aware Pedestrian Trajectory Prediction (BA-PTP), a novel approach to pedestrian trajectory prediction for ego-centric camera views. It incorporates behavioral features extracted from real-world traffic scene observations such as the body and head orientation of pedestrians, as well as their pose, in addition to positional information from body and head bounding boxes. For each input modality, we employed independent encoding streams that are combined through a modality attention mechanism. To account for the ego-motion of the camera in an ego-centric view, we introduced Spatio-Temporal Ego-Motion Module (STEMM), a novel approach to ego-motion prediction. Compared to the related works, it utilizes spatial goal points of the ego-vehicle that are sampled from its intended route. We experimentally validated the effectiveness of our approach using two datasets for pedestrian behavior prediction in urban traffic scenes. Based on ablation studies, we show the advantages of incorporating different behavioral features for pedestrian trajectory prediction in the image plane. Moreover, we demonstrate the benefit of integrating STEMM into our pedestrian trajectory prediction method, BA-PTP. BA-PTP achieves state-of-the-art performance on the PIE dataset, outperforming prior work by 7% in MSE-1.5 s and CMSE as well as 9% in CFMSE.

1. Introduction

One of the essential challenges for automated driving in urban traffic conditions is the task of pedestrian behavior prediction. Therefore, understanding the underlying intentions of pedestrians is crucial for automated vehicles to better understand their surroundings in order to make better and safer driving decisions and mitigate potential risks and hazardous scenarios [1]. Due to the inherently variable nature of pedestrians’ behavior, solving the task of predicting this behavior involves numerous challenges. Nonetheless, human drivers are able to anticipate the future actions of pedestrians by relying on visual observations and interpreting their behavioral cues. Inferring the potential future movement direction of pedestrians based on their body or head orientation and responding appropriately is a suitable example.
In this work, we focused on the problem of pedestrian trajectory prediction from an ego-centric perspective captured by a camera mounted on board a vehicle, e.g., behind the windshield. Hereby, the objective is to predict the future bounding boxes in the image plane. To accomplish this, it is necessary to extract rich information from camera-based observations within a remarkably dynamic environment. The majority of current state-of-the-art approaches for predicting future bounding boxes heavily depend on the past positions of pedestrians. These are utilized to encode their motion history [2,3] or condition the predictions based on estimated inherent goals [4,5]. Taking into account the past trajectory of pedestrians is one of the most important features for this prediction task and is particularly advantageous in scenarios where they exhibit linear motion patterns. However, pedestrians consistently adapt their intended paths due to the influence of the constantly evolving environment surrounding them. For example, a pedestrian may all of a sudden decide to cross the street after standing on the sidewalk for some time. In such scenarios, additional information beyond just the past trajectory is required to accurately predict the future path of pedestrians, especially in ego-centric camera views where the scene dynamically changes due to the ego-motion of the vehicle. By observing pedestrians from the perspective of an ego-vehicle camera, it becomes possible to capture behavioral features that provide valuable cues about their intended movement. These cues can be extracted from visual observations and explicitly modeled through features like the body and head orientation as well as their pose. Despite that, behavioral features inferred from visual observations are only used in a few works [6,7,8,9] to predict pedestrians’ future bounding boxes.
In this work, we proposed a pedestrian trajectory prediction method for ego-centric camera systems that uses behavioral features such as body orientation and head orientation, as well as pedestrian’s pose. The different input modalities are processed by independent encoding streams and the resulting encodings are fused to better model the motion history of pedestrians. By visually observing pedestrians, we make use of the cues they provide about their intended movements and incorporate them explicitly into our method. We refer to the method as Behavior-Aware Pedestrian Trajectory Prediction (BA-PTP). An overview of our proposed method is shown in Figure 1.
When predicting pedestrian trajectories in the image plane, there is a need to compensate for the ego-motion of the camera in this dynamically changing scene. Current approaches tackle this by incorporating estimated odometry information of the ego-vehicle into the trajectory prediction [2,3,10]. However, these approaches solely rely on past information without utilizing the future intended route of the ego-vehicle.
In contrast to that, we integrated a module into BA-PTP, which leverages spatial goal points from the intended route of the ego-vehicle for predicting its future ego-motion, called Spatio-Temporal Ego-Motion Module (STEMM). This route can be provided, e.g., by a navigation system. STEMM follows a similar scheme as the bounding box prediction, whereas it encodes the ego-vehicle’s past odometry and utilizes the goal points to predict future ego-motion. The predictions are used by BA-PTP to predict future pedestrian trajectories.
Contributions: (1) We propose BA-PTP, a novel method for pedestrian trajectory prediction in the image plane, which incorporates behavioral features of pedestrians using independent encoding streams. (2) We introduce STEMM, a novel approach for predicting the future ego-motion of the ego-vehicle using spatial goal points, which is integrated into BA-PTP. (3) We performed a detailed analysis of BA-PTP on two pedestrian behavior datasets, demonstrating the advantages of incorporating behavioral features as well as STEMM into a pedestrian trajectory prediction model. (4) Based on ablation studies, we evaluated the effectiveness of our proposed encoding strategy and explored the impact of various behavioral features on the performance of prediction.
Preliminary results of this research are published in a conference paper at ICMLA 2022 [11]. In this work, we extended this in the following ways:
  • We improved the performance of the BA-PTP method from [11] by focusing on a more stable encoding of pedestrian trajectories in the image plane. We did this by first incorporating information about the head position of pedestrians into our method. Compared to the body bounding boxes of a pedestrian, this modality is not prone to changes in shape. Second, by adapting the bounding box representation compared to our preliminary work.
  • To compensate for the ego-motion in pedestrian trajectory prediction from the ego-vehicle camera perspective, we integrated STEMM into BA-PTP. Compared to our previous work, this replaces the use of precise temporal future ego-vehicle odometry information. In a real-world application with unknown future odometry, spatial goal points may be extracted from the intended route provided by a navigation system.
  • BA-PTP achieves a state-of-the-art performance on the PIE dataset, outperforming prior work in MSE-1.5s, CMSE, and CFMSE.
This work is organized as follows: It begins with reviewing related works in Section 2. In Section 3, we present BA-PTP and STEMM as an approach to predicting future trajectories of pedestrians in the image plane. Then, we show experimental results and ablation studies in Section 4, before discussing our findings and elaborating on future research directions in Section 5. Finally, we conclude this work in Section 6.

2. Related Work

The field of pedestrian behavior prediction has witnessed increasing research attention in recent times. Predicting pedestrians’ future behavior is approached through two distinct modeling methods, which differ in terms of their output. The first approach involves the prediction or classification of the future action of a pedestrian, i.e., whether he will cross the street in the near future, often referred to as intention prediction [7,12,13,14,15,16,17,18,19,20,21,22,23,24,25]. In the second approach, the focus lies on predicting the future trajectory of a pedestrian for a defined prediction horizon [2,3,4,5,6,7,10,14,26,27,28,29,30,31,32,33,34,35]. The former approach primarily provides insights solely into a potential crossing action occurring at some point in the future, while the latter predicts positional information that can be utilized to derive the crossing action as well.
Moreover, methods working on the problem of trajectory prediction can be distinguished by the perspective in which they operate. Pedestrian trajectory prediction from the ego-vehicle camera perspective observes the environment from a moving first-person view. This results in a dynamically changing scene, because the relative position and size of observed objects depend on the ego-motion of the camera, as the ego-vehicle is also moving. To address this, some methods try to compensate for the ego-motion of the camera by incorporating ego-vehicle information [2,3,6,8,9,10,14,22,34,35].
Most of these approaches try to tackle this problem by predicting the future odometry of the ego-vehicle [2,3,6]. They first encode the history of the ego-vehicle’s odometry using a Recurrent Neural Network (RNN) encoder. Sometimes, additional information, such as action and interaction priors [10] or visual features [3], is also incorporated into the encoding. Subsequently, the encoded data are fed into an RNN decoder for predicting the future ego-vehicle odometry. Similar to the aforementioned approaches, ego-motion compensation is also performed by simply encoding the observed ego-vehicle’s odometry [8,9,10,14,22] or combining this with optical flow representations [35]. Ref. [34] proposed a method that explicitly disentangles the motion of pedestrians and the ego-motion of the vehicle. This is accomplished by using an ego-motion prediction network to observe and predict pedestrian behavior from an ego-motion compensated viewpoint.
When predicting the trajectory from a bird’s eye perspective, the environment is observed from a top-down view, where the positions of objects are represented by global coordinates and are not dependent on the ego-motion of the camera. This perspective allows for better modeling of interactions [29,30,31,32] as relative distances between objects can be inferred from the global coordinates. Recently, this has also been investigated in the ego-centric camera view domain [20,36,37]. These methods try to model the relationships between the target pedestrian and nearby objects in the spatial and temporal domain by constructing scene graphs using object locations in the image plane.
Nowadays, recurrent architectures [2,3,15], as well as CVAEs [4,5,26], are prevalent in state-of-the-art pedestrian behavior prediction and are used in the majority of recent works. Other approaches model spatio-temporal features by first extracting features using CNNs [10,12,15] or graph structures [6,13,20] and processing them later on with RNNs. Recently, partially attention-based [2,15,23,24] transformers [7,8,25,35,38], as well as goal-driven [4,5,9,33] approaches, have also gained growing interest.
Related work may also be differentiated by the information used as input to predict the future behavior of pedestrians. The pedestrian’s motion history is the most fundamental information as almost all methods rely on the past motion of the pedestrian independent of the domain [2,3,4,5,6,7,8,9,10,12,13,14,15,16,22,23,24,25,26,27,28,29,30,31,32,33,34,36,37], This information includes bounding boxes for ego-centric camera view methods or global coordinates for methods in bird’s eye view. Additionally, features like the distance to the ego-vehicle [26,27], semantic and contextual information [7,13,14,20,23,24,26,27,31,32,36], as well as visual or appearance features [2,3,6,10,12,15,16,17,19,20,23,24,25,34,36,37], are used as inputs to encode the past motion and behavior of pedestrians.
Behavioral cues such as body or head orientation [7,8,13,26,28], awareness [8,17,21,27], pose of the pedestrian [6,8,9,15,18,19,22,23,24], and pedestrians’ intention as the wish to cross the street [2,6,16] are also used to better understand and model pedestrians’ behavior.
However, in the domain of pedestrian trajectory prediction from the ego-vehicle camera perspective, such behavioral features are rarely used. [6] used pedestrians’ poses to improve the prediction of their intention module, resulting in only a marginal benefit for the trajectory prediction on PIE [2]. The body orientation of pedestrians was used in [7] to simultaneously predict intention and trajectory, but only benefits intention prediction, whereas this reduces trajectory prediction performance.
After the initial submission of our preliminary work [11], two further related works have been published [8,9]. Ref. [8] proposed an approach, where different modalities—namely pedestrian bounding boxes, their poses, and the ego-motion of the vehicle—are encoded using crossmodal transformers and concatenated with an embedding of a pedestrian attribute vector, which contains the pedestrian’s body orientation, bounding box and its awareness of the ego-vehicle as well as the vehicle ego-motion of the last observed timestep. This attribute vector should not only encode the basic state of the pedestrian but also implicitly the relationship between the pedestrian and the ego-vehicle [8]. Pedestrian poses are also used in the goal-driven approach of [9] in addition to bounding boxes and the ego-vehicle’s motion. Moreover, the authors made use of ground truth future odometry information to compensate for the ego-motion of the vehicle.
In contrast to the methods published before our preliminary work, we show how to incorporate behavioral features, such as the body and head orientation of pedestrians, as well as their pose, to benefit the prediction of future trajectories in the image plane. Further, we included information about the head position of pedestrians in our method to better encode the trajectories of pedestrians. We used independent encoding streams for each input modality and fused the learned embeddings with an attention mechanism to better encode the motion history of pedestrians and predict more accurate future trajectories. Additionally, we integrated a module that predicts the camera’s ego-motion into our trajectory prediction approach to account for the dynamically changing scene in ego-centric views. Compared to previous works, we utilized spatial goal points, that were sampled from the intended route of the ego-vehicle. Table 1 demonstrates a characteristic comparison of our approach against prior works for pedestrian trajectory prediction in ego-centric camera views, including our preliminary work [11].

3. Method

We define the task of forecasting pedestrian behavior as the prediction of future trajectories in terms of image positions. This prediction is based on real-world traffic scene observations obtained from the ego-vehicle camera perspective. Hereby, pedestrians’ trajectories are represented as a sequence of body bounding boxes in the image plane. However, these sequences of body bounding boxes are prone to fluctuations in their box widths due to the natural gait cycle of humans, which can be referred to as box wobbling. Figure 2 shows exemplary trajectories of (crossing) pedestrians, which show how the bounding box width changes during the observation. To overcome this downside and get a more stable representation of pedestrian trajectories, we proposed to also include information about the head position of pedestrians. Head bounding boxes do not have the property of changing shapes and can thus help to better encode the motion history of pedestrians and reduce the box wobbling effects. This can also be seen in Figure 2.
Bounding boxes are usually represented by either the top-left and bottom-right corners in pixel coordinates in the image b b = { x t l , y t l , x b r , y b r } , or the center coordinate as well as width and height in pixels: b b = { c x , c y , w , h } . Our previous publication [11] used the first representation. Here, we switched to the latter representation to further reduce box wobbling effects.
We aimed to predict the future trajectory Traj f u t = { b b i t + 1 , , b b i t + m } of a pedestrian i at a certain timestep t for a prediction horizon m, given the pedestrian’s observed past trajectory represented by body bounding boxes B B o b s = { b b i t n + 1 , , b b i t } and head bounding boxes H B o b s = { h b i t n + 1 , , h b i t } for an observation horizon of length n, as well as observed behavioral features of the pedestrian:
  • Body orientation B O o b s = { b o i t n + 1 , , b o i t } , where b o R ;
  • Head orientation H O o b s = { h o i t n + 1 , , h o i t } , where h o R ;
  • Pose P o b s = { p i t n + 1 , , p i t } , where p R 34 defines the pixel coordinates of a 17-joint skeleton.
Additionally, we made use of the predicted future ego-motion of the ego-vehicle E M f u t = { e m t + 1 , , e m t + m } , where e m t + j R d e m is a vector with dimensionality d e m . The ego-motion vectors are predicted by a separate module, which is introduced in Section 3.1.2. In different experiments, the predicted ego-motion of this module is either speed and yaw rate (i.e., d e m = 2 ) or an arbitrary feature embedding.

3.1. Architecture

Pedestrians implicitly offer numerous cues regarding their intended movement as they walk along a street. These cues can be inferred through visual observations and expressed as behavioral features, such as the body and head orientation of pedestrians as well as their pose. To account for this, we proposed Behavior-Aware Pedestrian Trajectory Prediction (BA-PTP), a method that uses this information as explicit features in addition to their motion history. Furthermore, our novel approach for ego-motion prediction, called Spatio-Temporal Ego-Motion Module (STEMM), provides predicted future ego-motion information to be utilized by BA-PTP.
Initially, we present our BA-PTP architecture, followed by the introduction of STEMM.

3.1.1. Pedestrian Trajectory Prediction

As depicted in Figure 3, BA-PTP consists of an RNN encoder-decoder architecture, where we process each input modality through an independent encoder, namely a body bounding box stream, a head bounding box stream, a pose stream, a head orientation stream, and a body orientation stream.
The encoding scheme of BA-PTP is inspired by [15]. We applied the same structure for each encoding stream e as follows: The RNN encoder processes the input sequence and produces hidden states for each observed timestep. We then applied the temporal attention mechanism from [15] on the hidden states. The role of the attention mechanism is to give higher importance to particular timesteps in the observation horizon compared to other ones, focusing on the most relevant information. The resulting embedding vectors e e of each encoding stream are concatenated and the modality attention module from [15] produces the final embedding vector e p e d by assigning importance to the different modality inputs.
The final embedding created by our encoding module contains a latent representation of the motion history of the pedestrian, which can be used to predict the future trajectory. To get a joint representation as input to the trajectory predictor, we concatenated the embedding vector e p e d with ego-motion information e m t + j for every future timestep t + j . We passed this through a self-attention unit from [2] to target the features of the encoding that are most relevant for the present prediction task. We then used this as the decoder inputs, resulting in hidden states for every future timestep in the prediction horizon. To generate the final predictions, we passed the decoder outputs through a dense layer that regresses the hidden states for every future timestep into bounding box predictions Traj f u t = { b b i t + 1 , , b b i t + m } .
We calculated the decoder’s initial state vector s t a t e p e d by taking the elementwise maximum (max-pooling [39]) of the final hidden states h e t of all encoding streams, with e being the index of an encoding stream:
s t a t e p e d = max 1 j d h e j t ,
where d is the dimensionality of the hidden states. We found this to result in the best performance.

3.1.2. Ego-Motion Prediction

In this work, we proposed Spatio-Temporal Ego-Motion Module (STEMM), a novel approach for predicting the future ego-motion of the ego-vehicle. The main contribution of STEMM compared to other approaches for odometry prediction, as described in Section 2, is the incorporation of spatial goal points of the ego-vehicle. These goal points represent target locations along the route to be followed, e.g., provided by a navigation system. This approach is inspired by recent methods for planning in the context of autonomous driving that also use goal points to describe the driver intention or the route to be followed [40,41,42]. For STEMM, the goal points should provide priors about where the ego-vehicle might move to better predict the future odometry that follows the intended path. Inspired by [41], a goal point g = { g x , g y , g y a w } is represented by the x- and y-coordinates, g x and g y , as well as the yaw angle g y a w of the ego-vehicle in a bird’s eye view and ego-centric coordinate system. As the used datasets in this work do not provide the intended route from the navigation system itself, we approximate it using odometry information. Details are described in Section 4.2.
Goal points are sampled using a simple yet effective approach. Due to the fact that we do not have precise temporal future odometry information at present, last observed, timestep t, we sample spatial future goal points from the intended route to be followed by the ego-vehicle. We utilized the present velocity of the ego-vehicle to estimate its distance traveled during the prediction horizon m, assuming constant velocity to calculate the spatial distance d i s t g t + j for a future timestep t + j . We obtain g t + j by extracting the point with spatial distance d i s t g t + j from the intended route of the ego-vehicle. In total, we sample m goal points G = { g t + 1 , , g t + m } , one for each future timestep. STEMM is named for its use of spatial information to predict temporal features.
In addition to the set of goal points G, the input to STEMM is the ego-vehicle odometry information for the n observed timesteps:
  • Speed S o b s = { s t n + 1 , , s t } ;
  • Yaw rate Y o b s = { y t n + 1 , , y t } ;
  • Past ego-vehicle positions P O S o b s = { p o s t n + 1 , , p o s t } , with p o s = { p o s x , p o s y , p o s y a w } similar to goal point g.
STEMM predicts the future ego-motion of the ego-vehicle E M f u t = { e m t + 1 , , e m t + m } , where the ego-motion vector e m t + j R d e m for timestep t + j has dimensionality d e m .
STEMM follows a similar scheme as the bounding box prediction, whereas the ego-vehicle’s odometry history is encoded using a speed stream, a yaw rate stream, and a past ego-vehicle position stream. The architecture of STEMM is illustrated in the lower part of Figure 3. The final embedding e v e h is concatenated with the future goal points g t + j for every future timestep t + j and is used as inputs to the decoder. The decoder then outputs the future ego-motion vectors e m t + j for every future timestep t + j .
We investigated two variants of STEMM, which differ in their final output. In the first variant, STEMM outputs the ego-vehicle’s future odometry, i.e., the speed and yaw rate for every future timestep: e m t + j = ( s t + j , y t + j ) . Thus, we set the dimensionality d e m = 2 for the future ego-motion vector e m t + j R d e m . Hereby, STEMM directly replaces the use of precise temporal future odometry information, which has been used as input in our preliminary work [11]. Moreover, STEMM is trained separately in a supervised manner on the true future odometry. In the second variant, STEMM outputs a feature embedding of arbitrary size, which is directly used by BA-PTP. Hereby, BA-PTP and STEMM have to be trained jointly end-to-end and no auxiliary supervised loss on the ego-motion output e m t + j is applied. We evaluated both variants in Section 4.

3.2. Loss Function

Our method is trained in a supervised manner. As the loss function for training BA-PTP, we use the root mean squared error (RMSE) between the predicted bounding boxes b ^ and the ground truth b for each of the I training samples with prediction horizon m:
RMSE = 1 I i = 1 I j = 1 m b i t + j b ^ i t + j 2 .
For training of the first variant of STEMM, we also used RMSE. However, we applied separate loss functions to the outputs of STEMM, i.e., speed and yaw rate, with λ being the loss weight for the yaw rate term:
L STEMM = RMSE speed + λ RMSE yawrate .

4. Experiments

In this Section, we evaluated our method BA-PTP on two datasets for pedestrian behavior prediction. We investigated which behavioral features contribute most to the trajectory prediction performance and verified the effectiveness of our independent encoding strategy based on ablation studies. Moreover, we demonstrated the benefit of integrating STEMM into a trajectory prediction method by comparing BA-PTP with state-of-the-art methods.

4.1. Datasets

We performed experiments on two different datasets:
1.
PIE [2]
PIE contains on-board camera videos recorded at 30 fps with a resolution of 1920 × 1080 px during daytime in Toronto, Canada. Annotations include pedestrian bounding boxes with track IDs and occlusion flags, as well as vehicle information such as speed and yaw from an on-board diagnostics sensor. The dataset contains 1842 pedestrians, divided into train, validation, and test sets by ratios of 50 % , 10 % and 40 % , respectively.
2.
ECP-Intention
The ECP-Intention dataset was created by selecting recordings collected for the EuroCity Persons (ECP) dataset [43]. The data were recorded with a two-megapixel camera (1920 × 1024 px) mounted behind the windshield at 20 fps in 31 different cities in Europe. Two-hundred-and-twenty-seven sequences from 27 cities with an average length of 16 s were selected. Each sequence was manually labeled at 5 fps (every 4th image), providing pedestrian bounding boxes with track IDs and occlusion/truncation flags, as well as head bounding boxes. Additionally, for all pedestrians, their body and head orientation relative to the line of sight of the camera in the range [ 0 , 360 ) was annotated. A value of 0 corresponds to an orientation directly towards the camera. For each image, odometry information is provided, captured by the vehicle’s sensors and an additional IMU/GNSS sensor: the vehicle’s speed and yaw rate.
Overall, ECP-Intention consists of 17,299 manually labeled images with 133,030 pedestrian bounding boxes containing 6344 unique pedestrian trajectories. The data were split into train, validation, and test subsets by ratios of 55%, 10%, and 35%, respectively. The dataset has not been published yet.
Since there were images at 20 fps available for ECP-Intention, we extended the 5 fps manual annotations to 20 fps by linear interpolation between two hand-labeled frames, called keyframes. This increased the amount of data in the observation horizon and enabled us to better train our method since it can rely on more information for encoding the motion history. In addition to that, we can generate more data samples.

4.2. Sampling of Future Goal Points

Section 3.1.2 explained how future goal points are sampled from the intended route of the ego-vehicle based on a constant velocity model. ECP-Intention as well as PIE do not provide data on the intended route from the navigation system itself. Therefore, we calculated the driven route based on the odometry information available in both datasets as an approximation of the intended route. This driven route only provides spatial information for goal point sampling. Precise temporal information was discarded.

4.3. Implementation

We used Gated Recurrent Units (GRUs) with 256 hidden units and tanh activation for the encoding streams as well as for the decoder. Observation (n) and prediction horizon (m) were set based on the dataset used: For PIE, we followed [2] and used 0.5 s of observations ( n = 15 ) and predicted 1.5 s ( m = 45 ). We sampled tracks with a sliding window approach and a stride of 7. For ECP-Intention, we observed 0.6 s ( n = 13 ) and predicted 1.6 s ( m = 32 ). Tracks were sampled with a stride of 2. If there were missing data frames within a sequence of the same pedestrian (e.g., due to occlusion), we split the sequence into multiple tracks to ensure consecutive tracks during training and testing. We trained every model for 100 epochs using the Adam [44] optimizer with a batch size of 128 and L 2 regularization of 5 × 10 3 for ECP-Intention and L 2 regularization of 10 3 for PIE. After each temporal attention block for both BA-PTP and STEMM, we used a dropout of 0.5 , following [15]. If not stated otherwise, the initial learning rate for ECP-Intention and PIE was set to 5 × 10 4 and 10 3 , respectively. If the validation loss had not improved for five epochs during training, the learning rate was reduced by a factor of 5.
Since we did not have annotated poses for ECP-Intention and no annotations for behavioral features at all for PIE, we needed to generate the data beforehand. To infer the pose and body orientation of pedestrians, we used the following architecture: For the detection and pose estimation of pedestrians, we used VRU-Pose SSD [45], extended by the orientation loss used in Pose-RCNN [46]. In this way, we simultaneously predicted the pose and body orientation of pedestrians. The model was trained on a combination of three datasets: ECPDP [47], TDUP [48], and additional non-public data. To associate these detections with the ground truth boxes of ECP-Intention and PIE, we matched every ground truth bounding box with the detection with the highest Intersection over Union (IoU). Missing attribute values are ignored by our method using a masking layer.

4.4. Data Preparation

For the trajectory inputs B B o b s and H B o b s , we used the ground truth bounding boxes. The reason for that is to make the trajectory prediction independent of a pedestrian detector’s performance as in the majority of previous works [2,4,5,10,14]. We normalized B B o b s and H B o b s by subtracting the first bounding box from the whole track, i.e., converting the absolute to relative pixel coordinates, as is the case in [2].
The pedestrian’s pose p i was normalized by dividing each x-coordinate by the image width and each y-coordinate by the image height. The values for body and head orientation wee normalized to the range [ 0 , 1 ) . For odometry information, we did not apply any pre-processing.
For normalization of the goal points for STEMM, we used the ego-centric coordinate system of the ego-vehicle at the final observed timestep t as the reference system. We obtained goal points within that reference system using common coordinate transformations. We used the front-left-up convention, meaning the x-axis points towards the front for the ego-vehicle. Therefore, the y-coordinate of the goal points was zero, as long as the intended route was straight. For normalization of the past ego-vehicle positions, we simply used the ego-centric coordinate system at the first observed timestep t n .

4.5. Metrics

We used the standard metrics [2,3,4] for evaluation on both datasets, PIE and ECP-Intention: the mean squared error (MSE) of the bounding boxes’ upper-left and lower-right corner was calculated over the different prediction horizons as well as the MSE of the bounding box centers over the whole prediction horizon (CMSE) and only for the final timestep (CFMSE). For ECP-Intention, we calculated the metrics only for the keyframes to evaluate the performance only on the hand-labeled data. To this end, we restricted test samples to end with keyframes during data generation enabling us to measure prediction performance exactly 1.6 s into the future.
We averaged the trajectory prediction results of all models over four different experiment runs. In the following sections, we report the average performance with standard deviation in integer pixel errors by mathematical rounding.

4.6. Results on the ECP-Intention Dataset

In this subsection, we performed experiments on the ECP-Intention dataset. Initially, we investigated the relevance of different behavioral features. In the following, we demonstrate the effectiveness of independently encoding multiple input features. Finally, we evaluated the benefit of integrating STEMM into BA-PTP and compared it with the state-of-the-art methods.

4.6.1. Relevance of Different Behavioral Features

We performed an ablation study on ECP-Intention to investigate which behavioral features contribute most to the prediction performance. In order to isolate the effect of specific behavioral features, we decoupled the camera’s ego-motion from the trajectory prediction task by utilizing ground truth future odometry, i.e., speed and yaw rate, for these experiments instead of inferring the future ego-motion using STEMM. We compared our results to PIEtraj+speed, which is an extension of PIEtraj [2] that additionally uses ground truth ego-vehicle speed information. The results of our method when we use only a subset of the behavioral features are shown in Table 2.
One initial finding is that our method outperforms PIEtraj+speed on all metrics by a significant margin even without incorporating any behavioral features into the prediction process (BA-PTPBB). We attribute this to the utilization of the yaw rate of the ego-vehicle, which PIEtraj+speed does not use, particularly enhancing prediction performance in turning scenarios. A performance comparison with our model that excludes the use of the yaw rate (BA-PTPBB-Y) validates this assumption. Upon incorporating the yaw rate to our model, we observed a notable performance increase of 38% for MSE-0.8 s and up to 55% for longer-term predictions (MSE-1.6 s). These findings suggest that integrating the ego-vehicle’s yaw rate is crucial for effectively accounting for the ego-motion of the camera in the sequences of ECP-Intention.
Furthermore, we observed that adding the head bounding boxes of pedestrians, in addition to their body bounding boxes, is beneficial. Adding this information to BA-PTP results in an increase of performance by 12–13% across all metrics (BA-PTPBB+HB). This confirms our motivation of adding head bounding boxes to obtain a more stable encoding of pedestrian trajectories and to reduce the box wobbling effects. In the rest of this work, we refer to the combination of body and head bounding boxes as the past trajectory.
Rows 5–7 illustrate the performance variation when utilizing only one of the behavioral features (BO, HO, P) alongside the past trajectory. Every individual feature contributes to the enhancement of trajectory prediction. Among the behavioral features, the pose shows the greatest impact on the prediction performance, resulting in an improvement of 6% for MSE-0.8 s and 15% for MSE-1.6 s when compared to solely using the past trajectory. Comparing the improvements in short-term (0.8 s) and long-term (1.6 s) predictions indicates that the advantage of incorporating behavioral features grows as the prediction horizon becomes longer.
The results of our method when integrating multiple behavioral features in combination are shown in rows 8–11 and allow us to derive several conclusions. The combination of two behavioral features yields an improved performance compared to using a single additional feature, suggesting a complementary relationship between these features. Utilizing all behavioral features, i.e., body orientation, head orientation, and pose (BA-PTPBB+HB+BO+HO+P) attains the most accurate prediction of future trajectories of pedestrians on the ECP-Intention dataset. The prediction performance increases by 23% in MSE-1.6 s, 24% in CMSE, and 28% in CFMSE compared to only using the past trajectory. This demonstrates the advantage of incorporating behavioral features of pedestrians in pedestrian trajectory prediction.
Experiments with using inferred body orientation instead of the ground truth annotations as input for both training and inference are shown in the last two rows of Table 2. We noticed a slight performance decrease across all metrics when compared to the models using ground truth annotations as input, which can be explained by the introduced noise from using inferred values. Nevertheless, notable improvements are observed in comparison to solely relying on bounding boxes.

4.6.2. Effect of Independent Encoding Streams

We conducted a comparison between our attention-based modality fusion (Independent) and a variant of the model in which all input features were concatenated first and processed using a single encoding stream (Concat) in Table 3. Hence, this model variant solely relies on temporal attention and does not incorporate modality attention. Also, we did not change the number of hidden units for this ablation. The only additional change we made is that we initialized the decoder’s initial state with zeros, which we found to result in the best performance for the Concat encoding strategy. We show this ablation study for two models: BA-PTPBB+HB+BO+P and BA-PTPBB+HB+BO+HO+P.
Our Independent encoding strategy outperforms the Concat encoding strategy in both models, yielding superior results. For example, for BA-PTPBB+HB+BO+HO+P MSE-1.6 s, CMSE and CFMSE improve by 38%, 40%, and 34%, respectively. Moreover, for the two models using the Concat encoding strategy, the additional use of the head orientation does not improve the prediction performance in contrast to the Independent encoding strategy. This also indicates the benefit of using independent encoding streams for different behavioral features. This result may be caused by the higher model complexity in terms of the number of weights for the Independent strategy with a fixed number of hidden units in the GRUs.

4.6.3. Results Using STEMM

We then performed experiments on the ECP-Intention dataset, evaluating our method BA-PTP using ego-motion predictions from STEMM for testing. The pre-trained variant of STEMM has been used for ECP-Intention, as it leads to better performance compared to the end-to-end variant. This means that STEMM directly predicts the speed and yaw rate of the ego-vehicle for every future timestep and replaces the use of precise temporal future odometry information from Section 4.6.1. The better performance of the pre-trained variant of STEMM for ECP-Intention can be explained by our findings from Section 4.6.1. We observed significant performance gains when adding the yaw rate of the ego-vehicle to the inputs, indicating its importance. Pre-training STEMM on the true future odometry seems to satisfy the yaw rate requirement more effectively compared to a latent feature vector in the end-to-end variant without an explicit supervised yaw rate loss. For training STEMM, we set the initial learning rate to 5 × 10 3 and used L 2 regularization of 10 4 . Furthermore, the loss weight λ for the yaw rate was set to 50. The results are shown in Table 4.
When replacing precise ground truth information with ego-motion predictions from STEMM in BA-PTP, we observed a decreasing performance across all metrics. For BA-PTPBB+HB, the prediction error increases by 8% for MSE-0.8 s and for a longer prediction horizon by up to 23% for MSE-1.6 s. However, comparing to our ablation, which does not use any odometry information (BA-PTPBB+HB†), underlines the benefit of integrating STEMM into BA-PTP. The integration of STEMM results in significant improvements, i.e., a lower MSE-1.6 s of 50% and a lower CMSE of 52%.
We then compared our behavior-aware model (BA-PTPBB+HB+BO+HO+P) with the current state-of-the-art for pedestrian trajectory prediction in the image plane: PIEtraj [2] and SGNet [4]. The latter is currently the best-performing model on the PIE dataset. As we already discussed in Section 4.6.1, the baselines achieve bad prediction results on the ECP-Intention dataset as they do not utilize information about the yaw rate. We have surpassed SGNet’s performance by a large margin. In terms of MSE-1.6 s, CMSE, and CFMSE, we outperformed them by 61%, 63% and 65%, respectively.

4.7. Results on the PIE Dataset

In this section, we evaluated our proposed method on the PIE dataset [2] using inferred values for body orientation and pose of pedestrians, as described in Section 4.3. For our experiments on the PIE dataset, we additionally compared our results to PIEfull [2], Action-Aware Enc-Dec Network [9], and Crossmodal Transformer [8]. PIEfull incorporates the estimated pedestrian’s intention to cross as well as predicted speed of the ego-vehicle for the bounding box predictions. As described in Section 2, the latter two methods also use behavioral features of pedestrians and [9] even utilizes ground truth future ego-vehicle odometry. The results for the baselines were taken from the respective publications.
Similar to our ablations in Section 4.6.1, we first used ground truth future odometry, i.e., speed, for these experiments, to decouple the influence of ego-motion prediction accuracy on the prediction performance. As there are no annotations for the head bounding boxes for the PIE dataset, we could not investigate this in our experiments.
Table 5 shows that using inferred behavioral features in addition to bounding boxes improves the prediction of pedestrians’ future trajectories also on the PIE dataset. Compared to only using bounding boxes (BA-PTPBB*), adding body orientation increases the prediction performance by a small margin (1% in MSE-1.5 s), while using pose information improves the performance significantly (11% in MSE-1.5 s). The best average prediction performance is achieved by BA-PTPBB+BO+P*. We observed a few points of improvement across all metrics when comparing to BA-PTPBB+P*. In terms of MSE-1.5 s, CMSE, and CFMSE, we outperformed Action-Aware Enc-Dec Network—which also uses ground truth odometry information—by 13%, 16%, and 15%, respectively.
Since we investigated the influence of including behavioral features for pedestrian trajectory predicted in an idealized setting using ground truth future odometry information, we then evaluated the performance of BA-PTP when it was trained in an end-to-end manner alongside STEMM for the purpose of ego-motion compensation. The end-to-end variant of STEMM has been used for PIE as it leads to better performance compared to the pre-trained variant. We explain this by the fact that the scenes in the PIE dataset mostly cover scenarios where the ego-vehicle drives in a straight line or stands still, in contrast to ECP-Intention. In Section 4.6.3, the pre-trained variant of STEMM performs better on the ECP-Intention dataset due to explicitly predicting the yaw rate of the ego-vehicle. This seems to be less important on the PIE dataset, where turning scenarios are underrepresented. For PIE, the latent features learned by the end-to-end variant of STEMM provide better means for ego-motion compensation to be used by BA-PTP. We changed the initial learning rate to 5 × 10 4 when using STEMM. The dimensionality of the future ego-motion vectors e m t + j is d e m = 64 . The results are shown in the final two rows of Table 5.
When we incorporated STEMM into BA-PTP instead of using precise future odometry information, the prediction performance drops slightly. For BA-PTPBB+BO+P we lose 4% in MSE-1.5 s and CMSE and 12% in CFMSE. However, using STEMM’s output instead of ground truth odometry results in a better performance for shorter prediction horizons, i.e., four points for 0.5 s and six points for 1 s. Moreover, when comparing BA-PTPBB with its counterpart, which neglects any odometry information (BA-PTPBB†), we observed a highly increasing performance across all metrics. Additionally, BA-PTPBB outperforms PIEfull—which also utilizes ego-motion predictions—by a large margin (11% in MSE-1.5 s). This aligns with our hypothesis that the latent features provided by the end-to-end variant of STEMM allow for more effective ego-motion compensations on PIE compared to explicit ego-motion predictions like speed, which is predicted by PIEfull. These findings demonstrate the advantage of integrating STEMM into our pedestrian trajectory prediction method.
Comparing to the previous best-performing approaches for pedestrian trajectory prediction shows that we outperform Crossmodal Transformer—which also uses the pose and body orientation of pedestrians as input—by 7% in MSE-1.5 s. Further, we observe that SGNet, which solely relies on pedestrian bounding boxes as input, is able to better capture the short-term motions of pedestrians (0.5 s and 1 s). However, BA-PTPBB+BO+P achieves state-of-the-art performance in terms of MSE-1.5 s and CMSE, as well as CFMSE, improving over SGNet by 7%, 7% and 9%, respectively. Therefore, we conclude that the importance of ego-motion compensation and the benefit of behavioral features increases with higher prediction horizons. For behavioral features, this has also been shown in Section 4.6.1.

4.8. Qualitative Results

Figure 4 shows the qualitative results of our proposed models on ECP-Intention and on PIE. Due to the difference in the used inputs for our models for ECP-Intention and PIE, we introduced two naming conventions here. First, we refer to our behavior-aware models as BA-PTPbeh, i.e., BA-PTPBB+HB+BO+HO+P for ECP-Intention and BA-PTPBB+BO+P for PIE. Second, we refer to ablations of our model, which solely relies on the past trajectory, as BA-PTPtraj, i.e., BA-PTPBB+HB for ECP-Intention and BA-PTPBB for PIE. We compared our behavior-aware model BA-PTPbeh with three other models:
  • An ablation of our model solely relying on the past trajectory and not using any odometry information, i.e., BA-PTPtraj†;
  • BA-PTPtraj, i.e., our method without using behavioral features;
  • PIEtraj+speed.
Images (a) and (b) depict two example situations in which the ego-vehicle is executing a turning maneuver. In these scenarios, both of our ego-motion-aware models effectively account for the camera’s ego-motion due to the incorporation of our newly proposed STEMM, thereby compensating for its impact, whereas the two other models lag behind. BA-PTPtraj† does not incorporate any odometry information and PIEtraj+speed is missing the ability to compensate for the horizontal movement of the image due to neglecting yaw rate information. These two examples underline the effectiveness of integrating STEMM into a pedestrian trajectory prediction model. However, in image (g) we show a failure case on the PIE dataset, where the ego-vehicle is also making a turn. We still see that the ego-motion-aware models perform better in this scenario and predict the turn of the ego-vehicle, but they cannot completely compensate for the ego-motion of the camera.
Examples (c) and (e) illustrate instances where BA-PTPbeh successfully predicts a change in motion, specifically that the pedestrians depicted in both images will start crossing the street. On the contrary, all other models fail to accurately estimate the correct future behavior in these cases. This observation can be attributed to the pedestrians’ orientations towards the street, highlighting the potential improvement in pedestrian trajectory prediction by leveraging behavioral features such as body and head orientation as well as pose. Moreover, we show a scenario on the ECP-Intention dataset, where a pedestrian suddenly stops crossing the street, in image (d). All models correctly predict that the pedestrian will not cross the street. Additionally, BA-PTPbeh performs best in compensating for the ego-motion of the vehicle surpassing PIEtraj+speed, which uses ground truth future speed information.
In the image (f), we also show a failure case, where a pedestrian decides to cross the street in high traffic. None of the models predict this behavior. BA-PTPbeh even estimates the future position on the sidewalk. This failure might be explained by the missing multimodality in BA-PTP’s output as this is a demonstrative scenario where two possible trajectories are plausible and the model cannot predict both. Image (h) displays a pedestrian walking across a zebra crossing, exhibiting a linear motion pattern. All models accurately predict the future trajectory in this case, as the positional information from past bounding boxes, which is utilized by all models, proves to be sufficient.

5. Discussion

This study demonstrated the benefit of using behavioral features in a pedestrian trajectory prediction method. It also showcased that extending the representation of pedestrian trajectories by head bounding boxes in addition to solely using body bounding boxes improves the prediction performance by learning better encodings of the motion history of pedestrians. Moreover, we have shown the effectiveness of including STEMM in our method to account for the ego-motion of the camera in ego-centric views. BA-PTP outperformed all prior state-of-the-art works on two datasets for pedestrian behavior prediction.
Compared to [11], we extended our preliminary work by including information about the head position of pedestrians as well as adapting the bounding box representation to reduce the box wobbling effects. One major limitation of our previous work was the use of ground truth future odometry information. In this work, we introduced STEMM and integrated it into BA-PTP, dropping the need for highly precise ego-motion information from the future. This makes our extended method usable in a real-world application with unknown future odometry. During the methodical extension of BA-PTP, we also had to revisit the choice of hyperparameters used to train the model (cf. Section 4.3) to cope with the new changes. In order to ensure fair quantitative comparisons, we compared our results using ground truth future odometry (Table 2 and Table 5) with our preliminary results reported in [11]. On both datasets, ECP-Intention and PIE, we observed remarkable improvements in the prediction performance for our extended version of our method. For the model variant of BA-PTP that solely utilizes the past body bounding boxes of pedestrians, we achieved an improvement of 5% in terms of MSE-1.5 s on PIE and 6% in terms of MSE-1.6 s on ECP-Intention. This can be partly attributed to the optimized hyperparameters, but particularly to the change in representing bounding boxes (Section 3), which contributes to reducing the box wobbling effects. When comparing the results of the best performing behavior-aware models, we note even more significant gains in terms of performance. On the ECP-Intention dataset, we outperformed our method from [11] by 19% across all metrics. The even greater improvement in performance indicates the advantage of incorporating the head bounding boxes, which were not utilized in our previous method. This extension results in a more stable encoding of pedestrian trajectories in the image plane. Also, this aligns with our hypothesis from Section 3. Another notable insight is the superior performance of our extended method BA-PTP using the ego-motion predictions from STEMM (cf. BA-PTPBB+BO+P in Table 5) against our results reported in [11] in terms of MSE-0.5 s, MSE-1 s, and MSE-1.5 s on the PIE dataset. Despite the utilization of ground truth future odometry information in [11], we still improved over our preliminary work, highlighting the contribution of STEMM to improving pedestrian trajectory prediction in the image plane.
We compared our proposed method to current state-of-the-art pedestrian trajectory prediction methods in the image plane in Section 4. The enhancement observed in comparison to [8] further demonstrates the capabilities of BA-PTP. Both methods share similar inputs, namely bounding boxes, poses, and the body orientation of pedestrians as well as the ego-motion of the vehicle. However, the crossmodal transformer based encoders in [8] are used to learn how the pedestrian modalities (box and pose) are influenced by the past ego-vehicle odometry information, whereas BA-PTP explicitly uses the ego-motion predictions provided by STEMM. This suggests that the integration of STEMM for ego-motion compensation helps to predict more accurate future bounding boxes. Additionally, only the final observed body orientation of a pedestrian was incorporated in [8] to encode the basic state of the pedestrian, in contrast to our method, which encodes a sequence of observed past body orientations.
We have also demonstrated superior performance compared to the method presented in [9], which also uses the pose of pedestrians in addition to bounding boxes as input. We attribute its lower performance to its encoding strategy, where the input modalities are concatenated first before being processed jointly by LSTM blocks contrary to our independent encoding approach. The benefit of using independent encoding streams for different input modalities has been shown for our method in Section 4.6.2. Moreover, Ref. [9] further utilizes ground truth future odometry information to compensate for the ego-motion of the camera, whereas BA-PTP utilizes ego-motion predictions provided by STEMM.
Our study has several limitations. In order to deploy STEMM in a real-world application, it is necessary to know the intended route of the ego-vehicle, e.g., provided by a navigation system. Nonetheless, in the domain of automated vehicles, knowing the intended route of the ego-vehicle can be considered given. The benefit and importance of accounting for the ego-motion of the camera for pedestrian trajectory prediction in the image plane has been demonstrated in Section 4.6.3 and Section 4.7.
Despite the state-of-the-art performance of BA-PTP, the quantitative as well as qualitative evaluations indicate that there is still room for improvement. Due to the highly variable behavior of pedestrians, we can observe scenarios in which multiple future trajectories of pedestrians are plausible. BA-PTP’s prediction performance is limited in such cases, because it is missing the capability of predicting multimodal trajectories. However, this fundamental extension is beyond the scope of this study and is left for future work. Furthermore, we focused on extracting the behavioral features from camera images and explicitly utilizing them with independent encoding streams in this work. An alternative to this approach would have been to directly use the observed images or image crops as input to a trajectory prediction method in an end-to-end manner. Future research directions should consider investigations regarding the effectiveness of such an approach.

6. Conclusions

In this work, we presented BA-PTP, a novel approach to pedestrian trajectory prediction from the ego-vehicle camera perspective. Behavioral features extracted from visual observations, such as the body and head orientation of pedestrians as well as their pose, were utilized by BA-PTP in addition to positional information from body and head bounding boxes. We added the head bounding boxes to the positional information to provide a more stable encoding of pedestrian trajectories in the image plane. To enhance the learned embeddings of pedestrians’ motion history, we employed independent encoding streams for each input modality and combined the resulting outputs. By adopting this approach, we explicitly leveraged the information provided by pedestrians regarding their intended movement. To overcome the challenge of ego-motion compensation for pedestrian trajectory prediction from the ego-vehicle camera perspective, we integrated STEMM, a novel approach for ego-motion prediction. It uses spatial goal points that are sampled from the intended route of the ego-vehicle and replaces the use of precise temporal future ego-vehicle odometry information.
By evaluating BA-PTP on two datasets for predicting pedestrian behavior, we showed that including behavioral features benefits pedestrian trajectory prediction and demonstrated the strength of incorporating STEMM into our method. Further, we have achieved a state-of-the-art performance on the PIE dataset, outperforming prior work by 7% in MSE-1.5 s and CMSE, as well as by 9% in CFMSE.

Author Contributions

Conceptualization, P.C. and M.B.; methodology, P.C.; software, P.C.; investigation, P.C.; resources, U.K.; data curation, P.C.; writing—original draft preparation, P.C.; writing—review and editing, M.B., U.K. and B.Y.; visualization, P.C.; supervision, M.B. and B.Y.; funding acquisition, U.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work is a result of the research project STADT:up—Solutions and Technologies for Automated Driving in Town: An urban mobility project. The project is supported by the Federal Ministry for Economic Affairs and Climate Action (BMWK), based on a decision taken by the German Bundestag. The authors are solely responsible for the content of this publication.

Data Availability Statement

Two datasets were used in this work. PIE is publicly accessible at https://data.nvision2.eecs.yorku.ca/PIE_dataset/ (accessed on 31 July 2023). The ECP-Intention dataset is not published yet.

Acknowledgments

We thank Sebastian Krebs for his valuable support throughout the research in this work. We thank Arij Bouazizi and Julian Wiederer for proofreading the manuscript and providing constructive feedback.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Zhang, C.; Berger, C. Pedestrian Behavior Prediction Using Deep Learning Methods for Urban Scenarios: A Review. IEEE Trans. Intell. Transp. Syst. 2023. early access. [Google Scholar] [CrossRef]
  2. Rasouli, A.; Kotseruba, I.; Kunic, T.; Tsotsos, J.K. PIE: A Large-Scale Dataset and Models for Pedestrian Intention Estimation and Trajectory Prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, ICCV 2019, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar] [CrossRef]
  3. Bhattacharyya, A.; Fritz, M.; Schiele, B. Long-term on-board prediction of people in traffic scenes under uncertainty. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4194–4202. [Google Scholar]
  4. Wang, C.; Wang, Y.; Xu, M.; Crandall, D.J. Stepwise goal-driven networks for trajectory prediction. IEEE Robot. Autom. Lett. 2022, 7, 2716–2723. [Google Scholar] [CrossRef]
  5. Yao, Y.; Atkins, E.; Johnson-Roberson, M.; Vasudevan, R.; Du, X. Bitrap: Bi-directional pedestrian trajectory prediction with multi-modal goal estimation. IEEE Robot. Autom. Lett. 2021, 6, 1463–1470. [Google Scholar] [CrossRef]
  6. Cao, D.; Fu, Y. Using graph convolutional networks skeleton-based pedestrian intention estimation models for trajectory prediction. J. Phys. Conf. Ser. 2020, 1621, 012047. [Google Scholar] [CrossRef]
  7. Sui, Z.; Zhou, Y.; Zhao, X.; Chen, A.; Ni, Y. Joint Intention and Trajectory Prediction Based on Transformer. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2021, Prague, Czech Republic, 27 September–1 October 2021; pp. 7082–7088. [Google Scholar] [CrossRef]
  8. Su, Z.; Huang, G.; Zhang, S.; Hua, W. Crossmodal transformer based generative framework for pedestrian trajectory prediction. In Proceedings of the International Conference on Robotics and Automation, ICRA 2022, Philadelphia, PA, USA, 23–27 May 2022; pp. 2337–2343. [Google Scholar] [CrossRef]
  9. Fu, J.; Zhao, X. Action-Aware Encoder-Decoder Network for Pedestrian Trajectory Prediction. J. Shanghai Jiaotong Univ. (Sci.) 2023, 28, 20–27. [Google Scholar] [CrossRef]
  10. Malla, S.; Dariush, B.; Choi, C. Titan: Future forecast using action priors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, 13–19 June 2020; pp. 11186–11196. [Google Scholar]
  11. Czech, P.; Braun, M.; Kreßel, U.; Yang, B. On-Board Pedestrian Trajectory Prediction Using Behavioral Features. In Proceedings of the 21st IEEE International Conference on Machine Learning and Applications (ICMLA), Nassau, Bahamas, 12–14 December 2022; pp. 437–443. [Google Scholar] [CrossRef]
  12. Lorenzo, J.; Parra, I.; Wirth, F.; Stiller, C.; Llorca, D.F.; Sotelo, M.A. Rnn-based pedestrian crossing prediction using activity and pose-related features. In Proceedings of the IEEE Intelligent Vehicles Symposium, IV 2020, Las Vegas, NV, USA, 19 October–13 November 2020; pp. 1801–1806. [Google Scholar] [CrossRef]
  13. Yau, T.; Malekmohammadi, S.; Rasouli, A.; Lakner, P.; Rohani, M.; Luo, J. Graph-sim: A graph-based spatiotemporal interaction modelling for pedestrian action prediction. In Proceedings of the IEEE International Conference on Robotics and Automation, ICRA 2021, Xi’an, China, 30 May–5 June 2021; pp. 8580–8586. [Google Scholar] [CrossRef]
  14. Rasouli, A.; Rohani, M.; Luo, J. Bifold and semantic reasoning for pedestrian behavior prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, 10–17 October 2021; pp. 15600–15610. [Google Scholar] [CrossRef]
  15. Kotseruba, I.; Rasouli, A.; Tsotsos, J.K. Benchmark for evaluating pedestrian action prediction. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision, WACV 2021, Waikoloa, HI, USA, 3–8 January 2021; pp. 1258–1268. [Google Scholar] [CrossRef]
  16. Yao, Y.; Atkins, E.; Johnson-Roberson, M.; Vasudevan, R.; Du, X. Coupling Intent and Action for Pedestrian Crossing Behavior Prediction. In Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI 2021, Virtual Event, 19–26 August 2021; pp. 1238–1244. [Google Scholar] [CrossRef]
  17. Rasouli, A.; Kotseruba, I.; Tsotsos, J.K. Are they going to cross? A benchmark dataset and baseline for pedestrian crosswalk behavior. In Proceedings of the IEEE International Conference on Computer Vision Workshops, ICCV Workshops 2017, Venice, Italy, 22–29 October 2017; pp. 206–213. [Google Scholar] [CrossRef]
  18. Fang, Z.; López, A.M. Is the pedestrian going to cross? Answering by 2d pose estimation. In Proceedings of the IEEE Intelligent Vehicles Symposium, IV 2018, Changshu, Suzhou, China, 26–30 June 2018; pp. 1271–1276. [Google Scholar] [CrossRef] [Green Version]
  19. Cadena, P.R.G.; Qian, Y.; Wang, C.; Yang, M. Pedestrian graph+: A fast pedestrian crossing prediction model based on graph convolutional networks. IEEE Trans. Intell. Transp. Syst. 2022, 23, 21050–21061. [Google Scholar] [CrossRef]
  20. Liu, B.; Adeli, E.; Cao, Z.; Lee, K.H.; Shenoi, A.; Gaidon, A.; Niebles, J.C. Spatiotemporal relationship reasoning for pedestrian intent prediction. IEEE Robot. Autom. Lett. 2020, 5, 3485–3492. [Google Scholar] [CrossRef] [Green Version]
  21. Varytimidis, D.; Alonso-Fernandez, F.; Duran, B.; Englund, C. Action and intention recognition of pedestrians in urban traffic. In Proceedings of the 14th International Conference on Signal-Image Technology & Internet-Based Systems, SITIS 2018, Las Palmas de Gran Canaria, Spain, 26–29 November 2018; pp. 676–682. [Google Scholar] [CrossRef] [Green Version]
  22. Yagi, T.; Mangalam, K.; Yonetani, R.; Sato, Y. Future person localization in first-person videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7593–7602. [Google Scholar] [CrossRef] [Green Version]
  23. Yang, D.; Zhang, H.; Yurtsever, E.; Redmill, K.; Ozguner, U. Predicting Pedestrian Crossing Intention with Feature Fusion and Spatio-Temporal Attention. IEEE Trans. Intell. Veh. 2022, 7, 221–230. [Google Scholar] [CrossRef]
  24. Ham, J.S.; Kim, D.H.; Jung, N.; Moon, J. CIPF: Crossing Intention Prediction Network Based on Feature Fusion Modules for Improving Pedestrian Safety. In Proceedings of the Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Vancouver, BC, Canada, 18–22 June 2023; pp. 3665–3674. [Google Scholar]
  25. Zhang, Z.; Tian, R.; Ding, Z. TrEP: Transformer-based Evidential Prediction for Pedestrian Intention with Uncertainty. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37. [Google Scholar]
  26. Herman, M.; Wagner, J.; Prabhakaran, V.; Möser, N.; Ziesche, H.; Ahmed, W.; Bürkle, L.; Kloppenburg, E.; Gläser, C. Pedestrian behavior prediction for automated driving: Requirements, metrics, and relevant features. IEEE Trans. Intell. Transp. Syst. 2021, 23, 14922–14937. [Google Scholar]
  27. Kooij, J.F.; Flohr, F.; Pool, E.A.; Gavrila, D.M. Context-based path prediction for targets with switching dynamics. Int. J. Comput. Vis. 2019, 127, 239–262. [Google Scholar] [CrossRef] [Green Version]
  28. Ridel, D.A.; Deo, N.; Wolf, D.; Trivedi, M. Understanding pedestrian-vehicle interactions with vehicle mounted vision: An LSTM model and empirical analysis. In Proceedings of the IEEE Intelligent Vehicles Symposium, IV 2019, Paris, France, 9–12 June 2019; pp. 913–918. [Google Scholar] [CrossRef] [Green Version]
  29. Alahi, A.; Goel, K.; Ramanathan, V.; Robicquet, A.; Fei-Fei, L.; Savarese, S. Social lstm: Human trajectory prediction in crowded spaces. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 961–971. [Google Scholar] [CrossRef] [Green Version]
  30. Sadeghian, A.; Kosaraju, V.; Sadeghian, A.; Hirose, N.; Rezatofighi, H.; Savarese, S. Sophie: An attentive gan for predicting paths compliant to social and physical constraints. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, 15–20 June 2019; pp. 1349–1358. [Google Scholar] [CrossRef] [Green Version]
  31. Lee, N.; Choi, W.; Vernaza, P.; Choy, C.B.; Torr, P.H.; Chandraker, M. Desire: Distant future prediction in dynamic scenes with interacting agents. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 336–345. [Google Scholar] [CrossRef] [Green Version]
  32. Salzmann, T.; Ivanovic, B.; Chakravarty, P.; Pavone, M. Trajectron++: Dynamically-feasible trajectory forecasting with heterogeneous data. In Proceedings of the 16th European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 683–700. [Google Scholar] [CrossRef]
  33. Zhao, H.; Wildes, R.P. Where are you heading? Dynamic trajectory prediction with expert goal examples. In Proceedings of the IEEE/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, 10–17 October 2021; pp. 7629–7638. [Google Scholar] [CrossRef]
  34. Neumann, L.; Vedaldi, A. Pedestrian and Ego-vehicle Trajectory Prediction from Monocular Camera. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 10204–10212. [Google Scholar] [CrossRef]
  35. Yin, Z.; Liu, R.; Xiong, Z.; Yuan, Z. Multimodal Transformer Networks for Pedestrian Trajectory Prediction. In Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI 2021, Virtual Event, 19–26 August 2021; pp. 1259–1265. [Google Scholar] [CrossRef]
  36. Song, X.; Kang, M.; Zhou, S.; Wang, J.; Mao, Y.; Zheng, N. Pedestrian Intention Prediction Based on Traffic-Aware Scene Graph Model. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2022, Kyoto, Japan, 23–27 October 2022; pp. 9851–9858. [Google Scholar] [CrossRef]
  37. Zhai, X.; Hu, Z.; Yang, D.; Zhou, L.; Liu, J. Social Aware Multi-Modal Pedestrian Crossing Behavior Prediction. In Proceedings of the 16th Asian Conference on Computer Vision, Macao, China, 4–8 December 2022; pp. 4428–4443. [Google Scholar]
  38. Achaji, L.; Moreau, J.; Fouqueray, T.; Aioun, F.; Charpillet, F. Is attention to bounding boxes all you need for pedestrian action prediction? In Proceedings of the IEEE Intelligent Vehicles Symposium, IV 2022, Aachen, Germany, 4–9 June 2022; pp. 895–902. [Google Scholar]
  39. Weng, J.J.; Ahuja, N.; Huang, T.S. Learning recognition and segmentation of 3-D objects from 2-D images. In Proceedings of the Fourth International Conference on Computer Vision, ICCV 1993, Berlin, Germany, 11–14 May 1993; pp. 121–128. [Google Scholar] [CrossRef]
  40. Chitta, K.; Prakash, A.; Geiger, A. Neat: Neural attention fields for end-to-end autonomous driving. In Proceedings of the IEEE/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, 10–17 October 2021; pp. 15793–15803. [Google Scholar] [CrossRef]
  41. Renz, K.; Chitta, K.; Mercea, O.B.; Koepke, A.; Akata, Z.; Geiger, A. PlanT: Explainable Planning Transformers via Object-Level Representations. arXiv 2022, arXiv:2210.14222. [Google Scholar] [CrossRef]
  42. Chitta, K.; Prakash, A.; Jaeger, B.; Yu, Z.; Renz, K.; Geiger, A. Transfuser: Imitation with transformer-based sensor fusion for autonomous driving. IEEE Trans. Pattern Anal. Mach. Intell. 2022. early access. [Google Scholar] [CrossRef]
  43. Braun, M.; Krebs, S.; Flohr, F.; Gavrila, D.M. EuroCity Persons: A Novel Benchmark for Person Detection in Traffic Scenes. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 1844–1861. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar] [CrossRef]
  45. Kumar, C.; Ramesh, J.; Chakraborty, B.; Raman, R.; Weinrich, C.; Mundhada, A.; Jain, A.; Flohr, F.B. Vru pose-ssd: Multiperson pose estimation for automated driving. Proc. AAAI Conf. Artif. Intell. 2021, 35, 15331–15338. [Google Scholar] [CrossRef]
  46. Braun, M.; Rao, Q.; Wang, Y.; Flohr, F. Pose-rcnn: Joint object detection and pose estimation using 3d object proposals. In Proceedings of the 19th IEEE International Conference on Intelligent Transportation Systems, ITSC 2016, Rio de Janeiro, Brazil, 1–4 November 2016; pp. 1546–1551. [Google Scholar] [CrossRef]
  47. Braun, M.; Flohr, F.B.; Krebs, S.; Kreßel, U.; Gavrila, D.M. Simple Pair Pose-Pairwise Human Pose Estimation in Dense Urban Traffic Scenes. In Proceedings of the IEEE Intelligent Vehicles Symposium, IV 2021, Nagoya, Japan, 11–17 July 2021; pp. 1545–1552. [Google Scholar] [CrossRef]
  48. Wang, S.; Yang, D.; Wang, B.; Guo, Z.; Verma, R.; Ramesh, J.; Weinrich, C.; Kreßel, U.; Flohr, F.B. UrbanPose: A new benchmark for VRU pose estimation in urban traffic scenes. In Proceedings of the IEEE Intelligent Vehicles Symposium, IV 2021, Nagoya, Japan, 11–17 July 2021; pp. 1537–1544. [Google Scholar] [CrossRef]
Figure 1. Our proposed pedestrian trajectory prediction method (BA-PTP) utilizes behavioral features such as body and head orientation as well as pose in addition to body and head bounding boxes. These features are combined with ego-vehicle information to predict the future bounding boxes of a pedestrian in the image.
Figure 1. Our proposed pedestrian trajectory prediction method (BA-PTP) utilizes behavioral features such as body and head orientation as well as pose in addition to body and head bounding boxes. These features are combined with ego-vehicle information to predict the future bounding boxes of a pedestrian in the image.
Make 05 00050 g001
Figure 2. Visualization of box wobbling. We display cropped images from consecutive timesteps with body (green) and head (blue) bounding boxes of crossing pedestrians at a frame rate of 5 fps. Moreover, we plotted the respective box width over time. This figure demonstrates the fluctuations of body bounding box widths due to the natural gait cycle of humans. In contrast to that, the head bounding boxes are not prone to changes in their shapes.
Figure 2. Visualization of box wobbling. We display cropped images from consecutive timesteps with body (green) and head (blue) bounding boxes of crossing pedestrians at a frame rate of 5 fps. Moreover, we plotted the respective box width over time. This figure demonstrates the fluctuations of body bounding box widths due to the natural gait cycle of humans. In contrast to that, the head bounding boxes are not prone to changes in their shapes.
Make 05 00050 g002
Figure 3. Diagram of our proposed method BA-PTP. The inputs to the method are body and head bounding boxes and behavioral features, such as pedestrian pose and head and body orientation for a defined observation horizon. Each input modality is processed by an independent encoding stream and the outputs of each stream are fused using a modality attention module, resulting in a final embedding vector. For each prediction timestep, we concatenated the final embedding vector with future ego-motion information and passed this through a self-attention unit. This is then used as the decoder inputs to predict future bounding boxes in the image. For predicting the future ego-motion of the ego-vehicle, we integrated STEMM into BA-PTP. STEMM follows a similar scheme as the bounding box prediction, but the inputs are speed, yaw rate, and ego-vehicle positions from the observation horizon. Moreover, STEMM utilizes spatial goal points sampled from the intended route to predict the future ego-motion of the ego-vehicle.
Figure 3. Diagram of our proposed method BA-PTP. The inputs to the method are body and head bounding boxes and behavioral features, such as pedestrian pose and head and body orientation for a defined observation horizon. Each input modality is processed by an independent encoding stream and the outputs of each stream are fused using a modality attention module, resulting in a final embedding vector. For each prediction timestep, we concatenated the final embedding vector with future ego-motion information and passed this through a self-attention unit. This is then used as the decoder inputs to predict future bounding boxes in the image. For predicting the future ego-motion of the ego-vehicle, we integrated STEMM into BA-PTP. STEMM follows a similar scheme as the bounding box prediction, but the inputs are speed, yaw rate, and ego-vehicle positions from the observation horizon. Moreover, STEMM utilizes spatial goal points sampled from the intended route to predict the future ego-motion of the ego-vehicle.
Make 05 00050 g003
Figure 4. Qualitative results on the ECP-Intention dataset (images ad) and on the PIE dataset (images eh). The images show the last observed timestep with the last observed bounding box (white box) as well as predictions into the future (1.6 s for ECP-Intention and 1.5 s for PIE). Visualized are the final predicted bounding box and a line connecting the centers of the predicted bounding boxes for the intermediate timesteps. Ground truth is displayed in green. Predictions of PIEtraj+speed in yellow, our model solely relying on the past trajectory and not using any odometry information BA-PTPtraj† in purple, BA-PTPtraj in red and our behavior-aware model BA-PTPbeh in blue.
Figure 4. Qualitative results on the ECP-Intention dataset (images ad) and on the PIE dataset (images eh). The images show the last observed timestep with the last observed bounding box (white box) as well as predictions into the future (1.6 s for ECP-Intention and 1.5 s for PIE). Visualized are the final predicted bounding box and a line connecting the centers of the predicted bounding boxes for the intermediate timesteps. Ground truth is displayed in green. Predictions of PIEtraj+speed in yellow, our model solely relying on the past trajectory and not using any odometry information BA-PTPtraj† in purple, BA-PTPtraj in red and our behavior-aware model BA-PTPbeh in blue.
Make 05 00050 g004
Table 1. Characteristic comparison of our approach against prior works for pedestrian trajectory prediction in ego-centric camera views. We compared in terms of which pedestrian features are used (BB = Body Bounding Box, HB = Head Bounding Box, P = Pose, BO = Body Orientation, HO = Head Orientation), whether predicted ego-motion is incorporated for bounding box prediction, and if the predictions are conditioned on estimated goals.
Table 1. Characteristic comparison of our approach against prior works for pedestrian trajectory prediction in ego-centric camera views. We compared in terms of which pedestrian features are used (BB = Body Bounding Box, HB = Head Bounding Box, P = Pose, BO = Body Orientation, HO = Head Orientation), whether predicted ego-motion is incorporated for bounding box prediction, and if the predictions are conditioned on estimated goals.
MethodPedestrian FeaturesEgo-MotionGoal-
BBHBPBOHOPredictionDriven
Bhattacharyya et al. [3]
Rasouli et al. [2]
Yao et al. [5]
Wang et al. [4]
Fu et al. [9]
Su et al. [8]
Czech et al. [11]
Ours
Table 2. Ablation study on the ECP-Intention dataset (BB = Body Bounding Box, HB = Head Bounding Box, BO = Body Orientation, HO = Head Orientation, P = Pose). * denotes no use of yaw rate.
Table 2. Ablation study on the ECP-Intention dataset (BB = Body Bounding Box, HB = Head Bounding Box, BO = Body Orientation, HO = Head Orientation, P = Pose). * denotes no use of yaw rate.
ModelMSE
Avg (Std)
CMSE
Avg (Std)
CFMSE
Avg (Std)
0.8 s1.6 s1.6 s1.6 s
1PIEtraj+speed [2] *417 ± 102120 ± 631997 ± 676516 ± 235
2BA-PTPBB-Y *392 ± 122078 ± 261982 ± 826510 ± 110
3BA-PTPBB242 ± 1917 ± 10830 ± 102510 ± 38
4BA-PTPBB+HB211 ± 4801 ± 10716 ± 112204 ± 24
5BA-PTPBB+HB+HO205 ± 2731 ± 9652 ± 91953 ± 24
6BA-PTPBB+HB+BO206 ± 2734 ± 16656 ± 131957 ± 74
7BA-PTPBB+HB+P197 ± 1679 ± 5597 ± 51788 ± 17
8BA-PTPBB+HB+BO+HO201 ± 3716 ± 14637 ± 121920 ± 48
9BA-PTPBB+HB+HO+P193 ± 5641 ± 13568 ± 111658 ± 30
10BA-PTPBB+HB+BO+P193 ± 5635 ± 14565 ± 131631 ± 37
11BA-PTPBB+HB+BO+HO+P187 ± 2615 ± 17544 ± 151579 ± 65
12BA-PTPBB+HB+infer. BO205 ± 4741 ± 18659 ± 161986 ± 63
13BA-PTPBB+HB+infer. BO+P194 ± 5656 ± 25579 ± 221703 ± 86
Table 3. Comparison of different encoding strategies on the ECP-Intention dataset (C = Concat, I = Independent).
Table 3. Comparison of different encoding strategies on the ECP-Intention dataset (C = Concat, I = Independent).
ModelMSE
Avg (Std)
CMSE
Avg (Std)
CFMSE
Avg (Std)
0.8 s1.6 s1.6 s1.6 s
BA-PTPBB+HB+BO+P-C367 ± 7989 ± 23895 ± 192424 ± 84
BA-PTPBB+HB+BO+P-I193 ± 5635 ± 14565 ± 131631 ± 37
BA-PTPBB+HB+BO+HO+P-C371 ± 9996 ± 20907 ± 162406 ± 65
BA-PTPBB+HB+BO+HO+P-I187 ± 2615 ± 17544 ± 151579 ± 65
Table 4. Comparison with SOTA on the ECP-Intention dataset (BB = Body Bounding Box, HB = Head Bounding Box, BO = Body Orientation, HO = Head Orientation, P = Pose). * denotes using ground truth future odometry. † denotes using no odometry information.
Table 4. Comparison with SOTA on the ECP-Intention dataset (BB = Body Bounding Box, HB = Head Bounding Box, BO = Body Orientation, HO = Head Orientation, P = Pose). * denotes using ground truth future odometry. † denotes using no odometry information.
ModelMSE
Avg (Std)
CMSE
Avg (Std)
CFMSE
Avg (Std)
0.8 s1.6 s1.6 s1.6 s
PIEtraj [2] †434 ± 242216 ± 712088 ± 706785 ± 136
PIEtraj+speed [2] *417 ± 102120 ± 631997 ± 676516 ± 235
SGNet [4] †390 ± 82137 ± 942033 ± 926708 ± 330
BA-PTPBB+HB369 ± 142018 ± 241912 ± 236382 ± 163
BA-PTPBB+HB229 ± 4991 ± 11903 ± 112882 ± 27
BA-PTPBB+HB+BO+HO+P207 ± 2824 ± 16751 ± 142339 ± 53
BA-PTPBB+HB *211 ± 4801 ± 10716 ± 112204 ± 24
BA-PTPBB+HB+BO+HO+P *187 ± 2615 ± 17544 ± 151579 ± 65
Table 5. Trajectory prediction results on the PIE dataset (BB = Body Bounding Box, BO = Body Orientation, P = Pose). * denotes using ground truth future odometry. † denotes using no odometry information.
Table 5. Trajectory prediction results on the PIE dataset (BB = Body Bounding Box, BO = Body Orientation, P = Pose). * denotes using ground truth future odometry. † denotes using no odometry information.
ModelMSE
Avg (Std)
CMSE
Avg (Std)
CFMSE
Avg (Std)
0.5 s1 s1.5 s1.5 s1.5 s
PIEtraj [2] †582006365962477
PIEfull [2] - - 559 520 2162
PIEtraj+speed [2] *60 ± 2173 ± 4498 ± 9450 ± 81782 ± 40
Action-Aware Enc-Dec Network [9] *431604574361683
Crossmodal Transformer [8]431494434131670
SGNet [4] †341334424131761
BA-PTPBB53 ± 1188 ± 3615 ± 11580 ± 112469 ± 44
BA-PTPBB *48 ± 1154 ± 2459 ± 4429 ± 41746 ± 24
BA-PTPBB+BO *50 ± 1155 ± 1452 ± 5421 ± 51692 ± 34
BA-PTPBB+P *51 ± 1146 ± 2405 ± 6376 ± 51473 ± 31
BA-PTPBB+BO+P *50 ± 1143 ± 3395 ± 4366 ± 41421 ± 9
BA-PTPBB47 ± 0158 ± 4495 ± 22463 ± 211940 ± 105
BA-PTPBB+BO+P46 ± 0137 ± 1411 ± 4381 ± 31593 ± 22
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Czech, P.; Braun, M.; Kreßel, U.; Yang, B. Behavior-Aware Pedestrian Trajectory Prediction in Ego-Centric Camera Views with Spatio-Temporal Ego-Motion Estimation. Mach. Learn. Knowl. Extr. 2023, 5, 957-978. https://doi.org/10.3390/make5030050

AMA Style

Czech P, Braun M, Kreßel U, Yang B. Behavior-Aware Pedestrian Trajectory Prediction in Ego-Centric Camera Views with Spatio-Temporal Ego-Motion Estimation. Machine Learning and Knowledge Extraction. 2023; 5(3):957-978. https://doi.org/10.3390/make5030050

Chicago/Turabian Style

Czech, Phillip, Markus Braun, Ulrich Kreßel, and Bin Yang. 2023. "Behavior-Aware Pedestrian Trajectory Prediction in Ego-Centric Camera Views with Spatio-Temporal Ego-Motion Estimation" Machine Learning and Knowledge Extraction 5, no. 3: 957-978. https://doi.org/10.3390/make5030050

APA Style

Czech, P., Braun, M., Kreßel, U., & Yang, B. (2023). Behavior-Aware Pedestrian Trajectory Prediction in Ego-Centric Camera Views with Spatio-Temporal Ego-Motion Estimation. Machine Learning and Knowledge Extraction, 5(3), 957-978. https://doi.org/10.3390/make5030050

Article Metrics

Back to TopTop