Pedestrian Behavior Prediction Models

Pedestrian Behavior Prediction Models

Last Updated: 01/23/2025 | All information is accurate and up-to-date

Two-Tower Ego-Centric Pedestrian Trajectory Prediction

Key Features

  1. Multi-modal inputs;
  2. Two-tower model to decompose egocentric pedestrian trajectories based on ego-vehicle and pedestrian movements;
  3. Inferences of pedestrian future moving directions.
A diagram of the Two-tower Ego-centric Pedestrian Trajectory prediction model structure.
Model Structure
Multiple images of pedestrians crossing roads.
Prediction Results at 1.5 Seconds

Algorithm Results on JAAD Benchmark Dataset

MethodAverage Displacement ErrorFinal Displacement Error
PIE22.8349.44
BiPed21.1348.88
Two-Tower Model17.9241.33

PIE: Rasouli, A., Kotseruba, I., Kunic, T. and Tsotsos, J.K., 2019. Pie: A large-scale dataset and models for pedestrian intention estimation and trajectory prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision (pp. 6262-6271).

BiPed: Rasouli, A.; Rohani, M.; and Luo, J. 2021. Bifold and Semantic Reasoning for Pedestrian Behavior Prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 15600–15610.

Algorithm Results on PSI Benchmark Dataset

MethodAverage Displacement ErrorFinal Displacement Error
PIE35.3961.50
Two-Tower Model22.3446.63

Pedestrian Intention Prediction Models

Key Features of VR-GCN

  1. Scene graph with 32 objects;
  2. CNN + GCN + LSTM as the main structure;
  3. Pose information incorporated;
Graph-Convolutional-Network-based Pedestrian Intent Prediction (VR-GCN)

Key Features of TrEP

  1. Transformer-based feature extraction and encoding;
  2. Evidential learning for robust performance and calibrated uncertainty.
Transformer-based Evidential Learning for Intent Prediction (TrEP)

Algorithm Results on nuScenesBenchmark Dataset

MethodAccuracyBalanced AccuracyF1
VR-GCN0.740.610.64
TrEP0.850.770.90

Pedestrian Intention and Trajectory Prediction