2020
DOI: 10.1109/lra.2020.2976305
|View full text |Cite
|
Sign up to set email alerts
|

Spatiotemporal Relationship Reasoning for Pedestrian Intent Prediction

Abstract: Reasoning over visual data is a desirable capability for robotics and vision-based applications. Such reasoning enables forecasting the next events or actions in videos. In recent years, various models have been developed based on convolution operations for prediction or forecasting, but they lack the ability to reason over spatiotemporal data and infer the relationships of different objects in the scene. In this paper, we present a framework based on graph convolution to uncover the spatiotemporal relationshi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
49
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 124 publications
(68 citation statements)
references
References 73 publications
0
49
0
Order By: Relevance
“…In the future, some possible lines of work are detailed below: Evaluate our model in additional datasets, such as STIP [ 25 ] or TITAN [ 24 ]; Train and test models in a combination of different available behavior datasets and analyze it in an unrelated scenario to see its generalization capabilities; Research deeper in the usage of data augmentation techniques in the training of multi-branch models; Consider the development of virtual scenarios to include more crossing cases and fight data imbalance; Explore new features from datasets, such as labeled information from the environment (e.g., the relative relationship between vehicles, crosswalk position, traffic signals). These new features will be 2D or 3D, depending on the availability of datasets in the literature; Experimentation with different data cleaning strategies in training time, applying a maximum occlusion level, pedestrian minimum size, etc.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…In the future, some possible lines of work are detailed below: Evaluate our model in additional datasets, such as STIP [ 25 ] or TITAN [ 24 ]; Train and test models in a combination of different available behavior datasets and analyze it in an unrelated scenario to see its generalization capabilities; Research deeper in the usage of data augmentation techniques in the training of multi-branch models; Consider the development of virtual scenarios to include more crossing cases and fight data imbalance; Explore new features from datasets, such as labeled information from the environment (e.g., the relative relationship between vehicles, crosswalk position, traffic signals). These new features will be 2D or 3D, depending on the availability of datasets in the literature; Experimentation with different data cleaning strategies in training time, applying a maximum occlusion level, pedestrian minimum size, etc.…”
Section: Discussionmentioning
confidence: 99%
“…Another dataset with crossing behavior, binary annotated in this case, is STIP dataset [ 25 ]. This dataset is recorded with a multi-camera setup and includes hand-labeled bounding boxes at a low annotation rate.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…However, the generated context features are not always applicable to current data sets and require different kinds of modalities or hardware constraints [60,61]. In that regard, Liu et al [62] propose a new data set for pedestrian intention prediction tailored to intent prediction in dense driving scenes. It defines a model based on graph convolutions to represent the spatiotemporal context of the scene where each identified object is presented as a node of a spatiotemporal graph for two different perspectives: pedestrian-centric and location-centric settings graphs.…”
Section: Pedestrian Intention Predictionmentioning
confidence: 99%
“…Haddad et al [5] used spatiotemporal graphs to capture both the temporal and spatial correlations of pedestrian predictions and considered physical cues in a scene and the interactions between pedestrians, thereby improving the performance of trajectory prediction. In addition, Liang et al [6] and Liu et al [7] considered pedestrian-scene and pedestrian-object relationships simultaneously and incorporated pedestrian intentions to model future paths and predict human activities and locations. However, their work ignored the multimodal nature of the prediction of future pedestrian trajectories.…”
Section: Introductionmentioning
confidence: 99%