2020 25th International Conference on Pattern Recognition (ICPR) 2021
DOI: 10.1109/icpr48806.2021.9412190
|View full text |Cite
|
Sign up to set email alerts
|

Transformer Networks for Trajectory Forecasting

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
153
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 231 publications
(162 citation statements)
references
References 28 publications
2
153
0
Order By: Relevance
“…Additionally, a few first transformer models have entered the field of scene prediction, currently with the focus more on human traffic prediction [120,121] or long-term traffic flow (time interval on several minutes) [122]. As in Natural Language Processing (NLP) transformers partially begin to challenge RNN-based sequence-to-sequence models in (human) trajectory too, demonstrating the importance of the attention module in those tasks.…”
Section: Gnns Attention and New Use Casesmentioning
confidence: 99%
“…Additionally, a few first transformer models have entered the field of scene prediction, currently with the focus more on human traffic prediction [120,121] or long-term traffic flow (time interval on several minutes) [122]. As in Natural Language Processing (NLP) transformers partially begin to challenge RNN-based sequence-to-sequence models in (human) trajectory too, demonstrating the importance of the attention module in those tasks.…”
Section: Gnns Attention and New Use Casesmentioning
confidence: 99%
“…The Trajectron [21] combines elements from CVAE, LSTM, and dynamic Spatio-temporal graphical structures to produce multimodal trajectories. Recently, a Transformer model was proposed by [15] to predict the future trajectories of the pedestrians conditioning on the previous displacement of each pedestrian in the scene. The transformer used has the same architecture as the Vanilla Transformer proposed in [11].…”
Section: Related Workmentioning
confidence: 99%
“…Transformers have first revolutionized the natural language processing problems by outperforming all the previously proposed solutions [12], [13]. However, it was only until recently that were proved equally efficient for non-NLP problems [14], [15], [16].…”
Section: Introductionmentioning
confidence: 99%
“…However, taking into account interactions between pedestrians requires to predict the coordinates one step at a time, so RNNs are generally preferred. Most of past years research focused on improving the interaction module, with only limited new methods since Social-LSTM [1], or in developing approaches that take inspiration in popular frameworks such as Transformers [14] or contrastive learning [15] in order to deter the model from predicting colliding or too uncomfortable trajectories. However, little work has been published on the influence of the encoder and thus on the importance of past coordinates, even if it would be easily applicable on all models that use this pipeline.…”
Section: A Encoder-interaction-decoder Pipelinementioning
confidence: 99%
“…Convolutional Neural Networks are faster than RNN-based methods due to parallelization, but the performances are significantly lower [18]. Some authors have explored the popular Transformers architecture, but the results are inferior to those of RNNs with state-of-the-art social interaction modules [14]. Research has also been conducted on applying Inverse Reinforcement Learning (IRL) to the pedestrian trajectory prediction problem [19], even though retrieving the pedestrian cost function requires much more computation than learning a predictor.…”
Section: A Encoder-interaction-decoder Pipelinementioning
confidence: 99%