2020 IEEE International Conference on Robotics and Automation (ICRA) 2020
DOI: 10.1109/icra40945.2020.9197340
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Head Attention for Multi-Modal Joint Vehicle Motion Forecasting

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
111
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 141 publications
(122 citation statements)
references
References 15 publications
1
111
0
Order By: Relevance
“…trajectories. Mercat et al [10] use a long short-term memory (LSTM) [18] to predict multiple future trajectories per traffic actor, jointly for all traffic actors in a scene, and scoring the trajectories using self-attention layers. The approach in [19] also utilizes LSTMs and attention for jointly predicting trajectories, but introduces latent variables for generating multiple trajectories per traffic actor.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…trajectories. Mercat et al [10] use a long short-term memory (LSTM) [18] to predict multiple future trajectories per traffic actor, jointly for all traffic actors in a scene, and scoring the trajectories using self-attention layers. The approach in [19] also utilizes LSTMs and attention for jointly predicting trajectories, but introduces latent variables for generating multiple trajectories per traffic actor.…”
Section: Related Workmentioning
confidence: 99%
“…Recently, advances in neural networks [8] as well as the availability of motion forecasting datasets, e.g. [9], gave rise to approaches based on neural networks [10], [11], [12], [13].…”
Section: Introductionmentioning
confidence: 99%
“…Social Interaction: Similar to previous work [17], we model the interaction between agents with multi-headed attention [22]. Each agent is represented with an embedding vector that encodes its observed trajectory and is used to create a key K, query Q, and value vector V for each of the H attention heads.…”
Section: A Network Structurementioning
confidence: 99%
“…A different line of work has used recurrent neural networks (RNN) to build a representation for the trajectories [17] [13]. On top of the recurrent architecture, contextual information is added to model the interaction between agents and to extract the road structure.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation