2021
DOI: 10.48550/arxiv.2104.00563
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Latent Variable Sequential Set Transformers For Joint Multi-Agent Motion Prediction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 0 publications
0
5
0
Order By: Relevance
“…As Transformers (Vaswani et al 2017) have gained popularity, an increasing number of studies (Liu et al 2021;Ngiam et al 2021;Jia et al 2023) have utilized the attention mechanism to encode scene context. Encouraged by the successful application of DETR (Carion et al 2020), many Transformerbased models (Girgis et al 2021;Varadarajan et al 2022;Nayakanti et al 2023) have adopted learnable queries in decoder to generate multiple potential future trajectories. In our study, we utilize the architecture presented in MTR (Shi et al 2022a), which is an advanced transformer framework incorporating a local attention based encoder and a decoder with intention queries.…”
Section: Architectures For Motion Predictionmentioning
confidence: 99%
See 2 more Smart Citations
“…As Transformers (Vaswani et al 2017) have gained popularity, an increasing number of studies (Liu et al 2021;Ngiam et al 2021;Jia et al 2023) have utilized the attention mechanism to encode scene context. Encouraged by the successful application of DETR (Carion et al 2020), many Transformerbased models (Girgis et al 2021;Varadarajan et al 2022;Nayakanti et al 2023) have adopted learnable queries in decoder to generate multiple potential future trajectories. In our study, we utilize the architecture presented in MTR (Shi et al 2022a), which is an advanced transformer framework incorporating a local attention based encoder and a decoder with intention queries.…”
Section: Architectures For Motion Predictionmentioning
confidence: 99%
“…In prediction-based matching methods (Ngiam et al 2021;Varadarajan et al 2022;Nayakanti et al 2023), the positive mixture component is chosen by directly comparing predicted trajectories to the ground truth. Some models (Tang and Salakhutdinov 2019;Girgis et al 2021) using the loss based on EM algorithm can also be viewed as prediction-based matching when its KL term converges. Due to the challenge of selecting representative future trajectories, these methods have opted to use well-designed aggregation techniques (Varadarajan et al 2022;Nayakanti et al 2023), or to directly utilize an end-to-end version (Ngiam et al 2021;Girgis et al 2021).…”
Section: Modeling For Multimodal Future Motionmentioning
confidence: 99%
See 1 more Smart Citation
“…Hence, the multimodal trajectory prediction for agent i can be regarded as estimating a mixture distribution Note that we primarily focus on marginal motion prediction in this paper, but our approach can be smoothly extended to joint motion prediction tasks by involving scene-level loss functions [8,33]. We leave it as an important future work.…”
Section: A Problem Formulationmentioning
confidence: 99%
“…Mercat et al [20], by introducing self-attention mechanisms considering interactions between vehicles, successfully achieved trajectory prediction for multiple vehicle agents. Roger et al [21] proposed the AutoBots model, which, through the use of social multi-head self-attention (MHSA) modules, efficiently performs single-pass forward inference for the entire future scene, demonstrating high performance in handling complex traffic scenarios with multi-agent interactions. The adoption of deep learning models like Transformers and MHSA modules has significantly advanced multi-agent trajectory prediction in complex traffic scenarios.…”
Section: Introductionmentioning
confidence: 99%