2022
DOI: 10.1080/15472450.2022.2109416
|View full text |Cite
|
Sign up to set email alerts
|

Multiagent reinforcement learning for autonomous driving in traffic zones with unsignalized intersections

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(3 citation statements)
references
References 23 publications
0
3
0
Order By: Relevance
“…TORCS and SUMO are demonstrated to be prominent tools for developing and testing short-term trajectory planning approaches [50], solutions that handle complex state and action spaces in a continuous domain [51], multi-agent deep reinforcement learning methods for self-driving vehicles capable of navigating through traffic networks with uncontrolled intersections [52], and many other advances or methodologies.…”
Section: State Of the Art Reviewmentioning
confidence: 99%
“…TORCS and SUMO are demonstrated to be prominent tools for developing and testing short-term trajectory planning approaches [50], solutions that handle complex state and action spaces in a continuous domain [51], multi-agent deep reinforcement learning methods for self-driving vehicles capable of navigating through traffic networks with uncontrolled intersections [52], and many other advances or methodologies.…”
Section: State Of the Art Reviewmentioning
confidence: 99%
“…Several research have introduced optimization strategies to further improve intersection performance, including bilevel programming [5,6], tree search [7], and machine learning [8][9][10][11], reservation optimization [12], other strategies [13,14], yielding satisfactory schedule results. However, Lioris et al [15] found that increasing intersection capacity is feasible when vehicles traverse intersections as platoons rather than individually.…”
Section: Introductionmentioning
confidence: 99%
“…In MARL, the interaction is usually modeled as a Stochastic Game(Markov Game) [100,101] instead of MDP, so collective policy can be learnt for all the agents. For example in [102][103][104] MARL is used for multiple AVs planning and decision-making. In [105,106] MARL is utilised to train altruistic autonomous agents to drive safely but also master coordination among agents.…”
Section: Game Theoretic -Game Vs Mdpmentioning
confidence: 99%