Proceedings of the 10th ACM/IEEE International Conference on Cyber-Physical Systems 2019
DOI: 10.1145/3302509.3313784
|View full text |Cite
|
Sign up to set email alerts
|

Simulation to scaled city

Abstract: Using deep reinforcement learning, we successfully train a set of two autonomous vehicles to lead a fleet of vehicles onto a roundabout and then transfer this policy from simulation to a scaled city without fine-tuning. We use Flow, a library for deep reinforcement learning in microsimulators, to train two policies, (1) a policy with noise injected into the state and action space and (2) a policy without any injected noise. In simulation, the autonomous vehicles learn an emergent metering behavior for both pol… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 42 publications
(13 citation statements)
references
References 19 publications
0
8
0
Order By: Relevance
“…We seek to derive the optimal control strategy g * ∈ G of the system represented in (17) which minimizes the following expected cost:…”
Section: Illustrative Examplementioning
confidence: 99%
See 3 more Smart Citations
“…We seek to derive the optimal control strategy g * ∈ G of the system represented in (17) which minimizes the following expected cost:…”
Section: Illustrative Examplementioning
confidence: 99%
“…If the dynamics of the actual system given in (17) were known, then we could use (24) to derive the optimal control strategy g * ∈ G. Since the primitive random variables are Gaussian with zero mean, variance 1, and covariance 0.5, we can use the linear least-squares estimator to compute the unique optimal solution of (24), which is…”
Section: A Optimal Control Strategymentioning
confidence: 99%
See 2 more Smart Citations
“…While testing single-agent AVs is challenging, the problem is even more complex in the presence of other agents [6]. AV testing in a multi-agent setting is a complex problem since it is hard to find which states of interactions will most likely lead them to fail.…”
Section: Introductionmentioning
confidence: 99%