2020
DOI: 10.48550/arxiv.2006.06412
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Modeling Human Driving Behavior through Generative Adversarial Imitation Learning

Abstract: Imitation learning is an approach for generating intelligent behavior when the cost function is unknown or difficult to specify. Building upon work in inverse reinforcement learning (IRL), Generative Adversarial Imitation Learning (GAIL) aims to provide effective imitation even for problems with large or continuous state and action spaces. Driver modeling is one example of a problem where the state and action spaces are continuous. Human driving behavior is characterized by non-linearity and stochasticity, and… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
12
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3
3

Relationship

1
5

Authors

Journals

citations
Cited by 8 publications
(12 citation statements)
references
References 31 publications
(43 reference statements)
0
12
0
Order By: Relevance
“…GAIL was expanded to perform multi-agent imitation learning in simulated environments (Song et al, 2018). GAIL has recently been tested with mixed levels of success on realworld highway driving scenarios (Bhattacharyya, Phillips, et al, 2018;Bhattacharyya, Wulfe, et al, 2020). GAIL has not yet been implemented in more complex driving environments (e.g.…”
Section: Introductionmentioning
confidence: 99%
“…GAIL was expanded to perform multi-agent imitation learning in simulated environments (Song et al, 2018). GAIL has recently been tested with mixed levels of success on realworld highway driving scenarios (Bhattacharyya, Phillips, et al, 2018;Bhattacharyya, Wulfe, et al, 2020). GAIL has not yet been implemented in more complex driving environments (e.g.…”
Section: Introductionmentioning
confidence: 99%
“…This inaccuracy may not affect significantly for the traffic flow analysis, but will cause significant evaluation biasedness on AV performance. To incorporate the external noise term (t), the most commonly used one is the Gaussian noise (Laval et al, 2014;He et al, 2015;Treiber and Kesting, 2017;Kuefler et al, 2017;Bhattacharyya et al, 2020). For example, Laval et al (2014) added an external Gaussian noise to Newell's car-following model (Newell, 2002) and showed that the stochastic car-following model can produce traffic oscillations that accord well with the observations.…”
Section: Introductionmentioning
confidence: 99%
“…For instance, as humans, we know that vehicles should not collide with each other. However, as found by Bhattacharyya et al [10], it is challenging for completely data-driven techniques to learn such rules. Second, because the datadriven models are not interpretable, it is difficult to verify and validate them, making them less attractive for safetycritical applications such as autonomous driving.…”
Section: Introductionmentioning
confidence: 99%
“…Behavioral cloning [21] is one way to learn such demonstrations from data. However, such supervised learning techniques have proven to be less successful in applications such as modeling multiagent traffic due to compounding errors and not taking into account multi-agent interactions [10]. Techniques such as inverse reinforcement learning [22], [23] and inverse reward design [24] attempt to directly model the underlying reward function of humans.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation