2018
DOI: 10.1016/j.trc.2018.10.024
|View full text |Cite
|
Sign up to set email alerts
|

Human-like autonomous car-following model with deep reinforcement learning

Abstract: This study proposes a framework for human-like autonomous car-following planning based on deep reinforcement learning (deep RL). Historical driving data are fed into a simulation environment where an RL agent learns from trial and error interactions based on a reward function that signals how much the agent deviates from the empirical data. Through these interactions, an optimal policy, or car-following model that maps in a human-like way from speed, relative speed between a lead and following vehicle, and int… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
177
1
1

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 362 publications
(180 citation statements)
references
References 32 publications
1
177
1
1
Order By: Relevance
“…In the above formula, U a (s) is the potential energy value of the gravitational field at the point s, U r (s) is the potential energy value of the repulsive field at the point s, and U(s) is the potential energy of the point s. U a (s) and U r (s) are obtained by Formulas (10) and (11), respectively.…”
Section: Improved Model Of Autonomous Path Planningmentioning
confidence: 99%
“…In the above formula, U a (s) is the potential energy value of the gravitational field at the point s, U r (s) is the potential energy value of the repulsive field at the point s, and U(s) is the potential energy of the point s. U a (s) and U r (s) are obtained by Formulas (10) and (11), respectively.…”
Section: Improved Model Of Autonomous Path Planningmentioning
confidence: 99%
“…Building the POMDP model is not an easy task because it is necessary to define not only the states, actions, and transitions but also all the transition probabilities and rewards. Some of those parameters can be learned from experimental data using some reinforcement learning or deep learning techniques [24,25]. A question remaining is to evaluate the performance of systems of this kind against other solutions, such as simpler finite state machines, where it is easier to model the traffic rules and change the model as the traffic rules are updated.…”
Section: Related Workmentioning
confidence: 99%
“…Wang et al [2] developed a car-following model that can better describe freeway driving behaviors based on an adaptive neuro-fuzzy inference system and wavelet analysis for denoising. Zhu et al [14] proposed a framework for a human-like autonomous car-following model based on the deep deterministic policy gradient algorithm of RL, which showed good capability of generalization under various driving situations. It can be driver-adapted by successive learning and is outperforming most of existing data-driven models.…”
Section: Data-driven Cf Modelsmentioning
confidence: 99%