2019
DOI: 10.1016/j.robot.2019.01.003
|View full text |Cite
|
Sign up to set email alerts
|

Advanced planning for autonomous vehicles using reinforcement learning and deep inverse reinforcement learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
84
0
1

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 212 publications
(85 citation statements)
references
References 39 publications
0
84
0
1
Order By: Relevance
“…The algorithms [130] without any prior knowledge and starting to play random gains, in 24 hours, achieved super-human level performance in chess, shogi as well as Go and convincingly defeated a work-champion program in each case. Since then the algorithm has been applied to solve more engineering problems like advanced planning of autonomous vehicles [131], lung cancer detection in medical treatment [132], smart agriculture [133], UAV cluster task scheduling [134], chatbots [135], autonomous building energy assessment [136].…”
Section: B Data-driven Modelingmentioning
confidence: 99%
“…The algorithms [130] without any prior knowledge and starting to play random gains, in 24 hours, achieved super-human level performance in chess, shogi as well as Go and convincingly defeated a work-champion program in each case. Since then the algorithm has been applied to solve more engineering problems like advanced planning of autonomous vehicles [131], lung cancer detection in medical treatment [132], smart agriculture [133], UAV cluster task scheduling [134], chatbots [135], autonomous building energy assessment [136].…”
Section: B Data-driven Modelingmentioning
confidence: 99%
“…where r t s is the safety penalty, r t e is the reward for agent's driving efficiency, r t t is the penalty for deviation from task's target and v i is the velocity of the i's traffic participant. As frequent lane changes can cause danger in traffic flow, a small penalty is set in r t s of (14) to discourage unnecessary lane changes. A reasonable agent drives as fast as possible within the speed limit of a lane.…”
Section: Rewards Designmentioning
confidence: 99%
“…To learn a simple and explainable driving policy, either states or actions or both in the discrete semantic form were used [11][12][13][14]. States can be "close to the front vehicle" and "far to the front vehicle" and the actions can be overtake, left change, give way and accelerate, respectively.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Although there exist several optimal control solutions to the problem of overtaking maneuvers [1], [2], [3], [4], machine-learning-based methods have also been successfully applied. A reinforcement-learning-based overtaking control strategy is proposed in [5], [6]. In [7] a Q-learning strategy is used in the design of driving algorithms for multi-lane environments.…”
Section: Introductionmentioning
confidence: 99%