2020
DOI: 10.1049/iet-its.2019.0317
|View full text |Cite
|
Sign up to set email alerts
|

Hierarchical reinforcement learning for self‐driving decision‐making without reliance on labelled driving data

Abstract: Decision making for self-driving cars is usually tackled by manually encoding rules from drivers' behaviors or imitating drivers' manipulation using supervised learning techniques. Both of them rely on mass driving data to cover all possible driving scenarios. This paper presents a hierarchical reinforcement learning method for decision making of self-driving cars, which does not depend on a large amount of labeled driving data. This method comprehensively considers both high-level maneuver selection and low-l… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
56
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
6
3
1

Relationship

1
9

Authors

Journals

citations
Cited by 153 publications
(62 citation statements)
references
References 33 publications
(43 reference statements)
0
56
0
Order By: Relevance
“…Global domain experts have also researched the factors that need to be considered for driverless automobiles to make decisions. Duan proposed a hierarchical reinforcement learning method for self-driving vehicle decisions, which was applied to a highway driving scene [ 13 ]. The Carnegie Mellon Navlab vehicles combine a rationally distributed system (PolySAPIENT) with a novel evolutionary optimization strategy (PBIL) and use route-level planning to achieve context-sensitive local decision-making and complex motion planning, using reasonable behavior theory to decide driving behavior [ 14 , 15 , 16 , 17 ].…”
Section: Introductionmentioning
confidence: 99%
“…Global domain experts have also researched the factors that need to be considered for driverless automobiles to make decisions. Duan proposed a hierarchical reinforcement learning method for self-driving vehicle decisions, which was applied to a highway driving scene [ 13 ]. The Carnegie Mellon Navlab vehicles combine a rationally distributed system (PolySAPIENT) with a novel evolutionary optimization strategy (PBIL) and use route-level planning to achieve context-sensitive local decision-making and complex motion planning, using reasonable behavior theory to decide driving behavior [ 14 , 15 , 16 , 17 ].…”
Section: Introductionmentioning
confidence: 99%
“…Compared to classic autonomous drive systems with environment perception, path planning, and dynamics control modules [11], the DRL method combines deep neural networks with a reinforcement learning frame. This combination can help UGV achieve a faster response to more complex traffic scenarios by learning strategies from high-dimensional perceptual input through the end-to-end model [12]. Therefore, the application of DRL in the UGV field is attracting an increasing research focus.…”
Section: Introductionmentioning
confidence: 99%
“…Some works address decentralized structure in the controller and/or apply data-based controller design. Although the works [4,5] show the effectiveness of decentralized control and data-based design, it is not always realistic to completely replace the existing reliable controller with a new one. As is pointed out in [6,7], the safety of the control systems may be impaired during the update of the controller.…”
Section: Introductionmentioning
confidence: 99%