2020
DOI: 10.1109/tvt.2020.2979493
|View full text |Cite
|
Sign up to set email alerts
|

Stabilization Approaches for Reinforcement Learning-Based End-to-End Autonomous Driving

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 47 publications
(17 citation statements)
references
References 22 publications
0
14
0
Order By: Relevance
“…Thus, it automatically adapts to the 4 Wireless Communications and Mobile Computing new environment. Moreover, it allows online learning to support real-time signal communication between V2V, V2I, and V2X links [28]. Hence, it takes advantages in a dynamic vehicular network.…”
Section: Background Of Deep Reinforcement Learningmentioning
confidence: 99%
“…Thus, it automatically adapts to the 4 Wireless Communications and Mobile Computing new environment. Moreover, it allows online learning to support real-time signal communication between V2V, V2I, and V2X links [28]. Hence, it takes advantages in a dynamic vehicular network.…”
Section: Background Of Deep Reinforcement Learningmentioning
confidence: 99%
“…For instance, Li et al [26] proposed a combined trajectory planning and tracking algorithm for vehicle control under the effects of the traffic environments and human driving styles. Chen et al, [27] suggested two techniques to improve the stability of the policy model training with as little manual data as possible on endto-end autonomous driving. Chen et al [28] developed a deep Monte Carlo Tree Search (deep-MCTS) control method for vision-based autonomous driving for predicting driving maneuvers to assist in enhancing the stability and performance of driving control.…”
Section: Literature Reviewmentioning
confidence: 99%
“…The selection strategy of 𝜌 L and Δ is based on the relation between u L and u K , see (5). Due to the expression I n×n •(𝜌 * L J 1×n )u K in (3) the control input u i is independent of u K, j , for all i ≠ j; i, j ∈ [i; n], as it is detailed in Section 2.…”
Section: Calculation Of the Scheduling Variable And The Measured Disturbancementioning
confidence: 99%
“…In the following, the representation of the system ( 15) is used for the design of the robust LPV control (𝜌 K , y K ), see (2). The purpose of the design is to derive the LPV controller which guarantees a minimum performance level for the closed-loop system, considering the predefined control rule (5). The output of the LPV controller u K is used in the expression u = I n×n •(𝜌 L J 1×n )u K + Δ L .…”
Section: Iterative Design Of the Lpv Controlmentioning
confidence: 99%
See 1 more Smart Citation