2022
DOI: 10.1109/tiv.2022.3167271
|View full text |Cite
|
Sign up to set email alerts
|

Receding-Horizon Reinforcement Learning Approach for Kinodynamic Motion Planning of Autonomous Vehicles

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2022
2022
2025
2025

Publication Types

Select...
9

Relationship

1
8

Authors

Journals

citations
Cited by 31 publications
(12 citation statements)
references
References 48 publications
0
10
0
Order By: Relevance
“…Challenges lie in how to collect massive training samples that are consistent and how to guarantee learning efficiency (e.g., free from overfitting). Reinforcement-learning-based methods obtain knowledge by trial-and-error operations, thus they rely less on the quality and quantity of external training samples [168].…”
Section: B Local Behavior/trajectory Planningmentioning
confidence: 99%
“…Challenges lie in how to collect massive training samples that are consistent and how to guarantee learning efficiency (e.g., free from overfitting). Reinforcement-learning-based methods obtain knowledge by trial-and-error operations, thus they rely less on the quality and quantity of external training samples [168].…”
Section: B Local Behavior/trajectory Planningmentioning
confidence: 99%
“…In [12], the unknown disturbances are considered in the kinodynamic motion problem and solved by Pontryagin's Maximum Principle. Zhang et al [13] proposed a receding-horizon RL algorithm with an NN-based uncertain dynamics model for motion planning of intelligent vehicles, while it is generally difficult to collect a sufficient number of samples. In [14], a barrier transformation was incorporated into the RL algorithm for generating optimal control inputs online, but there may exist issues of slow convergence speed and difficulty in selecting appropriate parameters.…”
Section: Related Workmentioning
confidence: 99%
“…RL and ADP have received significant attention in recent years for solving optimal planning and control problems [9], [10]. Previous RL approaches in solving planning problems with uncertain dynamics employ the integration of Pontryagin's Maximum Principle to address model uncertainty and kinodynamic constraints [11], [12], incorporate model learning strategy and barrier function into the cost function to separately handle model uncertainty and collision constraints [13], [14], measure and select the optimal solution among the planning candidates to address the performance degradation caused by model uncertainty [15], [16]. RL approaches necessitate the use of approximators e.g., actor-critic networks to approximate the optimal policy and value function for continuous control tasks, where the feature representation capability greatly impacts learning efficiency and performance [17], [18].…”
Section: Introductionmentioning
confidence: 99%
“…For many physical systems, accurate models play important roles in model-based planning, control, and trajectory forecasting [1][2][3][4]. Therefore, it is meaningful to develop data-driven modelling approaches for systems with unknown dynamics [5,6].…”
Section: Introductionmentioning
confidence: 99%