2022
DOI: 10.1109/tits.2021.3086033
|View full text |Cite
|
Sign up to set email alerts
|

Model-Reference Reinforcement Learning for Collision-Free Tracking Control of Autonomous Surface Vehicles

Abstract: Qingrui Zhang (Member, IEEE) received the B.S. degree in automatic control from the Harbin Institute of Technology, Harbin, China, in 2013, and the Ph.D. degree in aerospace science and engineering from the

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
19
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 43 publications
(23 citation statements)
references
References 47 publications
0
19
0
Order By: Relevance
“…Reinforcement Learning for uncertainty: Compared with the existing GP-based and neuralnetwork-based approaches, RL, an adaptive and interactive learning approach, is introduced to model highly dynamic uncertainties in recent work [15]. Distributional RL constructs the entire distributions of the action-value function instead of the traditional expectation, where, to some extent, it addresses the key challenge of traditional RL, i.e., biasing the actions with high variance values in policy optimization [9].…”
Section: Related Workmentioning
confidence: 99%
“…Reinforcement Learning for uncertainty: Compared with the existing GP-based and neuralnetwork-based approaches, RL, an adaptive and interactive learning approach, is introduced to model highly dynamic uncertainties in recent work [15]. Distributional RL constructs the entire distributions of the action-value function instead of the traditional expectation, where, to some extent, it addresses the key challenge of traditional RL, i.e., biasing the actions with high variance values in policy optimization [9].…”
Section: Related Workmentioning
confidence: 99%
“…Model Uncertainties: They are usually caused by the unknown dynamics [134]- [136], underactuated ASVs [137], high-speed maneuvering situation [138], sensor errors [135], or environmental disturbances [139], which may introduce unknown parameters, terms or functions into an ASV control system.…”
Section: A Definition and Key Problemsmentioning
confidence: 99%
“…In addition, based on DRL, the control law can be learned directly to compensate for uncertainties and disturbances [139]. The reward functions of DRL have a very large impact on the learned desired behavior.…”
Section: The Limitation Of DLmentioning
confidence: 99%
“…In comparison to existing data-driven approaches, Reinforcement Learning (RL), an interactive learning process, is able to learn complex and changeable disturbances -i.e., the errors between the true and estimated values -using much less model information [14]. The key challenge of most existing RL approaches [15] is that policy optimization biases toward actions with high variance value estimates, since some of these values will be overestimated by random chance [16].…”
Section: Introduction Accurate Trajectory Tracking For Autonomous Unm...mentioning
confidence: 99%