2019
DOI: 10.48550/arxiv.1912.08578
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Taming an autonomous surface vehicle for path following and collision avoidance using deep reinforcement learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4

Citation Types

0
4
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 21 publications
0
4
0
Order By: Relevance
“…The optimization-based algorithms can directly find a control law with collision avoidance by optimizing a certain objective function, e.g. model predictive control (MPC) [9] and reinforcement learning (RL) [18]. They potentially have a better performance than the motion planning approaches in dynamic environments.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…The optimization-based algorithms can directly find a control law with collision avoidance by optimizing a certain objective function, e.g. model predictive control (MPC) [9] and reinforcement learning (RL) [18]. They potentially have a better performance than the motion planning approaches in dynamic environments.…”
Section: Introductionmentioning
confidence: 99%
“…They will, therefore, experience dramatic degradation in performances for uncertain ASVs. In comparison to MPC, RL can learn an intelligent collision avoidance law from data samples [18], [19], which can significantly reduce the dependence on modeling efforts and thus make RL very suitable for uncertain ASVs.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…In comparison with model-based methods, RL is capable of learning a control law from data samples using much less model information [22]. Hence, it is more promising in controlling systems subject to massive uncertainties and disturbances as ASVs [12], [14], [23], [24], given the sufficiency and good quality of collected data. Nevertheless, it is challenging for model-free RL to ensure closed-loop stability, though some research attempts have been made [25].…”
Section: Introductionmentioning
confidence: 99%