2018 AIAA Information Systems-Aiaa Infotech @ Aerospace 2018
DOI: 10.2514/6.2018-2134
|View full text |Cite
|
Sign up to set email alerts
|

Flexible Heuristic Dynamic Programming for Reinforcement Learning in Quad-Rotors

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
2
1

Relationship

2
1

Authors

Journals

citations
Cited by 3 publications
(7 citation statements)
references
References 35 publications
0
7
0
Order By: Relevance
“…Moreover, it implies that the shaping aspect of a learning curriculum is not limited to the training distribution only, but also resides in the agent representation. This interpretation of curriculum learning can be recognized in the work by Helmer et al [21] on flexible heuristic dynamic programming.…”
Section: B Curriculum Learningmentioning
confidence: 88%
See 4 more Smart Citations
“…Moreover, it implies that the shaping aspect of a learning curriculum is not limited to the training distribution only, but also resides in the agent representation. This interpretation of curriculum learning can be recognized in the work by Helmer et al [21] on flexible heuristic dynamic programming.…”
Section: B Curriculum Learningmentioning
confidence: 88%
“…Although safe curriculum learning for flight control is still a largely unexplored research area, other (related) paradigms have been investigated that aim to address the said safety and efficiency issues that undermine online reinforcement learning applications. Recently, Helmer et al [21] proposed the framework of flexible heuristic dynamic programming (HDP), and showed how decomposing a target task into a pair of successive tasks of lower complexity can expedite learning in the context of flight control. The concept of flexible function approximation can be regarded as a form of transfer learning (TL), which is a branch of RL research that specifically deals with the question of how the large data dependency of the reinforcement learning paradigm can be reduced by transferring knowledge obtained from training on one task to another task in which learning has not yet taken place [22].…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations