2016 14th International Conference on Control, Automation, Robotics and Vision (ICARCV) 2016
DOI: 10.1109/icarcv.2016.7838739
|View full text |Cite
|
Sign up to set email alerts
|

Autonomous navigation of UAV by using real-time model-based reinforcement learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
42
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 111 publications
(51 citation statements)
references
References 11 publications
0
42
0
Order By: Relevance
“…The UAV motion model is the basis for completing navigation and target tracking missions. The UAV can be thought as a rigid body with forces and torques applied from the four rotors and gravity [22,23]. In navigation and target tracking scenario, we assume that UAV is flying at a fixed altitude.…”
Section: Uav Motion Modelmentioning
confidence: 99%
“…The UAV motion model is the basis for completing navigation and target tracking missions. The UAV can be thought as a rigid body with forces and torques applied from the four rotors and gravity [22,23]. In navigation and target tracking scenario, we assume that UAV is flying at a fixed altitude.…”
Section: Uav Motion Modelmentioning
confidence: 99%
“…Imanberdiyev et al [32] developed a model-based RL algorithm called TEXPLORE to efficiently plan trajectories in unknown environments subject to constraints such as battery life. In [33], the authors use a model predictive controller (MPC) to generate training data for an RL controller, thereby guiding the policy search and avoiding the potentially catastrophic early phase before an effective policy is found.…”
Section: Related Workmentioning
confidence: 99%
“…However, as RL usually requires a lot of trial and error for the learning, having some restrictions on the environment inevitably slows down the training speed. Such characteristics of RL approaches lead to making use of a virtual environment, where the agent can safely make trials and errors quickly without concerning the actual hardware [28,29]. By using sensory data from the accelerometer, it is shown that the RL model can plan swing-free trajectories while carrying suspended load [30].…”
Section: Drone Navigation With Reinforcement Learningmentioning
confidence: 99%