2018 International Conference on Unmanned Aircraft Systems (ICUAS) 2018
DOI: 10.1109/icuas.2018.8453449
|View full text |Cite
|
Sign up to set email alerts
|

Toward End-to-End Control for UAV Autonomous Landing via Deep Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
34
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
4
1

Relationship

1
9

Authors

Journals

citations
Cited by 73 publications
(41 citation statements)
references
References 24 publications
0
34
0
Order By: Relevance
“…Their controller generalizes to avoid multiple obstacles, compared to the singular obstacle avoided by the MPC in training, does not require full state information like the MPC does, and is computed at a fraction of the time. With the advent of DRL, it has also been used for more advanced tasks such as enabling intelligent cooperation between multiple UAVs [34], and for specific control problems such as landing [35]. RL algorithms have also been proposed for attitude control of other autonomous vehicles, including satellites [36] and underwater vehicles.…”
Section: Related Workmentioning
confidence: 99%
“…Their controller generalizes to avoid multiple obstacles, compared to the singular obstacle avoided by the MPC in training, does not require full state information like the MPC does, and is computed at a fraction of the time. With the advent of DRL, it has also been used for more advanced tasks such as enabling intelligent cooperation between multiple UAVs [34], and for specific control problems such as landing [35]. RL algorithms have also been proposed for attitude control of other autonomous vehicles, including satellites [36] and underwater vehicles.…”
Section: Related Workmentioning
confidence: 99%
“…Learning-based control methods for autonomous landing have also been studied to achieve the optimal control policy under uncertainties. Polvara et al [21] have proposed an approach based on a hierarchy of deep Q-networks (DQNs) that can be used as a high-end control policy for the navigation in different phases. With an optimal policy, they have demonstrated a quadcopter autonomously landing in a large variety of simulated environments.…”
Section: Related Workmentioning
confidence: 99%
“…However, filling the gap between real and simulated experiences is everything but simple. To reduce this gap, we built on top of our previous work [12], and we used domain randomization (DR) [10] to improve the generalization capabilities of the DQNs via random sampling of training environments. We show that, when the variability is large enough, the networks learn to generalize well across a large variety of unseen scenarios, including real ones.…”
Section: Introductionmentioning
confidence: 99%