2021
DOI: 10.3390/s21072534
|View full text |Cite
|
Sign up to set email alerts
|

Deep Reinforcement Learning for End-to-End Local Motion Planning of Autonomous Aerial Robots in Unknown Outdoor Environments: Real-Time Flight Experiments

Abstract: Autonomous navigation and collision avoidance missions represent a significant challenge for robotics systems as they generally operate in dynamic environments that require a high level of autonomy and flexible decision-making capabilities. This challenge becomes more applicable in micro aerial vehicles (MAVs) due to their limited size and computational power. This paper presents a novel approach for enabling a micro aerial vehicle system equipped with a laser range finder to autonomously navigate among obstac… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
9
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
8
1

Relationship

1
8

Authors

Journals

citations
Cited by 23 publications
(12 citation statements)
references
References 29 publications
0
9
0
Order By: Relevance
“…proposed a simple and efficient energy-based method [ 34 ] to prioritize playback of “posterior experience”, innovatively using “trace energy” instead of TD-error as a measure of priority. In tasks such as continuous control [ 35 ], an actor-critic reinforcement learning algorithm is used to achieve autonomous navigation and obstacle avoidance of UAVs, and robustness to unknown environments through localization noise. Junjie Zeng et al.…”
Section: Related Workmentioning
confidence: 99%
“…proposed a simple and efficient energy-based method [ 34 ] to prioritize playback of “posterior experience”, innovatively using “trace energy” instead of TD-error as a measure of priority. In tasks such as continuous control [ 35 ], an actor-critic reinforcement learning algorithm is used to achieve autonomous navigation and obstacle avoidance of UAVs, and robustness to unknown environments through localization noise. Junjie Zeng et al.…”
Section: Related Workmentioning
confidence: 99%
“…Liu et al [19] used RL and probability map to create a search algorithm and improve the detection ability of the algorithm. Finally, deep RL is used for local motion planning in an unknown environment in [20], and for trajectory tracking and altitude control in [21].…”
Section: Related Workmentioning
confidence: 99%
“…Further, compared to works such as [18] where a Deep Q-Network capable of generating discrete action is used, in our work, a policy gradient-based reinforcement learning algorithm capable of generating continuous action is used. Compared to another similar work [20], while we use RGB-D data as the input to our Deep RL and generate 3D-action, they use 2D lidar data and generates 2D action. Further, our goal is an image, and their goal is a point fed to the algorithm.…”
Section: Related Workmentioning
confidence: 99%
“…Kong et al [19] explored the generalization of various DRL algorithm by training them with different (but not unseen) environments. Doukui et al [20] tackle this issue by mapping exteroceptive sensors, robot state, and goal information to continuous velocity control inputs, but their exploration was only tested on unseen targets instead of unseen scenes.…”
Section: Introductionmentioning
confidence: 99%