2020
DOI: 10.3390/s20071890
|View full text |Cite
|
Sign up to set email alerts
|

Deep Reinforcement Learning Approach with Multiple Experience Pools for UAV’s Autonomous Motion Planning in Complex Unknown Environments

Abstract: Autonomous motion planning (AMP) of unmanned aerial vehicles (UAVs) is aimed at enabling a UAV to safely fly to the target without human intervention. Recently, several emerging deep reinforcement learning (DRL) methods have been employed to address the AMP problem in some simplified environments, and these methods have yielded good results. This paper proposes a multiple experience pools (MEPs) framework leveraging human expert experiences for DRL to speed up the learning process. Based on the deep determinis… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
20
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 34 publications
(26 citation statements)
references
References 34 publications
0
20
0
Order By: Relevance
“…A state may contain an obstacle or not. Considering the idea is to simulate a real environment, which does not have a specific pattern of obstacle disposition, it is settled to randomly generate obstacles following some rules of degrees of difficulty, as it appears in [21]. Therefore, the RL model will be prepared and assessed in environments with degrees of difficulty: Easy; Medium; Difficult.…”
Section: Environment and Statesmentioning
confidence: 99%
See 1 more Smart Citation
“…A state may contain an obstacle or not. Considering the idea is to simulate a real environment, which does not have a specific pattern of obstacle disposition, it is settled to randomly generate obstacles following some rules of degrees of difficulty, as it appears in [21]. Therefore, the RL model will be prepared and assessed in environments with degrees of difficulty: Easy; Medium; Difficult.…”
Section: Environment and Statesmentioning
confidence: 99%
“…It is the module responsible for planning and (re) defining, in real-time, a route to be determined by the vehicle, from the mapping data, and transmitting this information to the Acting module. For this activity, artificial intelligence techniques, especially reinforcement learning, have been applied, as is the case in [21], in which models based on two Deep Reinforcement Learning techniques are presented called Deep Deterministic Policy Gradient (DDPG) and Multiple Experience Pools DDPG (MEP-DDPG) for planning autonomous movement of aerial unmanned vehicles (AUVs). Likewise [22] presents the application of DDPG for planning land vehicle routes.…”
Section: Introductionmentioning
confidence: 99%
“…For multi-robot motion control, a decentralized RL model is presented to learn a sensor-level collision avoidance policy in multi-robot systems (Fan et al, 2020). Domains like UAVs (Hu et al, 2020;Wan et al, 2020) and underwater vehicles (Carlucho et al, 2018;Chu et al, 2020) have also exploited RL for motion control for various purposes, e.g., robust flying, path planning and remote surveillance. In our work, we present a novel approach of exploiting mobility of the walker and RL techniques for efficient sound source localization (SSL).…”
Section: Introductionmentioning
confidence: 99%
“…Obtaining an optimal policy in reasonable time, taking decisions and actions under large state-spaces using DRL have been applied to network access, wireless caching, cognitive spectrum sensing, and network security. Some of the more recent DRL applications include modeling multiple experience pools for UAV autonomous motion planning in complex unknown environments [ 9 ], learning output reference model tracking for higher-order nonlinear systems with unknown dynamics [ 10 ], and pick and place operations in logistics using a mobile manipulator controlled with DRL [ 11 ]. The DRL paradigm has been extended to domains such as autonomous vehicles and has opened new research avenues [ 12 ].…”
Section: Introductionmentioning
confidence: 99%