2021
DOI: 10.3390/app112210595
|View full text |Cite
|
Sign up to set email alerts
|

Hierarchical Active Tracking Control for UAVs via Deep Reinforcement Learning

Abstract: Active tracking control is essential for UAVs to perform autonomous operations in GPS-denied environments. In the active tracking task, UAVs take high-dimensional raw images as input and execute motor actions to actively follow the dynamic target. Most research focuses on three-stage methods, which entail perception first, followed by high-level decision-making based on extracted spatial information of the dynamic target, and then UAV movement control, using a low-level dynamic controller. Perception methods b… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(11 citation statements)
references
References 24 publications
0
11
0
Order By: Relevance
“…Nonetheless, the MAV dynamics are not fully exploited in the design of the control policy and the action space is discrete, which poses a hard limit on the achievable performance. A continuous action space is considered in [12], where a RL-based policy is coupled with a low-level PID control layer. However, the positioning of the MAV is constrained to a plane and thus the tracker is not free to move in 3D.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Nonetheless, the MAV dynamics are not fully exploited in the design of the control policy and the action space is discrete, which poses a hard limit on the achievable performance. A continuous action space is considered in [12], where a RL-based policy is coupled with a low-level PID control layer. However, the positioning of the MAV is constrained to a plane and thus the tracker is not free to move in 3D.…”
Section: Related Workmentioning
confidence: 99%
“…Model-based techniques (see, e.g., [23]) present design and integration issues that inherently limit their performance and entail tracking errors that may limit their applicability. On the other hand, existing learningbased approaches are affected by different constraints: (i) the target lies on a plane [22]; (ii) the tracker is controlled by discrete actions [11]; (iii) the agent is trained with continuous actions that are confined to a subset of the tracker action space [12]. To overcome these limitations, in this paper we provide the following contributions:…”
Section: A Contributionmentioning
confidence: 99%
See 1 more Smart Citation
“…In the event of a UAV failure or inability of the UAV to return to its "home" (initial) position, in the event of an accident or other loss (shooting down) of the UAV in such an area, it is essential to obtain to the location as soon as possible. Otherwise, the collected data that carry useful equipment, or even parts of the UAV, could be misused by the other party [8,9]. Since this is a specific, local area in which the UAV operates, it is possible to create alternative ways of tracing and detecting a particular target [10].…”
Section: Introductionmentioning
confidence: 99%
“…They are usually combined with reinforcement learning (RL) and have achieved great success [6][7][8][9]. Many advanced deep RL algorithms have been proposed and used to solve complex decision-making problems such as game playing [9,10], robotics [11][12][13][14], and autonomous driving [15][16][17][18]. In multi-agent systems [19,20], if an opponent's policy is fixed, the agent can treat it as part of the stationary environment and learn the response policy using single-agent RL algorithms [21][22][23].…”
Section: Introductionmentioning
confidence: 99%