2020 International Conference on Unmanned Aircraft Systems (ICUAS) 2020
DOI: 10.1109/icuas48674.2020.9213856
|View full text |Cite
|
Sign up to set email alerts
|

UAV Target Tracking in Urban Environments Using Deep Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
21
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 40 publications
(26 citation statements)
references
References 17 publications
0
21
0
Order By: Relevance
“…Next, the optimal reward function that minimizes the trajectory tracking error was found, and a reinforcement learning-based controller using this reward function was proposed. In the work of [39], Target Following DQN (TF-DQN), a deep reinforcement learning technique based on DQNs was proposed with a curriculum training framework for the UAV to persistently track the target in the presence of obstacles and target motion uncertainty. For the reward function, a piecewise reward was proposed to enable different rewards according to the status of the collision compared with the noncollision.…”
Section: Literature Surveymentioning
confidence: 99%
“…Next, the optimal reward function that minimizes the trajectory tracking error was found, and a reinforcement learning-based controller using this reward function was proposed. In the work of [39], Target Following DQN (TF-DQN), a deep reinforcement learning technique based on DQNs was proposed with a curriculum training framework for the UAV to persistently track the target in the presence of obstacles and target motion uncertainty. For the reward function, a piecewise reward was proposed to enable different rewards according to the status of the collision compared with the noncollision.…”
Section: Literature Surveymentioning
confidence: 99%
“…The framework consists of a global planner based on a modern online Partially Observable Markov Decision Process (POMDP) solver and a local continuous-environment exploration controller based on a DRL method. In [ 25 ], the authors proposed a target following method based on deep Q-networks, considering visibility obstruction from obstacles and uncertain target motion. In [ 26 ], the authors proposed a DRL-based method to enable a robot to explore unknown cluttered urban environments, in which a deep network with convolutional neural network (CNN) [ 27 ] was trained by asynchronous advantage actor-critic (A3C) approach to generate appropriate frontier locations.…”
Section: Introductionmentioning
confidence: 99%
“…Prior work on target tracking using UAVs has extensively covered a diverse range of training strategies and formulations in the last two decades. One of the most prevailing co-herent approaches to target tracking is via the development of guidance laws [Wise and Rysdyk, 2006;Choi and Kim, 2014;Oh et al, 2013;Regina and Zanzi, 2011;Chen et al, 2009;Theodorakopoulos and Lacroix, 2008;Pothen and Ratnoo, 2017]. In principle, the motion model for the target should be known apriori in order to design these laws that satisfy the FOV constraints.…”
Section: Introductionmentioning
confidence: 99%
“…In this paper, we also aim to utilize a simple reinforcement learning technique, modifying it for our specific task. For this, we extend a deep reinforcement learning approach called Target-Following Deep Q-Network (TF-DQN) [Bhagat and Sujit, 2020] to a Double DQN [Hasselt et al, 2016] that we refer to as TF-DDQN. We also propose a target tracking evaluation scheme that can be utilized to quantify the performance of any given target tracking algorithm based on factors like deviation from target's trajectory, proximity to checkpoints placed along the target's trajectory, and computational resources required for training and evaluation.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation