The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2020 16th International Conference on Control, Automation, Robotics and Vision (ICARCV) 2020
DOI: 10.1109/icarcv50220.2020.9305467
|View full text |Cite
|
Sign up to set email alerts
|

Design and Verification of UAV Maneuver Decision Simulation System Based on Deep Q-learning Network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 12 publications
(4 citation statements)
references
References 15 publications
0
4
0
Order By: Relevance
“…In particular, UAV pilots that utilize reinforcement learning (RL) are more flexible than rule-based pilots. With RL, UAV pilots have advanced to a level in which they can replace humans in decision-making [5,6]. In this situation, the air combat revolution was created B Jung Ho Bae deawith@gmail.com 1 Agency for Defense Development, Daejeon 34186, Korea by the defense advanced research project agency in 2019 to develop AI pilots to replace human pilots [7].…”
Section: Introductionmentioning
confidence: 99%
“…In particular, UAV pilots that utilize reinforcement learning (RL) are more flexible than rule-based pilots. With RL, UAV pilots have advanced to a level in which they can replace humans in decision-making [5,6]. In this situation, the air combat revolution was created B Jung Ho Bae deawith@gmail.com 1 Agency for Defense Development, Daejeon 34186, Korea by the defense advanced research project agency in 2019 to develop AI pilots to replace human pilots [7].…”
Section: Introductionmentioning
confidence: 99%
“…The results showed that the RL agent was able to outperform the baseline agent in terms of survival rate. More recently, Hu et al [18] trained long and short-term memory (LSTM) in a deep Q-network (DQN) framework for air combat maneuvering decisions, and this was more forward-looking and efficient in its decision-making than fully connected neural-network-and statistical-principle-based algorithms [19]. In addition, Li proposed a deep reinforcement learning method based on proximal policy optimization (PPO) to learn combat strategies from observation in an end-to-end manner [20,21], and the adversarial results showed that his PPO agent can beat the adversary with a win rate of approximately 97%.…”
Section: Introductionmentioning
confidence: 99%
“…Because rule-based techniques enable aircraft to perform predetermined maneuvers under given conditions, it is difficult to respond appropriately to unexpected situations [24]. Hence, recent studies have utilized RL techniques that are particularly suited to learning to make decisions quickly in unpredictable or uncertain situations [25][26][27][28][29][30][31]. Masadeh et al [6] utilized multi-agent deep deterministic policy gradient and Bayesian optimization to optimize the trajectory and network formation of UAVs for rapid data transmission and minimize energy consumption and transmission delay in a situation where multiple UAVs are used as repeaters in a wireless network.…”
Section: Introductionmentioning
confidence: 99%