2022
DOI: 10.1155/2022/3551508
|View full text |Cite
|
Sign up to set email alerts
|

Deep Reinforcement Learning for UAV Intelligent Mission Planning

Abstract: Rapid and precise air operation mission planning is a key technology in unmanned aerial vehicles (UAVs) autonomous combat in battles. In this paper, an end-to-end UAV intelligent mission planning method based on deep reinforcement learning (DRL) is proposed to solve the shortcomings of the traditional intelligent optimization algorithm, such as relying on simple, static, low-dimensional scenarios, and poor scalability. Specifically, the suppression of enemy air defense (SEAD) mission planning is described as a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
1

Relationship

1
7

Authors

Journals

citations
Cited by 10 publications
(7 citation statements)
references
References 17 publications
0
7
0
Order By: Relevance
“…ML algorithms are used in increment Path efficiency for multi objectives. Sequential decisionmaking problem-solving based on intelligent mission planning methods [37] can be used in the Suppression of Enemy Air Defense (SEAD). The experiments study DRL based on a PPO algorithm that considers the basis of DRL to gain high performance and robustness.…”
Section: ) Machine Learning Approaches In Uavmentioning
confidence: 99%
“…ML algorithms are used in increment Path efficiency for multi objectives. Sequential decisionmaking problem-solving based on intelligent mission planning methods [37] can be used in the Suppression of Enemy Air Defense (SEAD). The experiments study DRL based on a PPO algorithm that considers the basis of DRL to gain high performance and robustness.…”
Section: ) Machine Learning Approaches In Uavmentioning
confidence: 99%
“…In the tracking of the vehicle, it is necessary to determine each frame of the vehicle motion, establish a Kalman filter to predict the vehicle position, and let the state vector of the vehicle be s k = ½s k y k ŝk ŷk T . The first two components are the vector values of the vehicle on the x-axis and y-axis, and the last two components are the speed of the vehicle on the x-axis and y-axis [27].…”
Section: Combination Of Mean-shift and Kalman Filteringmentioning
confidence: 99%
“…The simulation sets the mission planning scenario in a square area with size 100 km × 100 km. The jammer and reconnaissance UAVs took off from coordinates (20,20) and (23,23), respectively. The two UAVs passed through three radar detection areas to reach the target protected by ground-to-missile, which is deployed in (80,80) and performed the reconnaissance task when the relative distance to the target is less than 20 km.…”
Section: Experiments Establishmentmentioning
confidence: 99%
“…When considering the suppression of enemy air defense mission planning, Ref. [23] established a general intelligent planning architecture based on the proximal policy optimization (PPO) algorithm, but it was not considered a dynamic situation; the experiments only involved direct strikes without situation changes during flight. Unfortunately, the UAV mission planning designed in these studies has weakened the obstacles and purpose requirements in the navigation mission planning, so there is no guarantee of the generalizability of the DRL agent or the robustness of navigation control, especially for target reconnaissance and jamming.…”
Section: Introductionmentioning
confidence: 99%