2022
DOI: 10.3390/aerospace9110658
|View full text |Cite
|
Sign up to set email alerts
|

Autonomous Maneuver Decision of Air Combat Based on Simulated Operation Command and FRV-DDPG Algorithm

Abstract: With the improvement of UAV performance and intelligence in recent years, it is particularly important for unmanned aerial vehicles (UAVs) to improve the ability of autonomous air combat. Aiming to solve the problem of how to improve the autonomous air combat maneuver decision ability of UAVs so that it can be close to manual manipulation, this paper proposes an autonomous air combat maneuvering decision method based on the combination of simulated operation command and the final reward value deep deterministi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(9 citation statements)
references
References 42 publications
0
9
0
Order By: Relevance
“…real-world tasks, it is very challenging to design ideal reward functions that apply to all situations. To realize the model in this paper to adapt the reward functions of the three models [9,15,49], we mainly adjust the f(ϕ u ), f(ϕ t ), and f(D) in Equations ( 12) and (13).…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…real-world tasks, it is very challenging to design ideal reward functions that apply to all situations. To realize the model in this paper to adapt the reward functions of the three models [9,15,49], we mainly adjust the f(ϕ u ), f(ϕ t ), and f(D) in Equations ( 12) and (13).…”
Section: Discussionmentioning
confidence: 99%
“…In this experiment, we reproduce the models and algorithms in three papers [9,15,49], and apply the hierarchical reinforcement learning framework established in this paper to learn and train them, respectively, while mapping the reward functions shaped in the three papers in the corresponding sub-state spaces; then, in the air combat environment established in this paper, different models are compared in the same test scenarios, and the performance of the three original models is compared with that of the models after applying HRL. We use the benchmark performance comparison method proposed in Section 4.2 to compare the models proposed in the paper, as shown in Table 1.…”
Section: Validation and Evolution Of The Hierarchical Agentsmentioning
confidence: 99%
See 1 more Smart Citation
“…Product assurance is a refined quality management method. Through the application of system engineering method, we can grasp the key points and weak links in product development and production, implement the responsibilities of key positions, refine and decompose the quality management requirements, identify and control the risks in the whole process, and realize the whole process management and control of product quality [1].…”
Section: Introductionmentioning
confidence: 99%
“…The shaping reward obtained by the inverse reinforcement learning algorithm for addressing reward sparsity effectively enhances air game strategy learning speed. To solve reinforcement learning algorithm convergence speed issues, Li et al [31] improved the deep deterministic policy gradient algorithm [32] (DDPG) convergence speed by proportionally returning the final reward value to other reward values in the same air game process. Targeting the DDPG exploration strategy's insufficient exploration and low data utilization, Wan et al [33] incorporated a heuristic exploration strategy to enhance the algorithm's exploration capacity.…”
Section: Introductionmentioning
confidence: 99%