2023
DOI: 10.1016/j.dt.2022.04.001
|View full text |Cite
|
Sign up to set email alerts
|

Task assignment in ground-to-air confrontation based on multiagent deep reinforcement learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
2

Relationship

3
4

Authors

Journals

citations
Cited by 19 publications
(10 citation statements)
references
References 7 publications
0
10
0
Order By: Relevance
“…To satisfy the rationality and completeness of the state space and action space and meet the needs of the game confrontation scenario, the state space, action space, and reward function in this paper are designed with reference to the literature [25].…”
Section: Mdp Modelingmentioning
confidence: 99%
See 2 more Smart Citations
“…To satisfy the rationality and completeness of the state space and action space and meet the needs of the game confrontation scenario, the state space, action space, and reward function in this paper are designed with reference to the literature [25].…”
Section: Mdp Modelingmentioning
confidence: 99%
“…We equate the CMDP problem in this paper to an unconstrained max-min optimization problem based on the RL algorithm of the literature [25], combined with the PPO-Lagrangian algorithm [33] to solve.…”
Section: ) Ppo-lagrangianmentioning
confidence: 99%
See 1 more Smart Citation
“…At the end of training, the critic is no longer used. The algorithm for executing the training of the agents is one of the critical issues studied in this paper and will be described in detail in Section “Model-based model predictive control with proximal policy optimization algorithm.” The training method for the scheduling agent refers to the proximal policy optimization for task assignment of general and narrow agents (PPO-TAGNA) algorithm in the literature ( Liu J. Y. et al, 2022 ) to ensure the training effect and demonstrate more intuitively the changes the executive agent brings.…”
Section: Hierarchical Architecture Design For Agentsmentioning
confidence: 99%
“…A centralized assignment solution is not fast enough, while a fully distributed assignment method does not respond effectively to unexpected events ( Lee et al, 2012 ). The one-general agent with multiple narrow agents (OGMN) architecture proposed in the literature ( Liu J. Y. et al, 2022 ), which divides agents into general and narrow agents, improves the computational speed and coordination ability. However, the narrow agent in the OGMN is entirely rule-driven.…”
Section: Introductionmentioning
confidence: 99%