2023
DOI: 10.23919/jsee.2023.000056
|View full text |Cite
|
Sign up to set email alerts
|

Deep reinforcement learning for UAV swarm rendezvous behavior

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 10 publications
0
2
0
Order By: Relevance
“…The above equation covers 10-dimensional vectors exhibiting different magnitudes and different degrees of sparsity, such that the degree of value change in 1 reconnaissance cycle is different. To be able to be sensitive to the changes of each vector and accelerate the convergence speed of the algorithm, the 10-dimensional vectors need to be normalised, as expressed in Equation (18). In Scouter, the mission area of the UAV is a rectangular area of 75 km (length) � 75 km (width) � 6 km (height), X max is 37.5 km, X min is −37.5 km, Y max is 37.5 km, Y min is −37.5 km; H max is 12 km, H min is 6 km, P u;min is the UAV reconnaissance payload sensitivity.…”
Section: Reward Shapingmentioning
confidence: 99%
See 1 more Smart Citation
“…The above equation covers 10-dimensional vectors exhibiting different magnitudes and different degrees of sparsity, such that the degree of value change in 1 reconnaissance cycle is different. To be able to be sensitive to the changes of each vector and accelerate the convergence speed of the algorithm, the 10-dimensional vectors need to be normalised, as expressed in Equation (18). In Scouter, the mission area of the UAV is a rectangular area of 75 km (length) � 75 km (width) � 6 km (height), X max is 37.5 km, X min is −37.5 km, Y max is 37.5 km, Y min is −37.5 km; H max is 12 km, H min is 6 km, P u;min is the UAV reconnaissance payload sensitivity.…”
Section: Reward Shapingmentioning
confidence: 99%
“…In recent years, deep reinforcement learning technology has proved to be successful in many difficult decision-making missions [18][19][20][21]. The trained intelligent agents impress CEW researchers with their excellent decision-making abilities.…”
Section: Introductionmentioning
confidence: 99%
“…Samir et al [15] combined DRL with joint optimization to achieve improved learning efficiency, although changes to the dynamic environment can hinder the implementation of this strategy. Zhang et al [16] investigated a double deep Q-network (DQN) framework for long period UAV swarm collaborative tasks and designed a guided reward function to solve the convergence problem caused by the sparse returns of long period tasks. Huda et al [17] investigated a surveillance application scenario using a hierarchical UAV swarm.…”
Section: Introductionmentioning
confidence: 99%