2023
DOI: 10.1007/978-981-99-0479-2_125
|Get access via publisher |Cite
|
Sign up to set email alerts

2v2 Air Combat Confrontation Strategy Based on Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
1
1
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 6 publications
0
3
0
Order By: Relevance
“…The barrier function serves as a formal safety certificate associated with a control policy, guaranteeing the state-wise safety of a dynamical system [21,22]. Classical control theory often relaxes the stringent conditions of the barrier function into optimization formulations like linear programs [23,24] and quadratic programs [25,26].…”
Section: State-wise Safe Reinforcement Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…The barrier function serves as a formal safety certificate associated with a control policy, guaranteeing the state-wise safety of a dynamical system [21,22]. Classical control theory often relaxes the stringent conditions of the barrier function into optimization formulations like linear programs [23,24] and quadratic programs [25,26].…”
Section: State-wise Safe Reinforcement Learningmentioning
confidence: 99%
“…Recent research has explored the joint learning of control policies and neural barrier functions to optimize state-wise safety constraints in reinforcement learning [27][28][29]. In the context of autonomous driving, ShieldNN [30] leverages CBF to design a safety filter neural network, providing safety assurances for environments with known bicycle dynamics models.…”
Section: State-wise Safe Reinforcement Learningmentioning
confidence: 99%
“…Various deep learning methods have addressed constraint learning through techniques such as backpropagation over (in)equality completions [27], differentiable projection layers that map interior points to boundary regions [43], convex programming layers [1], and problem-specific repair mechanisms [22]. In addition, safe reinforcement learning has approached feasibility through constrained MDPs with primal-dual techniques [26], soft barrier functions [79], and safety shields [3].…”
Section: Machine Learning For Optimization Problemsmentioning
confidence: 99%