2019
DOI: 10.1016/j.neucom.2019.01.033
|View full text |Cite
|
Sign up to set email alerts
|

Integral reinforcement learning off-policy method for solving nonlinear multi-player nonzero-sum games with saturated actuator

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
6
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
7
1

Relationship

2
6

Authors

Journals

citations
Cited by 26 publications
(6 citation statements)
references
References 24 publications
0
6
0
Order By: Relevance
“…Moreover, the system state x is UUB due to the property of the function ∇L i . Remark 3: Different from the method proposed in [37] that each player needs both critic NN and actor NN, the presented method in this paper requires only critic NN for each player to tackle the coupled HJ equations. By abnegating the actor NN, identifier-critic framework is employed, which can effectively simplify the algorithm complexity and save the computing resources.…”
Section: Stability Analysismentioning
confidence: 99%
See 1 more Smart Citation
“…Moreover, the system state x is UUB due to the property of the function ∇L i . Remark 3: Different from the method proposed in [37] that each player needs both critic NN and actor NN, the presented method in this paper requires only critic NN for each player to tackle the coupled HJ equations. By abnegating the actor NN, identifier-critic framework is employed, which can effectively simplify the algorithm complexity and save the computing resources.…”
Section: Stability Analysismentioning
confidence: 99%
“…The method in [36] also utilized event-triggered mechanism to solve nonlinear H ∞ control issues with constraints. In [37], an IRL method was used to figure out the optimal control policies for players in NZS games with saturated actuator. Nevertheless, this method employed both actor NN and critic NN, which perplexed the algorithm and aggravated the computing burdens.…”
Section: Introductionmentioning
confidence: 99%
“…Considering the constrained control inputs and partially unknown dynamics, 42 designs an online ADP approach to solve the finite‐horizon optimal control for NZS game. Consider unknown system dynamics and saturated actuator, an off‐policy IRL learning mechanism is introduced to obtain the solution of nonlinear NZS game 43 . Reference 44 proposes a data‐driven IRL method to solve the nonlinear optimization problem of NZS games.…”
Section: Introductionmentioning
confidence: 99%
“…Thence, IRL can also be classified into iterative methods for solving the HJB equations. For example, in Ren et al (2019), IRL technology was employed to settle the NZS game in ODE system when the actuators are constrained.…”
Section: Introductionmentioning
confidence: 99%