2014
DOI: 10.1109/tsmc.2013.2295351
|View full text |Cite
|
Sign up to set email alerts
|

Online Synchronous Approximate Optimal Learning Algorithm for Multi-Player Non-Zero-Sum Games With Unknown Dynamics

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

1
67
0

Year Published

2014
2014
2022
2022

Publication Types

Select...
9
1

Relationship

2
8

Authors

Journals

citations
Cited by 217 publications
(69 citation statements)
references
References 46 publications
1
67
0
Order By: Relevance
“…In [43] the optimal learning algorithm based on policy iteration is used to solve a multiplayer nonzero-sum game without the requirement of exact knowledge of dynamical systems. In [74] a near-optimal control scheme is proposed to solve the nonzero-sum differential games of continuous time nonlinear systems by using the neural networks.…”
Section: Chapter 5 Multiplayer Game In Orbitmentioning
confidence: 99%
“…In [43] the optimal learning algorithm based on policy iteration is used to solve a multiplayer nonzero-sum game without the requirement of exact knowledge of dynamical systems. In [74] a near-optimal control scheme is proposed to solve the nonzero-sum differential games of continuous time nonlinear systems by using the neural networks.…”
Section: Chapter 5 Multiplayer Game In Orbitmentioning
confidence: 99%
“…Policy iteration can be employed to reduce the computational cost and continuously update control policies by evaluating the interaction performance [20]. Methods of policy iteration for games with known and unknown dynamics have been developed by several research groups [21], [22], [23]. As mentioned above, however, the human's objective is generally unknown to the robot in a typical human-robot interaction scenario.…”
Section: Introductionmentioning
confidence: 99%
“…Additionally, in the field of optimal control, adaptive dynamic programming (ADP) [18][19][20][21] is a significant and hot topic. Many data-driven and model-free methods based on ADP have been established [22][23][24][25][26][27][28][29][30][31][32]. Different from the above methods, virtual reference feedback tuning (VRFT), which is originally proposed by Guardabassi and Savaresi [33], provides a global solution to a model reference control problem with oneshot off-line data.…”
Section: Introductionmentioning
confidence: 99%