2020
DOI: 10.1002/asjc.2276
|View full text |Cite
|
Sign up to set email alerts
|

Infinite time linear quadratic stackelberg game problem for unknown stochastic discrete‐time systems via adaptive dynamic programming approach

Abstract: In this paper, we propose an adaptive dynamic programming (ADP) approach to solve the infinite horizon linear quadratic (LQ) Stackelberg game problem for unknown stochastic discrete‐time systems with multiple decision makers. Firstly, the stochastic LQ Stackelberg game problem is converted into the deterministic problem by system transformation. Next, a value iteration ADP approach is put forword and the convergence is given. Thirdly, in order to implement the iterative method, back propagation neural network … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
14
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 14 publications
(15 citation statements)
references
References 33 publications
0
14
0
Order By: Relevance
“…Situation 1: if t ∈ [𝜏 𝑗 , 𝜏 𝑗+1 ), we can get the derivative of (40) when the events are not triggered as follows; substituting (18) and (19)…”
Section: Stability Analysismentioning
confidence: 99%
See 1 more Smart Citation
“…Situation 1: if t ∈ [𝜏 𝑗 , 𝜏 𝑗+1 ), we can get the derivative of (40) when the events are not triggered as follows; substituting (18) and (19)…”
Section: Stability Analysismentioning
confidence: 99%
“…This structure is similar to reinforcement learning (RL) and adaptive dynamic programming (ADP), so ACL, ADP, and RL are taken as synonyms. Because of the advantages of ACL technology, it has been widely used in continuous-times control [16,17], discrete-times control [18,19], robust control [20,21], fault-tolerant control [22], and many other fields [23]. Therefore, scholars have adopted ACL-based methods to design optimal security controllers and promoted the rapid development of this field in recent years.…”
Section: Introductionmentioning
confidence: 99%
“…Remark Note that when the control system in problem () is only disturbed by random variables or uncertain variables, the uncertain random system becomes stochastic system or uncertain system, and the LQ optimal control problems of the separate system [4,12,13,28] have been tackled in recent years. Compared with previous studies [4,12,13,28], we use the dynamic programming to solve optimal control problems. We study optimal control for uncertain random systems.…”
Section: Linear Quadratic Modelmentioning
confidence: 99%
“…In recent years, most of the researches on optimal control problems with stochastic noises or uncertain noises [27,28]. However, the system in practice is disturbed by not only stochastic noises but also uncertain noises [26].…”
Section: Introductionmentioning
confidence: 99%
“…The aim of every player in aggregative games is to find a Nash equilibrium to minimize its payoff function which not only depends its own decision variable but also the aggregate of all players' decision variables. To achieve this, some distributed Nash equilibrium seeking algorithms have been proposed [5][6][7][8][9][10][11][12][13][14][15][16]. The paper [16] proposed distributed consensus-based strategies to seek the Nash equilibrium of the box-constrained aggregative games.…”
Section: Introductionmentioning
confidence: 99%