2019
DOI: 10.1049/iet-cta.2018.5832
|View full text |Cite
|
Sign up to set email alerts
|

Optimal distributed learning for disturbance rejection in networked non‐linear games under unknown dynamics

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
1

Relationship

2
5

Authors

Journals

citations
Cited by 22 publications
(8 citation statements)
references
References 26 publications
0
8
0
Order By: Relevance
“…However, the PE condition might be hard to achieve or even might not be feasible in some scenarios, especially in the context of on-line learning. Concurrent learning [2][3][4][5][6] has emerged as a promising paradigm in the direction that guarantees the exponential convergence of the estimated parameters to their optimal values with relaxing the strict assumption of the PE condition to some easy-to-check verifiable conditions on the richness of data. Concurrent learning technique benefits from recorded experienced data along with current data to replace the PE condition on the regressor with a rank condition on the memory stack of the regressor recorded data.…”
Section: Introductionmentioning
confidence: 99%
“…However, the PE condition might be hard to achieve or even might not be feasible in some scenarios, especially in the context of on-line learning. Concurrent learning [2][3][4][5][6] has emerged as a promising paradigm in the direction that guarantees the exponential convergence of the estimated parameters to their optimal values with relaxing the strict assumption of the PE condition to some easy-to-check verifiable conditions on the richness of data. Concurrent learning technique benefits from recorded experienced data along with current data to replace the PE condition on the regressor with a rank condition on the memory stack of the regressor recorded data.…”
Section: Introductionmentioning
confidence: 99%
“…In control community, several model-based and data driven RL-based feedback controllers have been presented for control of uncertain dynamical systems. [23][24][25][26][27][28] In these traditional RL-based controllers, the reinforcement signal feedback is derived through a fixed quadratic objective function. 29,30 A fixed reward or objective function, however, cannot guarantee achieving desired specifications across all circumstances.…”
Section: Introductionmentioning
confidence: 99%
“…However, most of the works in ADP use known drift dynamics and the persistent excitation (PE) condition to guarantee the parameter convergence, for example, the recent work in [17]. Of course, the assumption of known drift dynamics makes the systems limited, and as shown in [18], PE is difficult to check online and makes the convergence process slow. Most recently, the concurrent learning technique of ADP is employed to relax the PE condition [19, 20] or is even combined with the PE condition to accelerate the convergence [21].…”
Section: Introductionmentioning
confidence: 99%
“…Unfortunately, external disturbance or input constraint is not included. ADP has also been employed to design scriptH optimal control schemes for systems with external disturbances and constrained inputs [17, 18, 22]. In the schemes, there are two policies of two players in a differential game, one for the optimal control law and the other for the worst disturbance law.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation