2022
DOI: 10.1002/rnc.6349
|View full text |Cite
|
Sign up to set email alerts
|

Performance‐guaranteed containment control for pure‐feedback multi agent systems via reinforcement learning algorithm

Abstract: In this article, a performance-guaranteed containment control scheme based on reinforcement learning (RL) algorithm is proposed for a class of pure-feedback multi agent systems (MASs) with unmeasurable states. The unknown nonlinear functions are approximated by the neural networks (NNs) and an adaptive NN state observer is designed for the states estimation. Based on estimated states, the algebraic loop problem can be removed by introducing filtered signals, and the actor-critic architecture of RL algorithm is… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 53 publications
0
2
0
Order By: Relevance
“…The nonlinear filtering of the intermediate control signals prevents the complexity explosion and achieves convergence within a predefined time period. Alternatively, the work in [20] proposes a novel containment control scheme for multi-agent systems using reinforcement learning and neural networks. It addresses unmeasurable states with an adaptive observer and filtered signals.…”
Section: Related Workmentioning
confidence: 99%
“…The nonlinear filtering of the intermediate control signals prevents the complexity explosion and achieves convergence within a predefined time period. Alternatively, the work in [20] proposes a novel containment control scheme for multi-agent systems using reinforcement learning and neural networks. It addresses unmeasurable states with an adaptive observer and filtered signals.…”
Section: Related Workmentioning
confidence: 99%
“…For the high-order nonlinear multi-agent system containing uncertainty, the optimal consistency control problem with specified performance is considered in [26], where the stability of a closed-loop system and the convergence of consensus errors within a certain range are proved. Based on an actor-critic network, the optimal control is investigated for robot [27] and pure feedback system [28] separately. It is noted that the above controllers rely on multiple neural networks to ensure the stability and optimality of the system.…”
Section: Introductionmentioning
confidence: 99%
“…(2) Different from the optimal control derived by the actor-critic ADP framework in [24,[26][27][28], a critic-only NN was designed for online learn optimal control without constructing an actor NN. In addition, to achieve the convergence rate with a preassigned region, a novel value function was minimized, such that tracking errors can evolve within the prescribed region with low control consumption.…”
Section: Introductionmentioning
confidence: 99%