2022
DOI: 10.1109/tsmc.2020.3003224
|View full text |Cite
|
Sign up to set email alerts
|

Adaptive Identifier-Critic-Based Optimal Tracking Control for Nonlinear Systems With Experimental Validation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
34
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
8
1

Relationship

1
8

Authors

Journals

citations
Cited by 79 publications
(34 citation statements)
references
References 47 publications
0
34
0
Order By: Relevance
“…With all transition samples participating in the AI-VIRL, this model-free Q-learninglike algorithm benefits from a form of experience replay, widely used in reinforcement learning. Under certain assumptions, convergence of the AI-VIRL C-NN to the optimal controller which implies stability of the closed-loop has been analyzed before in the literature [2,3,[6][7][8]11,12] and is not discussed here.…”
Section: The Ai-virl Solution For the Lrm Output Trackingmentioning
confidence: 99%
See 1 more Smart Citation
“…With all transition samples participating in the AI-VIRL, this model-free Q-learninglike algorithm benefits from a form of experience replay, widely used in reinforcement learning. Under certain assumptions, convergence of the AI-VIRL C-NN to the optimal controller which implies stability of the closed-loop has been analyzed before in the literature [2,3,[6][7][8]11,12] and is not discussed here.…”
Section: The Ai-virl Solution For the Lrm Output Trackingmentioning
confidence: 99%
“…Value Iteration (VI) is one popular approximate dynamic programming [1][2][3][4][5][6][7] and reinforcement learning algorithm [8][9][10][11][12][13], together with Policy Iteration. VI Reinforcement Learning (VIRL) algorithm comes in many implementation flavors, online or offline, offpolicy or on-policy, batch-wise or adaptive-wise, with known or unknown system dynamics.…”
Section: Introductionmentioning
confidence: 99%
“…In this paper, the SMOFCC scheme is designed based on known system dynamics, and it surely can be extended to model-free case as long as the system dynamics is available. To achieve this goal, one strategy is to employ the observer [15] or identifier [48] to estimate the system dynamics, and then directly applied it to propose the control method. While, another way is to develop a pure model-free control method, i.e., the controller is designed directly with system input-output data [28].…”
Section: Remarkmentioning
confidence: 99%
“…Additionally, compared with the data-driven method in [48], this work provides a neural network-based technique to avoid the Kronecker product in estimating the actor/critic term. e actor/critic-based approaches have been discussed in [49,50] for nonlinear affine systems using residual error δ hjb . However, in view of the consideration of the identifier in [49,50], it implies the difference in the computation of residual error δ hjb and training laws in actor/critic weights between the proposed method and the work in [49,50].…”
Section: Arl-based Control Design For Independent Jointsmentioning
confidence: 99%