2022
DOI: 10.1109/tcyb.2021.3108034
|View full text |Cite
|
Sign up to set email alerts
|

Hamiltonian-Driven Adaptive Dynamic Programming With Approximation Errors

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
52
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 75 publications
(56 citation statements)
references
References 47 publications
0
52
0
Order By: Relevance
“…In order to verify the correctness of the cantilever beam model established in this paper, the natural frequencies the first three modals are obtained by the following method: we assume that the initial displacement at node x 6 is 0.028 m, the open-loop response of the system can be obtained from (18). Figure 4 shows the open-loop time-domain response and frequency-domain response of node x 6 .…”
Section: Simulation Of Frequency Domain Responsementioning
confidence: 99%
See 1 more Smart Citation
“…In order to verify the correctness of the cantilever beam model established in this paper, the natural frequencies the first three modals are obtained by the following method: we assume that the initial displacement at node x 6 is 0.028 m, the open-loop response of the system can be obtained from (18). Figure 4 shows the open-loop time-domain response and frequency-domain response of node x 6 .…”
Section: Simulation Of Frequency Domain Responsementioning
confidence: 99%
“…The main control algorithms used in the active vibration control of piezoelectric cantilever beam are as follows: PD control based on optimal position feedback, 15 robust control, 16,17 adaptive control, [18][19][20][21][22][23][24][25][26][27][28][29] fuzzy control, 30 artificial neural network control, 31 LQR control method [32][33][34][35][36][37][38] etc. The purpose of control is to quickly suppress bending and torsional modal vibration.…”
Section: Introductionmentioning
confidence: 99%
“…In Reference 30, an integral reinforcement learning method incorporating experience replay technique is presented to solve the optimal control problem of partially unknown constrained‐input systems. Hamiltonian‐driven ADP method is proposed in Reference 31 and used to solve the infinite‐horizon optimal control problem of continuous‐time nonlinear systems in Reference 32. In Reference 33, consider a class of continuous time nonlinear systems with unmodeled dynamics, a robust actor‐critic RL method is presented.…”
Section: Introductionmentioning
confidence: 99%
“…For example, Hordijk and Kallenberg [17] used linear programming to nd the average optimal strategy of general Markov decision processes (abbreviated as MDPs), Denard and fox [18] obtained the algorithm of linear programming to solve some special cases of nondiscounted semi-Markov decision processes (semi-MDP), and Vrieze [19] gave a linear programming algorithm for nondiscounted stochastic games with nite state and action space. Recently, Yang et al [9,20,21] proposed some new algorithms for optimal control problems, such as a novel adaptive optimal control design method for the constrained control problem in [9], an iterative adaptive dynamic programming (ADP) algorithm for the infinite-horizon optimal control problem in continuous time for nonlinear systems in [20], and λ-policy iteration (λ-PI) for the discrete-time linear quadratic regulation problem in [21].…”
Section: Introductionmentioning
confidence: 99%