2023
DOI: 10.1109/tac.2022.3176439
|View full text |Cite
|
Sign up to set email alerts
|

Policy Optimization for Markovian Jump Linear Quadratic Control: Gradient Method and Global Convergence

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(8 citation statements)
references
References 36 publications
(47 reference statements)
0
8
0
Order By: Relevance
“…The coercive property, compactness of the sublevel set, and L-smoothness of the cost function in the SOF problem, can be deemed as partially observed counterparts to the properties of the state-feedback LQR cost. The associated proofs follow similar lines as the state-feedback LQR case [12], [19]. Different from these properties, to the best of our knowledge, we are the first to establish the M -Lipschitz continuous Hessian in both SOF and state-feedback LQR problems.…”
Section: Gradients and Hessianmentioning
confidence: 65%
See 4 more Smart Citations
“…The coercive property, compactness of the sublevel set, and L-smoothness of the cost function in the SOF problem, can be deemed as partially observed counterparts to the properties of the state-feedback LQR cost. The associated proofs follow similar lines as the state-feedback LQR case [12], [19]. Different from these properties, to the best of our knowledge, we are the first to establish the M -Lipschitz continuous Hessian in both SOF and state-feedback LQR problems.…”
Section: Gradients and Hessianmentioning
confidence: 65%
“…In this section, we give the analytical expression for both the gradient and Hessian. The derivations follow similar lines as the state-feedback LQR case [11], [19].…”
Section: Gradients and Hessianmentioning
confidence: 87%
See 3 more Smart Citations