2015
DOI: 10.1137/130932284
|View full text |Cite
|
Sign up to set email alerts
|

An Efficient Policy Iteration Algorithm for Dynamic Programming Equations

Abstract: Abstract. We present an accelerated algorithm for the solution of static Hamilton-JacobiBellman equations related to optimal control problems. Our scheme is based on a classic policy iteration procedure, which is known to have superlinear convergence in many relevant cases provided the initial guess is sufficiently close to the solution. This limitation often degenerates into a behavior similar to a value iteration method, with an increased computation time. The new scheme circumvents this problem by combining… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
84
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
7
3

Relationship

6
4

Authors

Journals

citations
Cited by 66 publications
(85 citation statements)
references
References 28 publications
1
84
0
Order By: Relevance
“…Another improvement can be obtained using efficient acceleration methods for the computation of the value function in every subdomain. In the framework of optimal control problems an efficient acceleration technique based on the coupling between value and policy iterations has been recently proposed and studied in [2]. The construction of a DP algorithm for time dependent problems has been addressed in [13] where also a-priori error estimates have been studied.…”
mentioning
confidence: 99%
“…Another improvement can be obtained using efficient acceleration methods for the computation of the value function in every subdomain. In the framework of optimal control problems an efficient acceleration technique based on the coupling between value and policy iterations has been recently proposed and studied in [2]. The construction of a DP algorithm for time dependent problems has been addressed in [13] where also a-priori error estimates have been studied.…”
mentioning
confidence: 99%
“…Therefore, the algorithm is a way to enhance PI with both efficiency and robustness features. We refer to [1] for a detailed description of the algorithm. Finally, we note that in both algorithms we penalize the value function outside of the numerical domain to impose state constraints boundary conditions.…”
Section: Numerical Methods For Dynamic Programming Equationsmentioning
confidence: 99%
“…To avoid this issue the Empirical Interpolation Method (EIM, [7]) and Discrete Empirical Interpolation Method (DEIM, [12]) were introduced. The computation of the POD basis functions for the nonlinear part is related to the set of the snapshots F (t j , y(t j )), where y(t j ) are already computed from (1). We denote by Φ ∈ R d×k the POD basis functions of rank k min{d, N + 1} of the nonlinear part.…”
Section: Pod For the State Equationmentioning
confidence: 99%