2020
DOI: 10.48550/arxiv.2002.08625
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Semiglobal optimal feedback stabilization of autonomous systems via deep neural network approximation

Abstract: A learning approach for optimal feedback gains for nonlinear continuous time control systems is proposed and analysed. The goal is to establish a rigorous framework for computing approximating optimal feedback gains using neural networks. The approach rests on two main ingredients. First, an optimal control formulation involving an ensemble of trajectories with 'control' variables given by the feedback gain functions. Second, an approximation to the feedback functions via realizations of neural networks. Based… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(7 citation statements)
references
References 21 publications
0
7
0
Order By: Relevance
“…Our work stems from the same framework as [23], which approximates the feedback control with an NN then optimizes the control cost on a distribution of initial states. The authors also provide a theoretical analysis of OC solutions via NN approximations.…”
Section: A High-dimensional Deterministic Optimal Controlmentioning
confidence: 99%
See 3 more Smart Citations
“…Our work stems from the same framework as [23], which approximates the feedback control with an NN then optimizes the control cost on a distribution of initial states. The authors also provide a theoretical analysis of OC solutions via NN approximations.…”
Section: A High-dimensional Deterministic Optimal Controlmentioning
confidence: 99%
“…The HJ t penalizer arises from the first equation in (21), whereas HJ fin and HJ grad are direct results of the final-time condition in (21) and its gradient, respectively. Penalizers prove helpful in problems similar to (23) [8]- [10], [47]. These penalizers improve the training convergence (Sec.…”
Section: A Main Formulationmentioning
confidence: 99%
See 2 more Smart Citations
“…In the literature there are different techniques to tackle this situation. Related recent works include, polynomial approximation [28], deep neural technique [31], tensor calculus [18], Taylor series expansions [13] and graph-tree structures [5].…”
Section: Introductionmentioning
confidence: 99%