2020
DOI: 10.48550/arxiv.2005.00611
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Neural Lyapunov Control

Abstract: We propose new methods for learning control policies and neural network Lyapunov functions for nonlinear control problems, with provable guarantee of stability. The framework consists of a learner that attempts to find the control and Lyapunov functions, and a falsifier that finds counterexamples to quickly guide the learner towards solutions. The procedure terminates when no counterexample is found by the falsifier, in which case the controlled nonlinear system is provably stable. The approach significantly s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
54
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 36 publications
(54 citation statements)
references
References 11 publications
0
54
0
Order By: Relevance
“…In light of the latter objective, it is desirable to increase the number of stable ReLUs in f N N (x), i.e., ReLUs that always output zero or one. For these ReLUs, we do not need to introduce binary variables in the MIQP (13). As a result, any training method that can increase the number of stable ReLUs can significantly improve the complexity of the verification.…”
Section: Neural Network Trainingmentioning
confidence: 99%
See 2 more Smart Citations
“…In light of the latter objective, it is desirable to increase the number of stable ReLUs in f N N (x), i.e., ReLUs that always output zero or one. For these ReLUs, we do not need to introduce binary variables in the MIQP (13). As a result, any training method that can increase the number of stable ReLUs can significantly improve the complexity of the verification.…”
Section: Neural Network Trainingmentioning
confidence: 99%
“…Since the approximation error bound (4) is written in terms of the ∞ norm, the training loss function is set to the mean absolute error. Furthermore, in order to alleviate the practical computational cost of the MIQP (13) in the analysis phase, we add 1 regularization with weight 10 −4 during training.…”
Section: -D Rational Systemmentioning
confidence: 99%
See 1 more Smart Citation
“…Meanwhile, in the control theory, there exist studies about provable energy-function-based safety guarantee of dynamic systems called the safety certificate, or safety index (Wieland and Allgöwer, 2007;Ames et al, 2014;Chang et al, 2020). These methods first synthesize an energy function such that the safe states are with low energy, and then design control laws satisfying the safe action constraint to make the system dissipate energy (Wei and Liu, 2019).…”
Section: Introductionmentioning
confidence: 99%
“…If there exists feasible control for all states in a safe set to satisfy the safe action constraint dissipating the energy, then the system will never leave the safe set (i.e., forward invariance). Despite its soundness, the safety index synthesis (SIS) by hand is extremely hard for complicated or unknown systems, which stimulates a rapidly growing interest in learning-based SIS (Chang et al, 2020;Saveriano and Lee, 2019;Srinivasan et al, 2020;Qin et al, 2021). These studies assume to know the dynamical models (either white-box or black-box).…”
Section: Introductionmentioning
confidence: 99%