Robotics: Science and Systems XVII 2021
DOI: 10.15607/rss.2021.xvii.063
|View full text |Cite
|
Sign up to set email alerts
|

Lyapunov-stable neural-network control

Abstract: Deep learning has had a far reaching impact in robotics. Specifically, deep reinforcement learning algorithms have been highly effective in synthesizing neural-network controllers for a wide range of tasks. However, despite this empirical success, these controllers still lack theoretical guarantees on their performance, such as Lyapunov stability (i.e., all trajectories of the closed-loop system are guaranteed to converge to a goal state under the control policy). This is in stark contrast to traditional model… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
19
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 49 publications
(19 citation statements)
references
References 28 publications
(39 reference statements)
0
19
0
Order By: Relevance
“…We can for example restrict the policy to obey stable system dynamics derived from first principles ( Greydanus et al, 2019 ; Lutter et al, 2019 ). Another approach is to design the model class such that the closed-loop system is passive for all parameterizations of the learned policy, thus guaranteeing stability in the sense of Lyapunov as well as bounded output energy given bounded input energy ( Brogliato et al, 2007 ; Yang et al, 2013 ; Dai et al, 2021 ). All these methods would require significant exploration in the environment, making it even more challenging to learn successful controllers in the real-world directly.…”
Section: Discussion and Outlookmentioning
confidence: 99%
“…We can for example restrict the policy to obey stable system dynamics derived from first principles ( Greydanus et al, 2019 ; Lutter et al, 2019 ). Another approach is to design the model class such that the closed-loop system is passive for all parameterizations of the learned policy, thus guaranteeing stability in the sense of Lyapunov as well as bounded output energy given bounded input energy ( Brogliato et al, 2007 ; Yang et al, 2013 ; Dai et al, 2021 ). All these methods would require significant exploration in the environment, making it even more challenging to learn successful controllers in the real-world directly.…”
Section: Discussion and Outlookmentioning
confidence: 99%
“…c) Learned Certifications for Model-based Controllers: Remaining methods for connecting safety and learning are to provide certification on learned stability and constraint set using existing model-based controllers. Stability criterion can be achieved by Lyapunov analysis on region of attraction (RoA) [8, 20,52] or by Lipschitz-based safety [32] during training. Safety with constraint set certifications can be learned using control synthesis such as feedback linearization controllers [68], CBFs [46,53,54,55,61,64,71], and Hamilton-Jacobi (HJ) reachability analysis [5,23,31].…”
Section: A Related Work 1) Safety and Learningmentioning
confidence: 99%
“…The quadratic Lyapunov candidate function (orange) violates the constraint in (2), since avoiding the obstacles implies moving away from the goal, while a valid Lyapunov function (blue) is monotonically decreasing along the demonstration path (red). Instead of explicitly specifying the Lyapunov candidate function, prior work [12] learns a function that satisfies the Lyapunov conditions in (1)- (3). In this work, we will adapt this structure to learn the Lyapunov function.…”
Section: A Lyapunov Functionmentioning
confidence: 99%
“…We follow the formulation in [12] to structure a Lyapunov candidate function as follows: The policy can be taken to be related to the negative gradient of the learned Lyapunov function…”
Section: B Value Function and Policy Learningmentioning
confidence: 99%
See 1 more Smart Citation