2022
DOI: 10.1109/lra.2022.3142743
|View full text |Cite
|
Sign up to set email alerts
|

Sablas: Learning Safe Control for Black-Box Dynamical Systems

Abstract: Control certificates based on barrier functions have been a powerful tool to generate probably safe control policies for dynamical systems. However, existing methods based on barrier certificates are normally for white-box systems with differentiable dynamics, which makes them inapplicable to many practical applications where the system is a black-box and cannot be accurately modeled. On the other side, model-free reinforcement learning (RL) methods for black-box systems suffer from lack of safety guarantees a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 18 publications
(7 citation statements)
references
References 28 publications
0
7
0
Order By: Relevance
“…• Classical control: We implement a model predictive control baseline via the CasADi solver [30]. • Safe control: Constrained policy optimization (CPO) [31] and a model-based safe control approach SABLAS [32]. The motivation for our choice of a diverse set of baseline algorithms stems from the perspective of a user trying to solve the problem of safe navigation, who would likely select one of these these established approaches found in the robotics literature.…”
Section: B Comparison With Baseline Methodsmentioning
confidence: 99%
“…• Classical control: We implement a model predictive control baseline via the CasADi solver [30]. • Safe control: Constrained policy optimization (CPO) [31] and a model-based safe control approach SABLAS [32]. The motivation for our choice of a diverse set of baseline algorithms stems from the perspective of a user trying to solve the problem of safe navigation, who would likely select one of these these established approaches found in the robotics literature.…”
Section: B Comparison With Baseline Methodsmentioning
confidence: 99%
“…In this case, L V is used to self-supervise both the certificate V and control policy π. The first example of simultaneous certificate/policy learning appeared in [7] for a Lyapunov certificate; later works learn barrier functions [10] and contraction metrics [13] alongside control policies as well, or extend to the case when dynamics are only partially-known [43].…”
Section: Learning Neural Certificatesmentioning
confidence: 99%
“…Other authors, notably Tsukamoto, Chung, and Slotine [34], [51] have extended the neural certificate approach to contraction metrics (see [5] for an overview of these approaches). In addition to different types of certificate, later works have expanded this framework to more challenging problems in control, such as multi-agent control in [10], black-box dynamical models in [43], the RL context in [8], and the robust case in [9]. Many of these later works are particularly notable for providing theoretical and algorithmic contributions to address difficulties that arise in practice (such as control from observations), which we discuss next in Section V.…”
Section: History Of Certificate Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…Then [22,23] propose a contraction-metric-based control framework, which extends neural networks to certificate learning for contraction metrics. Moreover, based on the framework proposed by Tsukamoto et al, [24,25,26,8] are used to address higher dimensional control problems.…”
Section: Related Workmentioning
confidence: 99%