2021
DOI: 10.1109/lra.2020.3044033
|View full text |Cite
|
Sign up to set email alerts
|

Chance-Constrained Trajectory Optimization for Safe Exploration and Learning of Nonlinear Systems

Abstract: Learning-based control algorithms require data collection with abundant supervision for training. Safe exploration algorithms ensure the safety of this data collection process even when only partial knowledge is available. We present a new approach for optimal motion planning with safe exploration that integrates chance-constrained stochastic optimal control with dynamics learning and feedback control. We derive an iterative convex optimization algorithm that solves an Information-cost Stochastic Nonlinear Opt… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
19
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
1

Relationship

2
4

Authors

Journals

citations
Cited by 28 publications
(19 citation statements)
references
References 21 publications
0
19
0
Order By: Relevance
“…Policy to be learned State tracking error x − x d Computational load (a) Learning-based motion planner [5]- [11] (x, o ,t) → u d Increases exponentially (Lemma 1) One neural net evaluation (b) Robust tube-based motion planner [13]- [26] (…”
Section: Table I Comparison Of the Proposed Methods With The Learning-based And Robust Tube-based Motion Planners Motion Planning Schemementioning
confidence: 99%
See 4 more Smart Citations
“…Policy to be learned State tracking error x − x d Computational load (a) Learning-based motion planner [5]- [11] (x, o ,t) → u d Increases exponentially (Lemma 1) One neural net evaluation (b) Robust tube-based motion planner [13]- [26] (…”
Section: Table I Comparison Of the Proposed Methods With The Learning-based And Robust Tube-based Motion Planners Motion Planning Schemementioning
confidence: 99%
“…One neural net evaluation demonstrations for training, so that the learned policy will not violate given state constraints even with the learning error and external disturbances. In this phase, LAG-ROS learns the contracting control law independently of a target trajectory, making it implementable without solving any motion planning problems online unlike [13]- [26]. The performance of LAG-ROS is evaluated in cart-pole balancing [28] and nonlinear motion planning of multiple robotic agents [29] in a cluttered environment, demonstrating that LAG-ROS indeed satisfies the formal exponential bound as in [13]- [21] with its computational load as low as that of existing learning-based motion planners [5]- [11].…”
Section: Table I Comparison Of the Proposed Methods With The Learning-based And Robust Tube-based Motion Planners Motion Planning Schemementioning
confidence: 99%
See 3 more Smart Citations