2018
DOI: 10.48550/arxiv.1806.04225
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

PAC-Bayes Control: Learning Policies that Provably Generalize to Novel Environments

Abstract: Our goal is to synthesize controllers for robots that provably generalize well to novel environments given a dataset of example environments. The key technical idea behind our approach is to leverage tools from generalization theory in machine learning by exploiting a precise analogy (which we present in the form of a reduction) between robustness of controllers to novel environments and generalization of hypotheses in supervised learning. In particular, we utilize the Probably Approximately Correct (PAC)-Baye… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
13
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
4

Relationship

3
1

Authors

Journals

citations
Cited by 4 publications
(13 citation statements)
references
References 35 publications
0
13
0
Order By: Relevance
“…On the theoretical front, an important direction for future work is to provide rigorous guarantees on generalization to novel domains. One potential avenue is to combine the algorithmic techniques presented here with recent results on PAC-Bayes generalization theory applied to control and RL settings [13,36]. On the algorithmic front, an interesting direction is to use domain randomization techniques to automatically generate new training domains that can be used to improve invariant policy learning (e.g., automatically generating domains with different colored keys in the colored-keys example).…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…On the theoretical front, an important direction for future work is to provide rigorous guarantees on generalization to novel domains. One potential avenue is to combine the algorithmic techniques presented here with recent results on PAC-Bayes generalization theory applied to control and RL settings [13,36]. On the algorithmic front, an interesting direction is to use domain randomization techniques to automatically generate new training domains that can be used to improve invariant policy learning (e.g., automatically generating domains with different colored keys in the colored-keys example).…”
Section: Discussionmentioning
confidence: 99%
“…Distributional robustness. The PAC-Bayes Control approach [13,14] provides a way to make provable generalization guarantees under distributional shifts. This approach is particularly useful in safety-critical applications where it is important to quantify the impact of switching between a training domain and a test domain.…”
Section: Related Workmentioning
confidence: 99%
“…Majumdar et al explore one potential theory for an automated guidance system, which leverages another form of machine learning. Majumdar et al utilizes a similar framework for both the navigation of a UAS, and the task of grasping an object [10]. One issue with the grasping framework presented in the paper can be found in the mass measurement used for crucial grasping calculations [10].…”
Section: Potential Computer Visionmentioning
confidence: 99%
“…Majumdar et al utilizes a similar framework for both the navigation of a UAS, and the task of grasping an object [10]. One issue with the grasping framework presented in the paper can be found in the mass measurement used for crucial grasping calculations [10]. The mass used in their calculations is a randomly generated number in the range [0.05, 0.15] kg [10], which could cause inaccurate results in a real setting.…”
Section: Potential Computer Visionmentioning
confidence: 99%
See 1 more Smart Citation