2016 IEEE International Conference on Robotics and Automation (ICRA) 2016
DOI: 10.1109/icra.2016.7487144
|View full text |Cite
|
Sign up to set email alerts
|

Automatic LQR tuning based on Gaussian process global optimization

Abstract: Abstract-This paper proposes an automatic controller tuning framework based on linear optimal control combined with Bayesian optimization. With this framework, an initial set of controller gains is automatically improved according to a pre-defined performance objective evaluated from experimental data. The underlying Bayesian optimization algorithm is Entropy Search, which represents the latent objective as a Gaussian process and constructs an explicit belief over the location of the objective minimum. This is… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
126
0
1

Year Published

2016
2016
2023
2023

Publication Types

Select...
4
2
1

Relationship

2
5

Authors

Journals

citations
Cited by 115 publications
(128 citation statements)
references
References 22 publications
0
126
0
1
Order By: Relevance
“…Deisenroth et al [9], for example, developed a Bayesian approach to tune a cart-pole system, whereas Berkenkamp et al [6] proposed to use Bayesian optimization to safely tune robotic controllers for quadrotors. Moreover, Marco et al [16] combined Bayesian optimization with optimal control to tune LQR regulators. In contrast to these approaches, Akrour et al [1] suggested to direct the optimization process by using a search distribution, however, the optimization loses expressibility on a global scope since it only optimizes locally.…”
Section: Related Workmentioning
confidence: 99%
“…Deisenroth et al [9], for example, developed a Bayesian approach to tune a cart-pole system, whereas Berkenkamp et al [6] proposed to use Bayesian optimization to safely tune robotic controllers for quadrotors. Moreover, Marco et al [16] combined Bayesian optimization with optimal control to tune LQR regulators. In contrast to these approaches, Akrour et al [1] suggested to direct the optimization process by using a search distribution, however, the optimization loses expressibility on a global scope since it only optimizes locally.…”
Section: Related Workmentioning
confidence: 99%
“…BO with GPs has successfully been used, for example, for gait learning with bipedal walkers [9], quadrupedal robots [10], and in cm-scale hexapodal robots [11]. It has also been proposed for automatic feedback controller tuning [12]- [14]. A major strength of GPs is that they allow one to include existing information about the system in the form of a probabilistic prior.…”
Section: Gait Learning Approachmentioning
confidence: 99%
“…We formulate the problem of gait learning in soft microrobots as a parametric controller tuning problem. This work builds on the automatic controller tuning methods proposed in [12], where an unknown controller cost function was learned in a data-efficient way using BO with GPs.…”
Section: A Controller Learningmentioning
confidence: 99%
“…BO for controller learning has recently also been suggested in [12], [20], [21], which include successful demonstrations in laboratory experiments. A discrete event controller is optimized for a walking robot in [12], and state-feedback controllers are tuned in [20] for a quadrotor and in [21] for a humanoid robot balancing a pole. Herein, we present results of applying BO for a typical control problem in the automotive industry (throttle valve control) and consider two types of control objectives, different from those in [12], [20], [21].…”
Section: Introductionmentioning
confidence: 99%
“…A discrete event controller is optimized for a walking robot in [12], and state-feedback controllers are tuned in [20] for a quadrotor and in [21] for a humanoid robot balancing a pole. Herein, we present results of applying BO for a typical control problem in the automotive industry (throttle valve control) and consider two types of control objectives, different from those in [12], [20], [21]. The proposed controller learning framework, which combines BO with ADRC, is different from the controllers in the mentioned references.…”
Section: Introductionmentioning
confidence: 99%