2021 European Control Conference (ECC) 2021
DOI: 10.23919/ecc54610.2021.9655042
|View full text |Cite
|
Sign up to set email alerts
|

Weights-varying MPC for Autonomous Vehicle Guidance: a Deep Reinforcement Learning Approach

Abstract: Determining the optimal cost function parameters of Model Predictive Control (MPC) to optimize multiple control objectives is a challenging and time-consuming task. Multiobjective Bayesian Optimization (BO) techniques solve this problem by determining a Pareto optimal parameter set for an MPC with static weights. However, a single parameter set may not deliver the most optimal closed-loop control performance when the context of the MPC operating conditions changes during its operation, urging the need to adapt… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 14 publications
(5 citation statements)
references
References 50 publications
0
5
0
Order By: Relevance
“…Comparatively, Reinforcement Learning (RL) can modify MPC parameters at each time step to quickly adapt to dynamic environments. Zarrouki et al (2021) and Song and Scaramuzza (2020) both learn a policy that can improve parameters of MPC's cost function, while Gros and Zanon (2020); Amos et al (2018) aim to modify both transition and cost function's parameters. Nevertheless, none of these methods consider the non-stationary scenario in an autonomous driving system, which violates the essential Markov property in RL.…”
Section: Related Workmentioning
confidence: 99%
“…Comparatively, Reinforcement Learning (RL) can modify MPC parameters at each time step to quickly adapt to dynamic environments. Zarrouki et al (2021) and Song and Scaramuzza (2020) both learn a policy that can improve parameters of MPC's cost function, while Gros and Zanon (2020); Amos et al (2018) aim to modify both transition and cost function's parameters. Nevertheless, none of these methods consider the non-stationary scenario in an autonomous driving system, which violates the essential Markov property in RL.…”
Section: Related Workmentioning
confidence: 99%
“…In [16], a neural network model predictive control(NNMPC) is d proposed, which used neural network (NN) to learn and predict vehicle dynamics based on measured states and input variables. The concept of using Reinforcement Learning (RL) to learn MPC cost function parameters is introduced in [17], and [18] proposes a weights-varying MPC based on deep reinforcement learning to adjust cost function weights in different situations. Prediction horizon refers to the time range used to predict the system's future behavior during the control process, and is the key parameter affecting performance and computational burden of the control system in MPC [19].…”
Section: Introductionmentioning
confidence: 99%
“…Ref. [20] proposes a weights-varying MPC using a deep reinforcement learning (DRL) algorithm to adjust cost function weights in different situations. Ref.…”
Section: Introductionmentioning
confidence: 99%
“…Ref. [21] proposed a novel approach limiting DRL actions within a safe learning space, and the proposed DRL algorithm can automatically learn context-dependent optimal parameter sets and dynamically adapt for a weights-varying MPC. Ref.…”
Section: Introductionmentioning
confidence: 99%