2018
DOI: 10.1109/lra.2018.2800106
|View full text |Cite
|
Sign up to set email alerts
|

Model-Plant Mismatch Compensation Using Reinforcement Learning

Abstract: Abstract-Learning-based approaches are suitable for the control of systems with unknown dynamics. However, learning from scratch involves many trials with exploratory actions until a good control policy is discovered. Real robots usually cannot withstand the exploratory actions and suffer damage. This problem can be circumvented by combining learning with model-based control. In this article, we employ a nominal model-predictive controller that is impeded by the presence of an unknown modelplant mismatch. To c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
22
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
6
3
1

Relationship

0
10

Authors

Journals

citations
Cited by 34 publications
(24 citation statements)
references
References 11 publications
(14 reference statements)
0
22
0
Order By: Relevance
“…Our framework also provides several new greedy algorithms, including parallel and constrained formulations. We further extend this method to include the effect of control inputs, making it applicable to robotic systems [ 55 ], and we use this procedure to discover Hamiltonians. Finally, our approach provides guidance on model selection, a comprehensive comparison with previous methods, and a careful analysis of noise robustness.…”
Section: Introductionmentioning
confidence: 99%
“…Our framework also provides several new greedy algorithms, including parallel and constrained formulations. We further extend this method to include the effect of control inputs, making it applicable to robotic systems [ 55 ], and we use this procedure to discover Hamiltonians. Finally, our approach provides guidance on model selection, a comprehensive comparison with previous methods, and a careful analysis of noise robustness.…”
Section: Introductionmentioning
confidence: 99%
“…Our work toward compensating for modeling error with datadriven learning is similar to Sun et al (2019) where authors use deep learning to predict physics-based modeling error of water resources, Kaheman et al (2019) where they present an algorithm to learn a discrepancy model on an double inverted pendulum, and Della Santina et al (2020) where the authors augment a model-based disturbance observer with a learned correction factor on a soft robot. Most similar to our work is that of Koryakovskiy et al (2018) where they augment a non-linear model predictive controller with various forms of learned actions to compensate for model-plant mismatch on a rigid humanoid robot. Other works that include using neural networks as the backbone for predictive control are Piche et al (2000) and Lu and Tsai (2008).…”
Section: Related Workmentioning
confidence: 96%
“…Koryakovskiy et al [23] argued that learning from scratch approaches are not applicable on real robots due to involving many trials with exploratory actions. To circumvent this problem, they combined RL with a model-based control using two different approaches to compensate model-plant mismatches.…”
Section: B Combination Of Model-based and MLmentioning
confidence: 99%