2022 International Conference on Robotics and Automation (ICRA) 2022
DOI: 10.1109/icra46639.2022.9811963
|View full text |Cite
|
Sign up to set email alerts
|

Offline Learning of Counterfactual Predictions for Real-World Robotic Reinforcement Learning

Abstract: Autonomous excavation is a challenging task. The unknown contact dynamics between the excavator bucket and the terrain could easily result in large contact forces and jamming problems during excavation. Traditional model-based methods struggle to handle such problems due to complex dynamic modeling. In this paper, we formulate the excavation skills with three novel manipulation primitives. We propose to learn the manipulation primitives with offline reinforcement learning (RL) to avoid large amounts of online … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 44 publications
0
1
0
Order By: Relevance
“…The advantage of this approach is that the control-rules are easy to explain to human operators, but since control is triggered by predictions that are continually updated in deployment the resultant controller adapts to changing conditions. An extension of this idea is to use GVF predictions-like the ones we learned in this work-as input to a neural-network based RL agent, similarly to how it was done for autonomous driving (Graves et al, 2020;Jin et al, 2022). This work provides the foundations for these next steps in industrial control with RL.…”
Section: Discussionmentioning
confidence: 99%
“…The advantage of this approach is that the control-rules are easy to explain to human operators, but since control is triggered by predictions that are continually updated in deployment the resultant controller adapts to changing conditions. An extension of this idea is to use GVF predictions-like the ones we learned in this work-as input to a neural-network based RL agent, similarly to how it was done for autonomous driving (Graves et al, 2020;Jin et al, 2022). This work provides the foundations for these next steps in industrial control with RL.…”
Section: Discussionmentioning
confidence: 99%
“…Counterfactuals can be used to analyse the current situation and envisage the different outcomes by changing the antecedents to add to the knowledge. Counterfactual predictions are used by Jin et al (2022) to guide the exploration of a reinforcement learner for robotic manipulation tasks. The active counterfactual predictions generated add to the existing body of knowledge about the task for the learning model to improve its performance.…”
Section: Proposed Counterfactual Learning-based Approach For Resilien...mentioning
confidence: 99%
“…At this juncture, it is exciting to look for the integration of artificial intelligence in process control as it can bring in a paradigm shift by enabling more intelligent, adaptive, and efficient control of industrial processes, ultimately contributing to increased productivity and sustainability . As such, a great deal of research is being done on model-free controlling techniques, particularly on model-free reinforcement learning (RL) algorithms. In essence, a trained RL has the capabilities of generating a control policy by optimizing the cumulative reward signal over successive interactions with an environment until a desired goal is achieved. …”
Section: Introductionmentioning
confidence: 99%