2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.01337
|View full text |Cite
|
Sign up to set email alerts
|

Coarse-to-Fine Q-attention: Efficient Learning for Visual Robotic Manipulation via Discretisation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
26
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 35 publications
(42 citation statements)
references
References 18 publications
0
26
0
Order By: Relevance
“…Akbulut et al [2] introduce a new framework called Adaptive Conditional Neural Movement Primitives, combining supervised learning and RL to conserve old skills learned from robot demonstrations while being adaptive to new environments. James and Davison [51] present a coarse-to-fine discrete RL algorithm to solve sparse reward manipulation tasks by using only a small amount of demonstration and exploration data (work extended by [49] and [50]). Celemin et al [19] include human corrective advice in the action domain through a learning-from-demonstration approach, while an RL algorithm guides the learning process by filtering out human feedback that does not maximize the reward.…”
Section: Learning From Demonstrationmentioning
confidence: 99%
See 1 more Smart Citation
“…Akbulut et al [2] introduce a new framework called Adaptive Conditional Neural Movement Primitives, combining supervised learning and RL to conserve old skills learned from robot demonstrations while being adaptive to new environments. James and Davison [51] present a coarse-to-fine discrete RL algorithm to solve sparse reward manipulation tasks by using only a small amount of demonstration and exploration data (work extended by [49] and [50]). Celemin et al [19] include human corrective advice in the action domain through a learning-from-demonstration approach, while an RL algorithm guides the learning process by filtering out human feedback that does not maximize the reward.…”
Section: Learning From Demonstrationmentioning
confidence: 99%
“…Using Multiple Guided RL Methods First, we find that the guided RL-compliant papers tend to use a variety of guided RL methods. For instance, [7], [27], and [51] utilize at least three guided RL approaches, while [43], [70], and [82] deploy five or more approaches to obtain improvements in all three guided RL dimensions.…”
Section: Key Insights On Guided Rl Compliancementioning
confidence: 99%
“…The traditional BC algorithms typically learn an explicit model that maps from the state space to the action space in the framework of supervised learning. Such an approach has been applied to Autonomous driving [9] and robotic manipulation [10], [11], [12], [13], [14]. Recently, implicit behavior cloning [3] is the first behavioral cloning algorithm to use an energybased model (EBM), showing impressive results for learning discontinuous behaviors in robotic manipulation.…”
Section: Related Workmentioning
confidence: 99%
“…Breyer et al ( 46) infer grasps at a resolution of just 40 3 . James et al circumvent this problem using hierarchy (48) in the context of Q learning. Specifically, they represent the Q function as a hierarchy of two Q models which we denote as Qcoarse and Q f ine .…”
Section: James Et Al (2022)mentioning
confidence: 99%