2023
DOI: 10.48550/arxiv.2303.03365
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Efficient Skill Acquisition for Complex Manipulation Tasks in Obstructed Environments

Abstract: Data efficiency in robotic skill acquisition is crucial for operating robots in varied small-batch assembly settings. To operate in such environments, robots must have robust obstacle avoidance and versatile goal conditioning acquired from only a few simple demonstrations. Existing approaches, however, fall short of these requirements. Deep reinforcement learning (RL) enables a robot to learn complex manipulation tasks but is often limited to small task spaces in the real world due to sample inefficiency and s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(5 citation statements)
references
References 34 publications
0
5
0
Order By: Relevance
“…Meanwhile, Lee et al (2020) quantify uncertainty in pose estimation to determine a binary switching strategy using model-based or RL policies. Additionally, Yamada et al (2023) implemented an object-centric generative model to identify goals for motion planning and a skill transition network to facilitate the movement of the endeffector from its terminal state in motion planning to viable starting states of a sample-efficient RL policy. However, these methods require the model of the object, in particular, the manual specification of a goal state in the robot's frame and control policy design (Yamada et al, 2023).…”
Section: Figurementioning
confidence: 99%
See 2 more Smart Citations
“…Meanwhile, Lee et al (2020) quantify uncertainty in pose estimation to determine a binary switching strategy using model-based or RL policies. Additionally, Yamada et al (2023) implemented an object-centric generative model to identify goals for motion planning and a skill transition network to facilitate the movement of the endeffector from its terminal state in motion planning to viable starting states of a sample-efficient RL policy. However, these methods require the model of the object, in particular, the manual specification of a goal state in the robot's frame and control policy design (Yamada et al, 2023).…”
Section: Figurementioning
confidence: 99%
“…Additionally, Yamada et al (2023) implemented an object-centric generative model to identify goals for motion planning and a skill transition network to facilitate the movement of the endeffector from its terminal state in motion planning to viable starting states of a sample-efficient RL policy. However, these methods require the model of the object, in particular, the manual specification of a goal state in the robot's frame and control policy design (Yamada et al, 2023). Additionally, they face difficulties in providing comprehensive guidance in both free space and contactrich regions due to the limited motion planning in tasks that require environmental interaction and the scarcity of visual servoing in addressing geometric constraints.…”
Section: Figurementioning
confidence: 99%
See 1 more Smart Citation
“…This enables the models to better understand the underlying structure of a scene and capture the relationships between its constituent objects. Early works have conducted unsupervised scene inference and generation in 2D (MONet [22], Slot Attention [9], GENESIS [23], GENESIS-V2 [10]), and for robotics applications using APEX [24], [25]. In both [24], [25] the 2D OCGM, APEX, is utilised for object matching using the learned object-centric latent representation in an object rearrangement task in simulation and a peg-in-hole task in the real world respectively.…”
Section: Related Workmentioning
confidence: 99%
“…Early works have conducted unsupervised scene inference and generation in 2D (MONet [22], Slot Attention [9], GENESIS [23], GENESIS-V2 [10]), and for robotics applications using APEX [24], [25]. In both [24], [25] the 2D OCGM, APEX, is utilised for object matching using the learned object-centric latent representation in an object rearrangement task in simulation and a peg-in-hole task in the real world respectively. Nevertheless, the 2D reconstruction and 2D bounding boxes predicted by such OCGM are of limited use in a 3D world.…”
Section: Related Workmentioning
confidence: 99%