2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2019
DOI: 10.1109/iros40897.2019.8968596
|View full text |Cite
|
Sign up to set email alerts
|

Affordance Learning for End-to-End Visuomotor Robot Control

Abstract: Training end-to-end deep robot policies requires a lot of domain-, task-, and hardware-specific data, which is often costly to provide. In this work, we propose to tackle this issue by employing a deep neural network with a modular architecture, consisting of separate perception, policy, and trajectory parts. Each part of the system is trained fully on synthetic data or in simulation. The data is exchanged between parts of the system as low-dimensional latent representations of affordances and trajectories. Th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
36
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 36 publications
(36 citation statements)
references
References 26 publications
(44 reference statements)
0
36
0
Order By: Relevance
“…James et al trained a visuomotor skill for pick-and-place task in simulation with domain randomization and transferred the visuomotor skill to the real world without any fine-tuning [3]. Hämäläinen et al applied an affordance detection method to compress task-related features and trained a visuomotor policy with these features to get a good generalization capability to new tasks and new objects [23]. Chen et al used an adversarial feature to enhance the robustness of the end-to-end visuomotor skills, which is integrated with RL framework [24].…”
Section: Related Workmentioning
confidence: 99%
“…James et al trained a visuomotor skill for pick-and-place task in simulation with domain randomization and transferred the visuomotor skill to the real world without any fine-tuning [3]. Hämäläinen et al applied an affordance detection method to compress task-related features and trained a visuomotor policy with these features to get a good generalization capability to new tasks and new objects [23]. Chen et al used an adversarial feature to enhance the robustness of the end-to-end visuomotor skills, which is integrated with RL framework [24].…”
Section: Related Workmentioning
confidence: 99%
“…They could not be applied on robot manipulation task directly. A trajectory latent space was also used for motion generation of robot manipulation, which embedded the entire task-specific trajectory into a low dimensional latent space [40], [41]. It used one RGB image to generate the entire manipulation trajectory just like our framework.…”
Section: Related Workmentioning
confidence: 99%
“…Compression processing is also widely used for redundant high-dimensional data such as images or video frames. Variational autoencoder [47] is often used to compress the perceptional information in field of robot vision [48][49][50]. It is verified that VAE has better generalization ability compared with traditional autoencoder (AE) in visuomotor framework [51].…”
Section: Perception Networkmentioning
confidence: 99%
“…Hoof et al used VAE to process the high-dimensional tactile and visual feedback for stabilization task [48]. Hämäläinen et al applied VAE to cope with the visual observation for affordance learning in a visuomotor control framework [49]. Piergiovanni compressed the image via a VAE for robot navigation [50].…”
Section: Perception Networkmentioning
confidence: 99%