2021 IEEE International Conference on Robotics and Automation (ICRA) 2021
DOI: 10.1109/icra48506.2021.9561994
|View full text |Cite
|
Sign up to set email alerts
|

A Joint Network for Grasp Detection Conditioned on Natural Language Commands

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 19 publications
(11 citation statements)
references
References 29 publications
0
11
0
Order By: Relevance
“…To address this problem, some recent researches have proposed to merge language grounding into vision-based manipulation and grasping pipelines [4]- [6], [9]- [13]. Conditioned on language, the robot can understand and execute a diverse range of VLG tasks.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…To address this problem, some recent researches have proposed to merge language grounding into vision-based manipulation and grasping pipelines [4]- [6], [9]- [13]. Conditioned on language, the robot can understand and execute a diverse range of VLG tasks.…”
Section: Related Workmentioning
confidence: 99%
“…11 templates are adapted from [28]. Similar to [13], we further augment the templates with QuillBot, an automatic paraphraser, to enrich the vocabulary and grammatical diversities. There are two types of instructions: (1) task with a target object (e.g., "Use…”
Section: Datasetmentioning
confidence: 99%
“…Natural language provides a human-interactive interface to link humans to robots, which is important for deploying robots in our lives. Many studies [12]- [16] have explored how robots follow language instructions, in which robots are required to complete tasks specified by the language. Some studies [17]- [19] have learned language-conditioned behaviors through imitation learning.…”
Section: Related Workmentioning
confidence: 99%
“…• Benchmark Performance: In simulation experiment, we evaluate model performances of the collision-free grasp using the object retrieval top-k recall (R@k) and top-k precision (P@k) metrics to evaluate multi-grasp detection (Hu et al, 2016). Chen et al (2021b) proposes above metric to evaluate language-based multi-grasping. We do not compare it with our work directly, because: (i) their work (including dataset) is not open-sourced.…”
Section: Settingsmentioning
confidence: 99%
“…It is useful for warehousing, manufacturing, medicine, retail, and service robots. One setting in robotic grasping is to grasp object orderly without disturbing the remaining in cluttered scenes (Chen et al, 2021b;Mees and Burgard, 2020;Zhang et al, 2021a) (called collision-free grasp). To solve this problem, a typical method parses the input into a scene graph first (Figure.…”
Section: Introductionmentioning
confidence: 99%