2022
DOI: 10.1007/s11063-022-10796-8
|View full text |Cite
|
Sign up to set email alerts
|

Improving Target-driven Visual Navigation with Attention on 3D Spatial Relationships

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
1
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 20 publications
(3 citation statements)
references
References 30 publications
0
3
0
Order By: Relevance
“…Rather than directly using semantic maps that only contain object classes, some works [4,5,18,22,31,34] consider the relationships among objects for semantic navigation as well. Yang et al [34] propose to use a Graph Convolutional Network (GCN) [14] for incorporating prior knowledge that encodes the object relationship extracted from the Visual Genome [16] dataset into a deep reinforcement learning framework.…”
Section: Relationship Guided Navigationmentioning
confidence: 99%
See 1 more Smart Citation
“…Rather than directly using semantic maps that only contain object classes, some works [4,5,18,22,31,34] consider the relationships among objects for semantic navigation as well. Yang et al [34] propose to use a Graph Convolutional Network (GCN) [14] for incorporating prior knowledge that encodes the object relationship extracted from the Visual Genome [16] dataset into a deep reinforcement learning framework.…”
Section: Relationship Guided Navigationmentioning
confidence: 99%
“…The robot uses the features from the knowledge graph for predicting corresponding actions. Some other works [5,18,22] also use GCNs to encode the relationship information, and the difference between them is the definition of the nodes. Different from the above works, which construct a graph to encode the relationship information, Druon et al [4] use a context grid to represent this kind of semantic information.…”
Section: Relationship Guided Navigationmentioning
confidence: 99%
“…The visual-based DRL has a broad range of applications in robot manipulation tasks as it has little requirement of accurate environment state, such as position and distance, and has achieved good performance in many tasks such as grasping [20,21], pushing [22,23], and navigation [24]. However, it could be really difficult to train an end-to-end visual-based DRL due to the sample complexity and inefficient reward shaping [16].…”
Section: Related Workmentioning
confidence: 99%