2022 International Conference on Robotics and Automation (ICRA) 2022
DOI: 10.1109/icra46639.2022.9812027
|View full text |Cite
|
Sign up to set email alerts
|

Object Memory Transformer for Object Goal Navigation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 18 publications
(6 citation statements)
references
References 10 publications
0
6
0
Order By: Relevance
“…VGM (Kwon et al, 2021) is constructed incrementally based on the similarities among the unsupervised representations of observed images, and these representations are learned from an unlabeled image dataset. OMT (Fukushima et al, 2022) uses transformer to salient objects stored in memory. DUET (Chen et al, 2022) proposes a joint long-term action planning to enable efficient exploration in global action space.…”
Section: A2 Memory Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…VGM (Kwon et al, 2021) is constructed incrementally based on the similarities among the unsupervised representations of observed images, and these representations are learned from an unlabeled image dataset. OMT (Fukushima et al, 2022) uses transformer to salient objects stored in memory. DUET (Chen et al, 2022) proposes a joint long-term action planning to enable efficient exploration in global action space.…”
Section: A2 Memory Methodsmentioning
confidence: 99%
“…According to the definition of meta-ability and thinking, we summarize the current mainstream object navigation methods and identify their limitations. As shown in Figure 2, object navigation methods are divided into four categories: association methods (Dang et al, 2022a;Zhang et al, 2021), memory methods (Chen et al, 2022;Fukushima et al, 2022), deadlock-specialized methods (Du et al, 2020;Lin et al, 2021) and SLAM methods (Ravichandran et al, 2022;Liang et al, 2021). The different inductive biases introduced by these four types of methods determine which meta-abilities are emphasized and which are overlooked.…”
Section: Introductionmentioning
confidence: 99%
“…In Object memory transformer [37], image and object representation are saved in the explicit memory every time step. Since TSGM only puts a new node into a graph memory based on the similarity between memory and current observations for both image and object graphs, it has less redundancy than [37]. [37] utilizes only the preceding T chunks of data are utilized.…”
Section: Appendices Appendix a Training Details And Experimental Sett...mentioning
confidence: 99%
“…Since TSGM only puts a new node into a graph memory based on the similarity between memory and current observations for both image and object graphs, it has less redundancy than [37]. [37] utilizes only the preceding T chunks of data are utilized. TSGM, on the other hand, makes use of all graph memory information derived from past exploration of the environment.…”
Section: Appendices Appendix a Training Details And Experimental Sett...mentioning
confidence: 99%
See 1 more Smart Citation