2018
DOI: 10.48550/arxiv.1807.06757
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On Evaluation of Embodied Navigation Agents

Peter Anderson,
Angel Chang,
Devendra Singh Chaplot
et al.
Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
382
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 188 publications
(383 citation statements)
references
References 0 publications
1
382
0
Order By: Relevance
“…While deep learning has demonstrated great success in various application domains [Russakovsky et al, 2015, Silver et al, 2016, large-scale annotated data for supervision inevitably becomes the bottleneck. Many works thus explore self-supervised learning via active perception [Wilkes & Tsotsos, 1992], interactive perception [Bohg et al, 2017], or interactive exploration [Wyatt et al, 2011] to learn visual representations [Fang et al, 2020, Jayaraman & Grauman, 2018, Weihs et al, 2019, Zakka et al, 2020, objects and poses [Caicedo & Lazebnik, 2015, Chaplot et al, 2020b, Choi et al, 2021, segmentation and parts [Eitel et al, 2019, Gadre et al, 2021, Katz & Brock, 2008, Kenney et al, 2009, Lohmann et al, 2020, Pathak et al, 2018, Van Hoof et al, 2014, physics and dynamics [Agrawal et al, 2016, Ehsani et al, 2020, Janner et al, 2018, Li et al, 2016, Lohmann et al, 2020, Mottaghi et al, 2016, Wu et al, 2015, manipulation skills [Agrawal et al, 2016, Batra et al, 2020, Zeng et al, 2018, navigation policies [Anderson et al, 2018, Chaplot et al, 2020a, Ramakrishnan et al, 2021, etc. In this work, we design interactive policies to explore novel 3D indoor rooms and learn our newly proposed inter-object functional relationships.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…While deep learning has demonstrated great success in various application domains [Russakovsky et al, 2015, Silver et al, 2016, large-scale annotated data for supervision inevitably becomes the bottleneck. Many works thus explore self-supervised learning via active perception [Wilkes & Tsotsos, 1992], interactive perception [Bohg et al, 2017], or interactive exploration [Wyatt et al, 2011] to learn visual representations [Fang et al, 2020, Jayaraman & Grauman, 2018, Weihs et al, 2019, Zakka et al, 2020, objects and poses [Caicedo & Lazebnik, 2015, Chaplot et al, 2020b, Choi et al, 2021, segmentation and parts [Eitel et al, 2019, Gadre et al, 2021, Katz & Brock, 2008, Kenney et al, 2009, Lohmann et al, 2020, Pathak et al, 2018, Van Hoof et al, 2014, physics and dynamics [Agrawal et al, 2016, Ehsani et al, 2020, Janner et al, 2018, Li et al, 2016, Lohmann et al, 2020, Mottaghi et al, 2016, Wu et al, 2015, manipulation skills [Agrawal et al, 2016, Batra et al, 2020, Zeng et al, 2018, navigation policies [Anderson et al, 2018, Chaplot et al, 2020a, Ramakrishnan et al, 2021, etc. In this work, we design interactive policies to explore novel 3D indoor rooms and learn our newly proposed inter-object functional relationships.…”
Section: Related Workmentioning
confidence: 99%
“…An agent is provided with large-scale scenes to explore for learning in the training stage and is asked to predict the functional scene graph (S, R S ) for a novel scene at the test time. We also abstract away complexities on robotic navigation [Anderson et al, 2018, Ramakrishnan et al, 2021 and manipulation [Mo et al, 2021a, which are orthogonal to our contribution to estimating inter-object functional relationships.…”
Section: Problem Formulationmentioning
confidence: 99%
“…Meanwhile, to avoid collisions, the works [10,18] imply an intrinsic collision reward with an additional collision detector module. Some works [1,9,48] introduce more information from the environment to improve navigation performance. For example, Chen et al [9] utilize additional topological guidance of scenes for visual navigation.…”
Section: Visual Navigation With Rlmentioning
confidence: 99%
“…For example, Chen et al [9] utilize additional topological guidance of scenes for visual navigation. In [1,13,26,36], natural language instructions are introduced to guide the agent. Tang et al [52] introduce an auto-navigator to design a specialized network for visual navigation.…”
Section: Visual Navigation With Rlmentioning
confidence: 99%
See 1 more Smart Citation