2020
DOI: 10.1609/aaai.v34i06.6556
|View full text |Cite
|
Sign up to set email alerts
|

NeoNav: Improving the Generalization of Visual Navigation via Generating Next Expected Observations

Abstract: We propose improving the cross-target and cross-scene generalization of visual navigation through learning an agent that is guided by conceiving the next observations it expects to see. This is achieved by learning a variational Bayesian model, called NeoNav, which generates the next expected observations (NEO) conditioned on the current observations of the agent and the target view. Our generative model is learned through optimizing a variational objective encompassing two key designs. First, the latent distr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5

Relationship

1
4

Authors

Journals

citations
Cited by 10 publications
(7 citation statements)
references
References 12 publications
0
7
0
Order By: Relevance
“…In addition, we investigate three techniques to improve robot navigation performance in the real world, including feature space dynamics, premature collision prediction and additional target checking. We show that the proposed method significantly outperforms prior work [16], boosting the success rate from 17.5% to 28.7% and reducing approximately 16.3% of the collisions for a navigation task in the unseen scenes from the Active Vision Dataset [17]. Furthermore, we steer a wheeled robot, TurtleBot, around office scenes and show that the learned navigation policy can generalize to novel targets in unseen real-world environments.…”
Section: Introductionmentioning
confidence: 84%
See 4 more Smart Citations
“…In addition, we investigate three techniques to improve robot navigation performance in the real world, including feature space dynamics, premature collision prediction and additional target checking. We show that the proposed method significantly outperforms prior work [16], boosting the success rate from 17.5% to 28.7% and reducing approximately 16.3% of the collisions for a navigation task in the unseen scenes from the Active Vision Dataset [17]. Furthermore, we steer a wheeled robot, TurtleBot, around office scenes and show that the learned navigation policy can generalize to novel targets in unseen real-world environments.…”
Section: Introductionmentioning
confidence: 84%
“…The mapping is a surjection, which means there is only one appropriate action a t for (x t , x t+1 ). In this way, the multimodality essentially affects the generation of x t+1 , which is learned by a generative module as [16]. Generative Module.…”
Section: A Navigation Policymentioning
confidence: 99%
See 3 more Smart Citations