2020 25th International Conference on Pattern Recognition (ICPR) 2021
DOI: 10.1109/icpr48806.2021.9412628
|View full text |Cite
|
Sign up to set email alerts
|

Explore and Explain: Self-supervised Navigation and Recounting

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3
1

Relationship

3
6

Authors

Journals

citations
Cited by 12 publications
(8 citation statements)
references
References 21 publications
0
7
0
Order By: Relevance
“…The resulting learned language groundings are often tightly coupled to the agent's observations and actions, making them inflexible to new objects or concepts. Researchers have explored numerous methods to mitigate this, including data augmentation (Blukis et al, 2020), dual-coding memory (Hill et al, 2021), auxiliary language generation objectives (Yan et al, 2022;Bigazzi et al, 2021), or interactive supervision (Kulick et al, 2013;Mohan & Laird, 2014;She et al, 2014;Thomason et al, 2017;Co-Reyes et al, 2018;Chai et al, 2018;Nguyen et al, 2021). Recent approaches instead leverage internet-scale language models to achieve generalization.…”
Section: Learning From Human Interactionsmentioning
confidence: 99%
“…The resulting learned language groundings are often tightly coupled to the agent's observations and actions, making them inflexible to new objects or concepts. Researchers have explored numerous methods to mitigate this, including data augmentation (Blukis et al, 2020), dual-coding memory (Hill et al, 2021), auxiliary language generation objectives (Yan et al, 2022;Bigazzi et al, 2021), or interactive supervision (Kulick et al, 2013;Mohan & Laird, 2014;She et al, 2014;Thomason et al, 2017;Co-Reyes et al, 2018;Chai et al, 2018;Nguyen et al, 2021). Recent approaches instead leverage internet-scale language models to achieve generalization.…”
Section: Learning From Human Interactionsmentioning
confidence: 99%
“…However, many state-of-the-art architectures are still considered black boxes, as their behavior lacks explainability [2]. In this respect, some attempts have been made to make the robot navigation and decision-making processes more interpretable for the end-user by letting it produce natural language descriptions of what it observes [5,12,3]. In this work, we consider a curiositydriven exploration agent [30] and equip it with the ability to produce natural language descriptions of what it observes while navigating the environment, also exploiting explainable maps to enhance the interpretability of the descriptions.…”
Section: Related Workmentioning
confidence: 99%
“…Both autonomous robotics [5,14] and embodied AI [8,11,18,6,12,21] have recently witnessed a boost of interest, which has been enabled by the release of photorealistic 3D simulated environments. In such environments, algorithms for intelligent exploration and navigation can be developed safely and more quickly than in the real-world, before being easily deployed on real robotic platforms [15,7,2].…”
Section: Related Workmentioning
confidence: 99%