2016
DOI: 10.1007/978-3-319-32192-9_8
|View full text |Cite
|
Sign up to set email alerts
|

Audio Visual Attention Models in the Mobile Robots Navigation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2017
2017
2021
2021

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 32 publications
0
2
0
Order By: Relevance
“…Audio-visual correspondence also allows audio signals to serve as supervision for visual learning [24]. In the context of robotics and situated interactions, joint reasoning using audio-visual inputs supports target localization, tracking, and hence navigation (e.g., [25]- [28]). For example, a deep neural network that utilizes both vision and auditory inputs outperforms conventional methods that rely on single modalities in object tracking [25].…”
Section: Audio-visual Joint Reasoningmentioning
confidence: 99%
“…Audio-visual correspondence also allows audio signals to serve as supervision for visual learning [24]. In the context of robotics and situated interactions, joint reasoning using audio-visual inputs supports target localization, tracking, and hence navigation (e.g., [25]- [28]). For example, a deep neural network that utilizes both vision and auditory inputs outperforms conventional methods that rely on single modalities in object tracking [25].…”
Section: Audio-visual Joint Reasoningmentioning
confidence: 99%
“…Joint reasoning using audio-visual inputs, in general, supports target localization, tracking, and hence navigation (e.g., [18]- [21]). For example, a deep neural network that utilizes both vision and auditory inputs outperforms conventional methods that rely on single modalities in object tracking [18].…”
Section: Audio-visual Joint Reasoningmentioning
confidence: 99%