2013
DOI: 10.1007/s12559-013-9212-5
|View full text |Cite
|
Sign up to set email alerts
|

A Multimodal Connectionist Architecture for Unsupervised Grounding of Spatial Language

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
8
0

Year Published

2014
2014
2020
2020

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 13 publications
(9 citation statements)
references
References 36 publications
1
8
0
Order By: Relevance
“…The development of associations between co-occurring stimuli for multimodal binding has been strongly supported by neurophysiological evidence [62,112]. Similar to References [44,113,114] and based on our experimental results, we argue that the co-occurrence of sensory inputs is a sufficient source of information to create robust multimodal representations with the use of associative links between unimodal representations that can be incrementally learned in an unsupervised fashion.…”
Section: A Universal Multimodal Association Model?supporting
confidence: 88%
“…The development of associations between co-occurring stimuli for multimodal binding has been strongly supported by neurophysiological evidence [62,112]. Similar to References [44,113,114] and based on our experimental results, we argue that the co-occurrence of sensory inputs is a sufficient source of information to create robust multimodal representations with the use of associative links between unimodal representations that can be incrementally learned in an unsupervised fashion.…”
Section: A Universal Multimodal Association Model?supporting
confidence: 88%
“…This approach postulates the unsupervised role of linguistic labels that can affect categorisation during the acquisition process, which has also been supported by Taniguchi et al [42]. Vavrečka and Farkaš [46] recently introduced a multimodal architecture for grounding of spatial words using a biologically inspired approach (separate "what" and "where" visual subsystems) in which the visual scenes (two objects in 2D space in a spatial relation) are associated with their linguistic descriptions, hence leading to integration of modalities.…”
Section: Introductionmentioning
confidence: 70%
“…We have investigated some researches which apply these self-organizing incremental neural networks into multimodality integration. For example, Vavrecka and Farkas [32] presented a two-layer connectionist architecture based on SOMs to bind spatial location, shape and color of objects. Parisi et al [33]- [35] built a series of hierarchical GWR networks to fuse multimodal action representations.…”
Section: (B) the Process Mainly Involves Audiovisual Integration Andmentioning
confidence: 99%