2021 30th IEEE International Conference on Robot &Amp; Human Interactive Communication (RO-MAN) 2021
DOI: 10.1109/ro-man50785.2021.9515374
|View full text |Cite
|
Sign up to set email alerts
|

Neural Variational Learning for Grounded Language Acquisition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 42 publications
0
3
0
Order By: Relevance
“…Indeed, we are not the first to make these assumptions. Visual question answering models have already been used to explore neural networks' capacity to learn meaningful representations of referential words, such as nouns and predicates when trained on language tasks grounded in the visual world (Jiang et al, 2023;Mao, Gan, Kohli, Tenenbaum, & Wu, 2019;Pillai, Matuszek, & Ferraro, 2021;Wang, Mao, Gershman, & Wu, 2021;Zellers et al, 2021). As for function words, Hill, Hermann, Blunsom, and Clark (2018) briefly consider how visually grounded models learn negation, and Kuhnle and Copestake (2019) studied how these models interpret the quantifier most.…”
Section: The Current Studymentioning
confidence: 99%
“…Indeed, we are not the first to make these assumptions. Visual question answering models have already been used to explore neural networks' capacity to learn meaningful representations of referential words, such as nouns and predicates when trained on language tasks grounded in the visual world (Jiang et al, 2023;Mao, Gan, Kohli, Tenenbaum, & Wu, 2019;Pillai, Matuszek, & Ferraro, 2021;Wang, Mao, Gershman, & Wu, 2021;Zellers et al, 2021). As for function words, Hill, Hermann, Blunsom, and Clark (2018) briefly consider how visually grounded models learn negation, and Kuhnle and Copestake (2019) studied how these models interpret the quantifier most.…”
Section: The Current Studymentioning
confidence: 99%
“…In (Matuszek, 2018), machine-learned classifiers ground words and phrases provided by a human in an agent's perception of the world. Language can also be grounded more directly in perception, by machine learning the relevant perceptual categories from data, rather than pre-specifying them in a formal semantics (Pillai et al, 2019). In (Liu et al, 2016), an agent learns cloth folding through rich verbal communication, based on AND-OR graphs.…”
Section: Related Workmentioning
confidence: 99%
“…In (Matuszek, 2018), machine-learned classifiers ground words and phrases provided by a human in an agent's perception of the world. Language can also be grounded more directly in perception, by machine learning the relevant perceptual categories from data, rather than pre-specifying them in a formal semantics (Pillai et al, 2019). In , an agent learns cloth folding through rich verbal communication, based on AND-OR graphs.…”
Section: Related Workmentioning
confidence: 99%