2019
DOI: 10.48550/arxiv.1906.10197
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Mutual exclusivity as a challenge for deep neural networks

Abstract: Strong inductive biases allow children to learn in fast and adaptable ways. Children use the mutual exclusivity (ME) bias to help disambiguate how words map to referents, assuming that if an object has one label then it does not need another. In this paper, we investigate whether or not standard neural architectures have a ME bias, demonstrating that they lack this learning assumption. Moreover, we show that their inductive biases are poorly matched to early-phase learning in several standard tasks: machine tr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 16 publications
0
4
0
Order By: Relevance
“…Human infants are not as tabula rasa as models like InferSent but rather encode useful inductive biases (Chomsky & Lightfoot, 2002;Lightfoot & Julia, 1984;Mitchell, 1980;Pearl & Goldwater, 2016;Seidenberg, 1997). Building such biases into our models (Battaglia et al, 2018;Dubey, Agrawal, Pathak, Griffiths, & Efros, 2018;Gandhi & Lake, 2019;Lake et al, 2018) is a promising direction towards scalably learning systematic representations. We also showed how analysis and controlled testing for heuristic strategies in the learning environment can provide rich insights into the representations learned.…”
Section: Discussion and Future Workmentioning
confidence: 99%
“…Human infants are not as tabula rasa as models like InferSent but rather encode useful inductive biases (Chomsky & Lightfoot, 2002;Lightfoot & Julia, 1984;Mitchell, 1980;Pearl & Goldwater, 2016;Seidenberg, 1997). Building such biases into our models (Battaglia et al, 2018;Dubey, Agrawal, Pathak, Griffiths, & Efros, 2018;Gandhi & Lake, 2019;Lake et al, 2018) is a promising direction towards scalably learning systematic representations. We also showed how analysis and controlled testing for heuristic strategies in the learning environment can provide rich insights into the representations learned.…”
Section: Discussion and Future Workmentioning
confidence: 99%
“…When children endeavour to learn a new word, they rely on inductive biases to narrow the space of possible meanings: they prefer to predict that the novel word refers to the novel object. However, deep learning algorithms lack this bias [15]. To demonstrate this assumption, we calculate the percentage of novel object in the wrongly labeled vision regions.…”
Section: Zero-shot Learning With Vsepsmentioning
confidence: 99%
“…Failure to generalize structurally or failure to produce novel labels? It is known that neural models find it challenging to produce labels they have not seen during training (Gandhi and Lake, 2019). Handling this problem is a necessary part of solving depth generalization, since each of the outputs of the depth generalization cases, such as (5b) below, contains more constants than the training outputs, such as the output of (5a):…”
Section: Lexical Vs Structural Generalizationmentioning
confidence: 99%