Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1438
|View full text |Cite
|
Sign up to set email alerts
|

Compositional Generalization for Primitive Substitutions

Abstract: Compositional generalization is a basic mechanism in human language learning, but current neural networks lack such ability. In this paper, we conduct fundamental research for encoding compositionality in neural networks. Conventional methods use a single representation for the input sentence, making it hard to apply prior knowledge of compositionality. In contrast, our approach leverages such knowledge with two representations, one generating attention maps, and the other mapping attended input words to outpu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
76
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 49 publications
(83 citation statements)
references
References 28 publications
1
76
0
Order By: Relevance
“…The table shows results (mean test accuracy (%) ± standard deviation) on the test splits of the dataset. Syntactic Attention is compared to the previous models, which were a CNN (Dessì and Baroni, 2019), GRUs augmented with an attention mechanism ("+ attn"), which either included or did not include a dependency ("-dep") in the decoder on the previous action (Bastings et al, 2018), and the recent model of Li et al (2019). Lake ( 2019) showed that a meta-learning architecture using an external memory achieves 99.95% accuracy on a meta-seq2seq version of the SCAN task.…”
Section: Compositional Generalization Resultsmentioning
confidence: 99%
“…The table shows results (mean test accuracy (%) ± standard deviation) on the test splits of the dataset. Syntactic Attention is compared to the previous models, which were a CNN (Dessì and Baroni, 2019), GRUs augmented with an attention mechanism ("+ attn"), which either included or did not include a dependency ("-dep") in the decoder on the previous action (Bastings et al, 2018), and the recent model of Li et al (2019). Lake ( 2019) showed that a meta-learning architecture using an external memory achieves 99.95% accuracy on a meta-seq2seq version of the SCAN task.…”
Section: Compositional Generalization Resultsmentioning
confidence: 99%
“…One illustrative example is the poor performance of LSTMs on a SCAN split that requires generalizing from shorter to longer sequences. While several models have made significant improvements over other SCAN splits, progress on the length split remains minimal (Li et al, 2019;Gordon et al, 2020).…”
Section: Comparison To Related Workmentioning
confidence: 99%
“…Crucially, their training/evaluation split required compositional generalization. A number of models have been developed that have improved performance on SCAN (Li et al, 2019;Gordon et al, 2020). However, since the semantic representation used by SCAN only covers a small subset of English grammar, SCAN does not enable testing various systematic linguistic abstractions that humans are known to make (e.g., verb argument structure alternation).…”
Section: Introductionmentioning
confidence: 99%
“…Many past works in the rich body of literature about analyzing NNs focus on compositional structure (Hupkes et al, 2020(Hupkes et al, , 2018Hewitt and Manning, 2019;Li et al, 2019) and systematicity (Lake and Baroni, 2018;Goodwin et al, 2020). Two of the most popular analysis techniques are the behavioral and probing approaches.…”
Section: Analysis Of Nnsmentioning
confidence: 99%