Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.657
|View full text |Cite
|
Sign up to set email alerts
|

Discriminatively-Tuned Generative Classifiers for Robust Natural Language Inference

Abstract: While discriminative neural network classifiers are generally preferred, recent work has shown advantages of generative classifiers in term of data efficiency and robustness. In this paper, we focus on natural language inference (NLI). We propose GenNLI, a generative classifier for NLI tasks, and empirically characterize its performance by comparing it to five baselines, including discriminative models and large-scale pretrained language representation models like BERT. We explore training objectives for discr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2

Relationship

2
5

Authors

Journals

citations
Cited by 8 publications
(4 citation statements)
references
References 36 publications
0
4
0
Order By: Relevance
“…We are motivated by the insight that generative classifier is more effective in low data regime than discriminative classifier which is demonstrated by [17]. Although the conclusion is drawn on simple linear models [17], similar results are also observed on deep neural networks (DNNs) [18], [19] recently. It should be noticed that in online CIL setting the data is seen only once, not fully trained, so it is analogous to the low data regime in which the generative classifier is preferable.…”
Section: Introductionmentioning
confidence: 86%
See 1 more Smart Citation
“…We are motivated by the insight that generative classifier is more effective in low data regime than discriminative classifier which is demonstrated by [17]. Although the conclusion is drawn on simple linear models [17], similar results are also observed on deep neural networks (DNNs) [18], [19] recently. It should be noticed that in online CIL setting the data is seen only once, not fully trained, so it is analogous to the low data regime in which the generative classifier is preferable.…”
Section: Introductionmentioning
confidence: 86%
“…As discussed above, online CIL is in a low data setting where generative classifiers are preferable compared to discriminative classifiers [17]. Moreover, generative classifiers are more robust to continual learning [18] and imbalanced data settings [19]. At each iteration, B n ∪ B r is also highly imbalanced.…”
Section: B Inference With a Generative Classifiermentioning
confidence: 99%
“…The trigger bias proposed in our paper belongs to selection bias and model overamplification bias. Bias has also been investigated in natural language inference [1,6,7,13,[21][22][23], question answering [24], ROC story cloze [2,28], lexical inference [17], visual question answering [12], etc. To our best knowledge, we are the first to present the biases in FSEC, i.e., trigger overlapping and trigger separability.…”
Section: Few-shot Event Classificationmentioning
confidence: 99%
“…Sanchez et al (2018) analysed the behaviour of NLI models and the factors to be more robust. Ding et al (2020) proposed efficient methods to mitigate a particular known bias in NLI. Benchmark collection in NLI: GLUE (Wang et al, 2019b,a) benchmark contains several NLIrelated benchmark datasets.…”
Section: Related Workmentioning
confidence: 99%