Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1425
|View full text |Cite
|
Sign up to set email alerts
|

Topics to Avoid: Demoting Latent Confounds in Text Classification

Abstract: Despite impressive performance on many text classification tasks, deep neural networks tend to learn frequent superficial patterns that are specific to the training data and do not always generalize well. In this work, we observe this limitation with respect to the task of native language identification. We find that standard text classifiers which perform well on the test set end up learning topical features which are confounds of the prediction task (e.g., if the input text mentions Sweden, the classifier pr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
40
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 30 publications
(41 citation statements)
references
References 30 publications
(31 reference statements)
0
40
0
Order By: Relevance
“…For the primary prediction models, we use the same architectures as Kumar et al (2019), including training multiple (2) adversaries. We perform minimal hyper-parameter tuning, primarily using the same parameters as Kumar et al (2019), with the exception of the learning rate, which we changed slightly to decrease fluctuations in validation accuracy, and the number of training epochs for each phase of the model, which we increased or decreased as needed based on how long the validation accuracy improved for. These changes were determined by manual tuning over < 10 trials.…”
Section: Discussionmentioning
confidence: 99%
See 4 more Smart Citations