Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.170
|View full text |Cite
|
Sign up to set email alerts
|

Changing the World by Changing the Data

Abstract: NLP community is currently investing a lot more research and resources into development of deep learning models than training data. While we have made a lot of progress, it is now clear that our models learn all kinds of spurious patterns, social biases, and annotation artifacts. Algorithmic solutions have so far had limited success. An alternative that is being actively discussed is more careful design of datasets so as to deliver specific signals. This position paper maps out the arguments for and against da… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
28
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 37 publications
(37 citation statements)
references
References 42 publications
(48 reference statements)
0
28
0
Order By: Relevance
“…Web-scale textual or speech data is harder to obtain for Maltese compared to languages such as English or Mandarin and we expect that similar challenges arise for many other under-resourced languages (Besacier et al, 2014). Furthermore, as recent critiques of large-scale pretraining approaches have emphasised, even where web-scale data is available, there are significant risks arising from its 'un-fathomable' nature, not least that it is likely to be extremely noisy, while not guaranteeing representativeness across demographic or ethnic groups, and/or across language varieties Bender et al (2021); Rogers (2021). Lastly, the computational resources needed for such experiments are not available to all research teams.…”
Section: Introductionmentioning
confidence: 93%
“…Web-scale textual or speech data is harder to obtain for Maltese compared to languages such as English or Mandarin and we expect that similar challenges arise for many other under-resourced languages (Besacier et al, 2014). Furthermore, as recent critiques of large-scale pretraining approaches have emphasised, even where web-scale data is available, there are significant risks arising from its 'un-fathomable' nature, not least that it is likely to be extremely noisy, while not guaranteeing representativeness across demographic or ethnic groups, and/or across language varieties Bender et al (2021); Rogers (2021). Lastly, the computational resources needed for such experiments are not available to all research teams.…”
Section: Introductionmentioning
confidence: 93%
“…One aim of this work is to document failures in evaluation processes that frequently happen, many of which can be directly addressed or pointed out in reviews. In the future, we also suggest creating model evaluation checklists like those by Rogers et al (2021) for responsible data use or Dodge et al (2019) for reporting hyperparameters and compute infrastructure.…”
Section: Model Audits and Evaluation Reportsmentioning
confidence: 99%
“…The argument for curating datasets Whether or not datasets should be curated to alter underlying distributions is a foundational issue. Rogers [45] summarize the arguments for and against curation that followed after the publication of work by Bender et al [7] which argued strongly for curation. One of the core questions is whether we should study the world as it is or the world as we want it to be, where "world" refers to extant sources of data, such as Wikipedia.…”
Section: Synthetic Dataset Creation and Augmentationmentioning
confidence: 99%