2022
DOI: 10.48550/arxiv.2212.09272
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Statistical Dataset Evaluation: Reliability, Difficulty, and Validity

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 0 publications
0
2
0
Order By: Relevance
“…To understand the performance of the methods on the datasets, we calculated the difficulty of the datasets and the similarity between the train and test sets of datasets. As difficulty metrics, we use 2 metrics: Entity Ambugity Degree (EAD), and Text Complexity (TC) (Wang et al, 2022a). We also use Target Vocabulary Covered (TVC) as similarity metric (Dai et al, 2019).…”
Section: A Dataset Detailsmentioning
confidence: 99%
“…To understand the performance of the methods on the datasets, we calculated the difficulty of the datasets and the similarity between the train and test sets of datasets. As difficulty metrics, we use 2 metrics: Entity Ambugity Degree (EAD), and Text Complexity (TC) (Wang et al, 2022a). We also use Target Vocabulary Covered (TVC) as similarity metric (Dai et al, 2019).…”
Section: A Dataset Detailsmentioning
confidence: 99%
“…We further show that near-duplicate analysis is useful in at least two ways. First, it allows us to inspect and refine a dataset, in a manner similar to measuring data (Wang et al, 2022;Mitchell et al, 2022, inter alia), by identifying phenomena that might otherwise go unnoticed, e.g. texts that are assigned to different classes but have no actual dialectal differences or spotting artefacts due to the selection of text sources or to the processing pipeline (e.g.…”
Section: Introductionmentioning
confidence: 99%