RANLP 2017 - Recent Advances in Natural Language Processing Meet Deep Learning 2017
DOI: 10.26615/978-954-452-049-6_026
|View full text |Cite
|
Sign up to set email alerts
|

Towards Replicability in Parsing

Abstract: We investigate parsing replicability across 7 languages (and 8 treebanks), showing that choices concerning the use of grammatical functions in parsing or evaluation and the influence of the rare word threshold, as well as choices in test sentences and evaluation script options have considerable and often unexpected effects on parsing accuracies. All of those choices need to be carefully documented if we want to ensure replicability.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(4 citation statements)
references
References 15 publications
0
4
0
Order By: Relevance
“…Branco et al (2016), Branco et al (2018) and Branco et al (2020)), encouragingly, there is an increasing awareness around the need for reproducibility in the field (e.g. Dakota andKübler (2017), Wieling et al (2018) and Cohen et al (2018)), while some of the major conferences (e.g. ACL-IJCNLP 22 ) now encourage authors to submit supplementary material to facilitate others in reproducing their results.…”
Section: Discussionmentioning
confidence: 97%
“…Branco et al (2016), Branco et al (2018) and Branco et al (2020)), encouragingly, there is an increasing awareness around the need for reproducibility in the field (e.g. Dakota andKübler (2017), Wieling et al (2018) and Cohen et al (2018)), while some of the major conferences (e.g. ACL-IJCNLP 22 ) now encourage authors to submit supplementary material to facilitate others in reproducing their results.…”
Section: Discussionmentioning
confidence: 97%
“…According to Fokkens et al (2013) and Wieling et al (2018), the main challenge is the unavailability of the source code and data. Dakota and Kübler (2017) study reproducibility for text mining. They show that 80% of the failed reproduction attempts were due to the lack of information about the datasets.…”
Section: Reproducibility In Nlpmentioning
confidence: 99%
“…Using the Berkeley parser (Petrov and Klein 2007), the Trance parser (Watanabe and Sumita 2015) and Berkeley neural parser (Kitaev and Klein 2018;Kitaev et al 2019), we train and evaluate the phrase structure Korean Sejong treebank. As shown in Dakota and Kübler (2017) for other languages and treebank, we define the best practices for replicability in constituent parsing for Korean, including correct comparison to future work by proposing the standard corpus division for the Sejong treebank. The other main contribution of this paper is to present two important factors on constituent parsing for Korean: detailed qualitative and quantitative parsing error analyses (Sections 3 and 4, respectively) using the Sejong treebank.…”
Section: Goal Of the Papermentioning
confidence: 99%