Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing 2017
DOI: 10.18653/v1/d17-1136
|View full text |Cite
|
Sign up to set email alerts
|

How much progress have we made on RST discourse parsing? A replication study of recent results on the RST-DT

Abstract: This article evaluates purported progress over the past years in RST discourse parsing. Several studies report a relative error reduction of 24 to 51% on all metrics that authors attribute to the introduction of distributed representations of discourse units. We replicate the standard evaluation of 9 parsers, 5 of which use distributed representations, from 8 studies published between 2013 and 2017, using their predictions on the test set of the RST-DT. Our main finding is that most recently reported increases… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
76
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 57 publications
(81 citation statements)
references
References 13 publications
3
76
0
Order By: Relevance
“…The human agreement results are 86.8%, 72.2%, and 58.2%, according to the span, nuclearity, and relation levels respectively. This level of agreement is similar to the inter-annotator agreement rates on the RST Discourse Treebank, i.e., 88.3% on span, 77.3% on nuclearity, and 64.7% on relation, respectively (Joty et al, 2015;Morey et al, 2017).…”
Section: Human Annotationssupporting
confidence: 80%
See 2 more Smart Citations
“…The human agreement results are 86.8%, 72.2%, and 58.2%, according to the span, nuclearity, and relation levels respectively. This level of agreement is similar to the inter-annotator agreement rates on the RST Discourse Treebank, i.e., 88.3% on span, 77.3% on nuclearity, and 64.7% on relation, respectively (Joty et al, 2015;Morey et al, 2017).…”
Section: Human Annotationssupporting
confidence: 80%
“…In this paper, all the parsers we built were evaluated with the micro F1 score. When using the gold standard syntax trees and EDU segmentations, the F1 scores on three levels of span, nuclearity, and relation can reach 84.1%, 69.6%, and 56.5% respectively, which are close to state-of-the-art accuracy, as reported in Morey et al (2017).…”
Section: Parser Trainingsupporting
confidence: 67%
See 1 more Smart Citation
“…The SciDTB annotations use 17 coarse-grained relation types and 26 finegrained relations. Polynary discourse relations in RST are binarized in SciDTB following a criteria similar to the "right-heavy" transformation used in other works that represent discourse structures as dependency trees [19,28,13], which makes it particularly suitable as input of sequence tagging algorithms.…”
Section: Scidtb Corpusmentioning
confidence: 99%
“…The second scenario is that several EDUs are of equal importance in a polynary relation. For this case, we link each former EDU to its neighboring EDU with the same relation, forming a relation chain similar to "right-heavy" binarization transformation in (Morey et al, 2017).…”
Section: Discourse Relationsmentioning
confidence: 99%