Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 2021
DOI: 10.18653/v1/2021.findings-acl.343
|View full text |Cite
|
Sign up to set email alerts
|

Scaling Within Document Coreference to Long Texts

Abstract: State of the art end-to-end coreference resolution models use expensive span representations and antecedent prediction mechanisms. These approaches are expensive both in terms of their memory requirements as well as compute time, and are particularly ill-suited for long documents. In this paper, we propose an approximation to end-to-end models which scales gracefully to documents of any length. Replacing span representations with token representations, we reduce the time/memory complexity via token windows and… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2

Relationship

2
5

Authors

Journals

citations
Cited by 9 publications
(6 citation statements)
references
References 16 publications
(56 reference statements)
0
6
0
Order By: Relevance
“…In Table 3 On the other hand, the c2f-coref opt + ELECTRA-large model reaches inferior performance than CorefQA+ SpanBERT-large, but without neither resorting to data augmentation to improve its generalization capability nor processing hundreds of individual context-questionanswer instances for a single document, substantially worsening execution time, as reported by [13]. As further stated in [21], [24], this method has resulted very computationally expensive since it needs to run a transformer-based model to perform a different query on the same document many times. It also exhibiting some difficulties to scale to long documents.…”
Section: A Quantitative Analysismentioning
confidence: 77%
“…In Table 3 On the other hand, the c2f-coref opt + ELECTRA-large model reaches inferior performance than CorefQA+ SpanBERT-large, but without neither resorting to data augmentation to improve its generalization capability nor processing hundreds of individual context-questionanswer instances for a single document, substantially worsening execution time, as reported by [13]. As further stated in [21], [24], this method has resulted very computationally expensive since it needs to run a transformer-based model to perform a different query on the same document many times. It also exhibiting some difficulties to scale to long documents.…”
Section: A Quantitative Analysismentioning
confidence: 77%
“…Much of the recent work on coreference can be organized into three categories: span based representations (Lee et al, 2017;Joshi et al, 2020), token-wise representations (Thirukovalluru et al, 2021;Kirstain et al, 2021) and memory networks / incremental models (Toshniwal et al, 2020b,a). We consider one approach from all three categories.…”
Section: Modelsmentioning
confidence: 99%
“…Our proposed dataset, LongtoNotes, restores documents to their original form, revealing dramatic increases in length in certain genres. Sachan et al, 2015;Wiseman et al, 2016;Lee et al, 2017;Joshi et al, 2020;Toshniwal et al, 2020b;Thirukovalluru et al, 2021;Kirstain et al, 2021).…”
Section: Introductionmentioning
confidence: 99%
“…Swayamdipta et al (2018) also leverage syntactic span classification as an auxiliary task to assist coreference. Thirukovalluru et al (2021), Kirstain et al (2021), andDobrovolskii (2021) explore token-level representations to both reduce memory consumption and increase performance on longer documents. Miculicich and Henderson (2020) and Yu et al (2020) both improve the mention detector with better neural network structures.…”
Section: Coreference Resolutionmentioning
confidence: 99%