Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.407
|View full text |Cite
|
Sign up to set email alerts
|

Multilevel Text Alignment with Cross-Document Attention

Abstract: Text alignment finds application in tasks such as citation recommendation and plagiarism detection. Existing alignment methods operate at a single, predefined level and cannot learn to align texts at, for example, sentence and document levels. We propose a new learning approach that equips previously established hierarchical attention encoders for representing documents with a cross-document attention component, enabling structural comparisons across different levels (document-to-document and sentence-to-docum… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
44
3

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 13 publications
(51 citation statements)
references
References 34 publications
2
44
3
Order By: Relevance
“…Despite these factors, BERT-HAN's large performance drop on PAN is still surprising. However, we emphasize that even when using Zhou et al (2020)'s original numbers, BERT-HAN still lags behind both our lexical overlap baselines and fine-tuned BERT models, so our overall takeaways from §7 still stand. For the S2D task, our results are not directly comparable to the original numbers of Zhou et al (2020) for two reasons: positively-labeled target sentences.…”
Section: B Implementation Of Bert-han and Gru-hanmentioning
confidence: 90%
See 4 more Smart Citations
“…Despite these factors, BERT-HAN's large performance drop on PAN is still surprising. However, we emphasize that even when using Zhou et al (2020)'s original numbers, BERT-HAN still lags behind both our lexical overlap baselines and fine-tuned BERT models, so our overall takeaways from §7 still stand. For the S2D task, our results are not directly comparable to the original numbers of Zhou et al (2020) for two reasons: positively-labeled target sentences.…”
Section: B Implementation Of Bert-han and Gru-hanmentioning
confidence: 90%
“…However, we emphasize that even when using Zhou et al (2020)'s original numbers, BERT-HAN still lags behind both our lexical overlap baselines and fine-tuned BERT models, so our overall takeaways from §7 still stand. For the S2D task, our results are not directly comparable to the original numbers of Zhou et al (2020) for two reasons: positively-labeled target sentences. When there are fewer than k positively-labeled target sentences in an example, a perfect system will still have a P@k < 1.…”
Section: B Implementation Of Bert-han and Gru-hanmentioning
confidence: 90%
See 3 more Smart Citations