Proceedings of the 2018 Conference of the North American Chapter Of the Association for Computational Linguistics: Hu 2018
DOI: 10.18653/v1/n18-2097
|View full text |Cite
|
Sign up to set email alerts
|

A Discourse-Aware Attention Model for Abstractive Summarization of Long Documents

Abstract: Neural abstractive summarization models have led to promising results in summarizing relatively short documents. We propose the first model for abstractive summarization of single, longer-form documents (e.g., research papers). Our approach consists of a new hierarchical encoder that models the discourse structure of a document, and an attentive discourse-aware decoder to generate the summary. Empirical results on two large-scale datasets of scientific papers show that our model significantly outperforms state… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
470
0
1

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 450 publications
(533 citation statements)
references
References 29 publications
1
470
0
1
Order By: Relevance
“…Earlier on, it was realized that summarizing scientific papers requires different approaches than what was used for summarizing news articles, due to differences in document length, writing style and rhetorical structure. For instance, (Teufel and Moens, 2002) presented a supervised Naive Bayes (Cohan et al, 2018), the length is in terms of the number of words classifier to select content from a scientific paper based on the rhetorical status of each sentence (e.g., whether it specified a research goal, or some generally accepted scientific background knowledge, etc.). More recently, researchers have extended this work by applying more sophisticated classifiers to identify more fine-grain rhetorical categories, as well as by exploiting citation contexts.…”
Section: Extractive Summarization On Scientific Papersmentioning
confidence: 99%
“…Earlier on, it was realized that summarizing scientific papers requires different approaches than what was used for summarizing news articles, due to differences in document length, writing style and rhetorical structure. For instance, (Teufel and Moens, 2002) presented a supervised Naive Bayes (Cohan et al, 2018), the length is in terms of the number of words classifier to select content from a scientific paper based on the rhetorical status of each sentence (e.g., whether it specified a research goal, or some generally accepted scientific background knowledge, etc.). More recently, researchers have extended this work by applying more sophisticated classifiers to identify more fine-grain rhetorical categories, as well as by exploiting citation contexts.…”
Section: Extractive Summarization On Scientific Papersmentioning
confidence: 99%
“…Since record encoders with record fusion gate provide record-level representation and row-level encoder provides row-level representation. Inspired by Cohan et al (2018), we can modify the decoder in base model to first choose important row then attend to records when generating each word. Following notations in Section 2.3, β t,i ∝ exp(score(d t , row i )) obtains the attention weight with respect to each row.…”
Section: Decoder With Dual Attentionmentioning
confidence: 99%
“…In addition, we implemented three other hierarchical encoders that encoded tables' row dimension information in both record-level and row-level to compare with the hierarchical structure of encoder in our model. The decoder was equipped with dual attention (Cohan et al, 2018). The one with LSTM cell is similar to the one in Cohan et al (2018) with 1 layer from {1,2,3}.…”
Section: Automatic Evaluationmentioning
confidence: 99%
See 1 more Smart Citation
“…1 Introduction is to analyze and understand the impact on the models' generalization ability from a dataset perspective. With the emergence of more and more summarization datasets (Sandhaus, 2008;Nallapati et al, 2016;Cohan et al, 2018;Grusky et al, 2018), the time is ripe for us to bridge the gap between the insufficient understanding of the nature of datasets themselves and the increasing improvement of the learning methods.…”
mentioning
confidence: 99%