Findings of the Association for Computational Linguistics: EMNLP 2021 2021
DOI: 10.18653/v1/2021.findings-emnlp.377
|View full text |Cite
|
Sign up to set email alerts
|

An Exploratory Study on Long Dialogue Summarization: What Works and What’s Next

Abstract: Dialogue summarization helps readers capture salient information from long conversations in meetings, interviews, and TV series. However, real-world dialogues pose a great challenge to current summarization models, as the dialogue length typically exceeds the input limits imposed by recent transformer-based pretrained models, and the interactive nature of dialogues makes relevant information more context-dependent and sparsely distributed than news articles. In this work, we perform a comprehensive study on lo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 17 publications
(6 citation statements)
references
References 21 publications
0
4
0
Order By: Relevance
“…This extra step reduces the burden on the neural summarizers that have to generate an abstract summary and select important content at the same time. Some also refer models that utilize the hybrid approach as retrieve-the-summarize model because it involves retrieving a subset of long document text before summarizing it [131]. TLM+Ext [94] irst implemented this method by limiting inputs of the scientiic articles in arXiv datasets as the introduction of the document, a subset of carefully selected sentences of the original article using extractive summarization approach, and, inally, include the remaining text if there remains extra space for Transformer-based decoder.…”
Section: ) Discourse Biasmentioning
confidence: 99%
See 2 more Smart Citations
“…This extra step reduces the burden on the neural summarizers that have to generate an abstract summary and select important content at the same time. Some also refer models that utilize the hybrid approach as retrieve-the-summarize model because it involves retrieving a subset of long document text before summarizing it [131]. TLM+Ext [94] irst implemented this method by limiting inputs of the scientiic articles in arXiv datasets as the introduction of the document, a subset of carefully selected sentences of the original article using extractive summarization approach, and, inally, include the remaining text if there remains extra space for Transformer-based decoder.…”
Section: ) Discourse Biasmentioning
confidence: 99%
“…Other than the discourse bias mechanism, we observe that (a) eicient attention and (b) content selection mechanisms are the two most notable long document mechanisms. As the content selection mechanism requires a separate retriever to extract salient content from the source (i.e., the hybrid approach), we distinguish Transformer models with content selection mechanism as the retrieve-then-summarize model [131] and the pure encoder-decoder Transformer without this mechanism as an end-to-end model for the rest of this work. Lastly, it is also important to note that both mechanisms can be jointly implemented within a single architecture, where the content selection mechanism will extract a longer subset of input to be processed by a Transformer with eicient attention [77].…”
Section: Supervised Hybridmentioning
confidence: 99%
See 1 more Smart Citation
“…(3) Our summaries adhere to people-personalized preferences. In comparisons with strong LLM summarization methods on the standard datasets MACSum [4] and arXiv [5], which contain long document summarization examples, our approach demonstrated clear advantages.…”
Section: Introductionmentioning
confidence: 98%
“…The results of the Longformer-based model in AMI and ICSI are from(Fabbri et al 2021), and the results of it in QMSum come from(Zhang et al 2021b). In screenplay domain, the results of Longformer are fromChen et al (2021a).3 BART-large-CNN refers to further fine-tuning BART-large on the news summarization dataset CNN/DailyMail.…”
mentioning
confidence: 99%