2020
DOI: 10.48550/arxiv.2010.04529
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

What Have We Achieved on Text Summarization?

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(10 citation statements)
references
References 35 publications
0
10
0
Order By: Relevance
“…The author further argues that community has exerted over crafting algorithms to improve performance evaluation scores on the benchmark data-sets, thereby limiting progress in the science of automatic extractive document summarization. Similar searching questions have been asked by Huang et al (2020), who design a multi-dimensional quality metric and quantify major sources of errors on well-known summarization models. The authors underscore faithfulness and factual-consistency of extractive summaries compared to abstractive counterparts, based on the designed metric.…”
Section: Introductionmentioning
confidence: 91%
See 2 more Smart Citations
“…The author further argues that community has exerted over crafting algorithms to improve performance evaluation scores on the benchmark data-sets, thereby limiting progress in the science of automatic extractive document summarization. Similar searching questions have been asked by Huang et al (2020), who design a multi-dimensional quality metric and quantify major sources of errors on well-known summarization models. The authors underscore faithfulness and factual-consistency of extractive summaries compared to abstractive counterparts, based on the designed metric.…”
Section: Introductionmentioning
confidence: 91%
“…During the last two years, there has been a spurt in research related to metrics for summary quality (Peyrard, 2019b;Bhandari et al, 2020a;Huang et al, 2020;Vasilyev & Bohannon, 2020;Fabbri et al, 2020;Bhandari et al, 2020b). Most of these works have argued against the ROUGE metric because it fails to robustly match paraphrases resulting in misleading scores, which do not correlate well with human judgements (Zhang et al, 2019;Huang et al, 2020). Zhong et al (2020a) argue that high semantically similarity with the source document is highly desirable for a good summary.…”
Section: Evaluation Metricsmentioning
confidence: 99%
See 1 more Smart Citation
“…There has been much work analyzing Seq2Seq models which is always task-specific and based on automatic or human evaluation. For example, Huang et al (2020) analyze the common models' performance on summarization.…”
Section: Related Workmentioning
confidence: 99%
“…We want to know the pros and cons of different Seq2Seq models on this task, and the factors influencing the generation performance. Particularly, following Multidimensional Quality Metric(MQM) (Mariana, 2014), similar to the job on summarization evaluation (Huang et al, 2020), we use 8 metrics on the Accuracy and Fluency aspects to count errors, respectively. Therefore, compared with existing manual evaluation reports, it is more informative and objective.…”
Section: Introductionmentioning
confidence: 99%