DOI: 10.1007/978-3-540-85853-9_32
|View full text |Cite
|
Sign up to set email alerts
|

Extrinsic Summarization Evaluation: A Decision Audit Task

Abstract: In this work we describe a large-scale extrinsic evaluation of automatic speech summarization technologies for meeting speech. The particular task is a decision audit, wherein a user must satisfy a complex information need, navigating several meetings in order to gain an understanding of how and why a given decision was made. We compare the usefulness of extractive and abstractive technologies in satisfying this information need, and assess the impact of automatic speech recognition (ASR) errors on user perfor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
22
0

Publication Types

Select...
6
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 19 publications
(24 citation statements)
references
References 21 publications
2
22
0
Order By: Relevance
“…ROUGE). We agree with [30] that the evaluation of abstractive summaries must be extrinsic, i.e. based on human judgement or indirectly observed as its helpfulness in performing a given task.…”
Section: Future Worksupporting
confidence: 84%
“…ROUGE). We agree with [30] that the evaluation of abstractive summaries must be extrinsic, i.e. based on human judgement or indirectly observed as its helpfulness in performing a given task.…”
Section: Future Worksupporting
confidence: 84%
“…Hence, another trend is to use the sentences selected in the summaries as starting point for browsing the meetings. This helps users recontextualize the information and improve their ability to locate information as shown by [71]. To this end, in [69], we proposed a user interface for improving the capture of a user's information need by presenting automatically extracted keyphrases that can be refined and used to generate summaries for meeting browsing.…”
Section: Summarizationmentioning
confidence: 99%
“…However, ROUGE does not correlate well with human judgements for meeting summarisation (Liu and Liu 2008). In order to better evaluate our meeting summarisation systems, we carried out a subjective evaluation based on a decision audit task (Murray et al 2009). Users were asked to find the factors that led to a particular decision that resulted from a sequence of four meetings in the AMI corpus.…”
Section: Summarisationmentioning
confidence: 99%