Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing 2018
DOI: 10.18653/v1/d18-1445
|View full text |Cite
|
Sign up to set email alerts
|

APRIL: Interactively Learning to Summarise by Combining Active Preference Learning and Reinforcement Learning

Abstract: We propose a method to perform automatic document summarisation without using reference summaries. Instead, our method interactively learns from users' preferences. The merit of preference-based interactive summarisation is that preferences are easier for users to provide than reference summaries. Existing preference-based interactive learning methods suffer from high sample complexity, i.e. they need to interact with the oracle for many rounds in order to converge. In this work, we propose a new objective fun… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
38
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 29 publications
(47 citation statements)
references
References 18 publications
1
38
0
Order By: Relevance
“…This work extends our earlier work (Gao et al 2018) in three aspects. (i) We present a new user study on the reliability and usability of the preferencebased interaction ( §5).…”
Section: Introductionsupporting
confidence: 84%
See 2 more Smart Citations
“…This work extends our earlier work (Gao et al 2018) in three aspects. (i) We present a new user study on the reliability and usability of the preferencebased interaction ( §5).…”
Section: Introductionsupporting
confidence: 84%
“…Note that in all previous work we are aware of (P.V.S. and Meyer 2017; Kreutzer et al 2017;Gao et al 2018), the evaluation was based on simulations with a perfect user oracle. Therefore, we expect that our results with real user interaction better reflect the true results.…”
Section: April Vs Sppimentioning
confidence: 99%
See 1 more Smart Citation
“…However, besides the fact that results were obtained from other domains than the presented telecommunication domain in this work, in both works, the authors did not apply any pre-qualification test or did not provide information about crowdsourcing task details, which can also cause a rather large influencing effect. Following, Gao et al (2018); Falke et al (2017); Fan et al (2018) have used crowdsourcing as the source of human evaluation to rate their automatic summarization systems. Nevertheless, they did not question the robustness of crowdsourcing for this task and compared the crowd with expert data.…”
Section: Crowdsourcing For Summarization Evaluationmentioning
confidence: 99%
“…In particular, the human scores for some documents might be on average higher than for other documents, which easily confuses the regression. Preference learning (PL) is robust to these issues, by learning the relative ordering induced by the human scores (Gao et al, 2018). PL can be formulated as a binary classification task (Maystre, 2018), where the input is a pair of data points {(S i , D i , h i ), (S j , D j , h j )} and the output is a binary flag indicating whether S i is better than S j , i.e., h i > h j :…”
Section: Inferring K With Human Judgmentsmentioning
confidence: 99%