Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2016
DOI: 10.18653/v1/n16-1012
|View full text |Cite
|
Sign up to set email alerts
|

Abstractive Sentence Summarization with Attentive Recurrent Neural Networks

Abstract: ive Sentence Summarization generates a shorter version of a given sentence while attempting to preserve its meaning. We introduce a conditional recurrent neural network (RNN) which generates a summary of an input sentence. The conditioning is provided by a novel convolutional attention-based encoder which ensures that the decoder focuses on the appropriate input words at each step of generation. Our model relies only on learned features and is easy to train in an end-to-end fashion on large data sets. Our expe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
593
0
1

Year Published

2017
2017
2019
2019

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 779 publications
(625 citation statements)
references
References 15 publications
(13 reference statements)
2
593
0
1
Order By: Relevance
“…Using OpenNMT, we were able to replicate the sentence summarization results of Chopra et al (2016), reaching a ROUGE-1 score of 33.13 on the Gigaword data. We have also trained a model on 14 million sentences of the OpenSubtitles data set based on the work Vinyals and Le (2015), achieving comparable perplexity.…”
Section: Benchmarksmentioning
confidence: 97%
“…Using OpenNMT, we were able to replicate the sentence summarization results of Chopra et al (2016), reaching a ROUGE-1 score of 33.13 on the Gigaword data. We have also trained a model on 14 million sentences of the OpenSubtitles data set based on the work Vinyals and Le (2015), achieving comparable perplexity.…”
Section: Benchmarksmentioning
confidence: 97%
“…Our baseline model is a strong, multi-layered encoder-attention-decoder model with bilinear attention, similar to Luong et al (2015) and following the details in Chopra et al (2016). Here, we encode the source document with a two-layered LSTM-RNN and generate the summary using another two-layered LSTM-RNN decoder.…”
Section: Baseline Modelmentioning
confidence: 99%
“…The pure data-driven end-to-end automatic summarization generation method was originally borrowed from the neural network model of machine translation [3,4]. K Lopvrev et al built an abstract generation model based on the encoder-decoder framework in 2015 by using RNN(Recurrent Neural Network) with unit of LSTM(Long Short-Term Memory) [5] and used an attention mechanism to generate news headlines [6].Secondly, the two papers [7,8] published by Rush et al from the Facebook Artificial Intelligence Research Institute from 2015 to 2016 to solve the text abstract generation task, based on the Encoder-Decoder architecture, proposed different encoder approaches based CNN(Convolutional Neural Network) and attention mechanisms, and decoder architecture based on the RNNLM(Recurrent Neural Network Language Model). Hu et al [9] applied RNN-based Encoder-Decoder architecture to Chinese text digest tasks and constructed a Chinese text digest dataset LCSTS to facilitate the study of Chinese comprehension abstracts.…”
Section: Introductionmentioning
confidence: 99%