2019
DOI: 10.3390/app9081665
|View full text |Cite
|
Sign up to set email alerts
|

Abstract Text Summarization with a Convolutional Seq2seq Model

Abstract: text summarization aims to offer a highly condensed and valuable information that expresses the main ideas of the text. Most previous researches focus on extractive models. In this work, we put forward a new generative model based on convolutional seq2seq architecture. A hierarchical CNN framework is much more efficient than the conventional RNN seq2seq models. We also equip our model with a copying mechanism to deal with the rare or unseen words. Additionally, we incorporate a hierarchical attention mechanism… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
22
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 47 publications
(22 citation statements)
references
References 24 publications
0
22
0
Order By: Relevance
“…The sequence to sequence model (Seq2Seq) has been widely used in processing tasks of variable length input and output sequences, including speech recognition, machine translation and so on [27][28] [29]. Its core idea is to map a variable length input sequence to variable length output sequence using cyclic neural network.…”
Section: B Seq2seq Modelmentioning
confidence: 99%
“…The sequence to sequence model (Seq2Seq) has been widely used in processing tasks of variable length input and output sequences, including speech recognition, machine translation and so on [27][28] [29]. Its core idea is to map a variable length input sequence to variable length output sequence using cyclic neural network.…”
Section: B Seq2seq Modelmentioning
confidence: 99%
“…The seq2seq framework introduced by Google was initially applied to NMT tasks [13,25]. Later, in the field of NLP, seq2seq models were also used for text summarization [26], parsing [27], or generative chatbots (as presented in Section 2). These models can address the challenge of a variable input and output length.…”
Section: Seq2seq Modelsmentioning
confidence: 99%
“…Zhou et al [9] proposed a joint learning model for sentence scoring and selection to lead the two tasks interact simultaneously, and multiple layer perceptron (MLP) is introduced to score sentences according to both the previously selected sentences and remains. Zhang et al [3] developed a hierarchical convolution model with an attention mechanism to extract keywords and key sentences simultaneously, and a copy mechanism was incorporated to resolve the problem of out of vocabulary (OOV) [22]. In addition, reinforcement learning (RL) has been proven to be effective in improving the performance of the summarization system [12,23] by allowing directly maximize the measure metric of summary quality, such as the ROUGE score between the generated summary and the ground truth.…”
Section: Related Workmentioning
confidence: 99%
“…Automatic summarization systems have been made great progress in many applications, such as headline generation [1], single or multi-document summarization [2,3], opinion mining [4], text categorization, etc. The system aims to shorten the input and retain the salient information from the source document.…”
Section: Introductionmentioning
confidence: 99%