2020
DOI: 10.48550/arxiv.2006.03210
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Sentence Compression as Deletion with Contextual Embeddings

Abstract: Sentence compression is the task of creating a shorter version of an input sentence while keeping important information. In this paper, we extend the task of compression by deletion with the use of contextual embeddings. Different from prior work usually using non-contextual embeddings (Glove or Word2Vec), we exploit contextual embeddings that enable our model capturing the context of inputs. More precisely, we utilize contextual embeddings stacked by bidirectional Long-short Term Memory and Conditional Random… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 11 publications
(41 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?