Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics 2019
DOI: 10.18653/v1/p19-1605
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Paraphrasing without Translation

Abstract: Paraphrasing exemplifies the ability to abstract semantic content from surface forms. Recent work on automatic paraphrasing is dominated by methods leveraging Machine Translation (MT) as an intermediate step. This contrasts with humans, who can paraphrase without being bilingual. This work proposes to learn paraphrasing models from an unlabeled monolingual corpus only. To that end, we propose a residual variant of vector-quantized variational auto-encoder. We compare with MT-based approaches on paraphrase iden… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
48
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
6
2
2

Relationship

0
10

Authors

Journals

citations
Cited by 45 publications
(54 citation statements)
references
References 29 publications
0
48
0
Order By: Relevance
“…The general approach is to build a paraphrase generation model, usually a neural model (Prakash et al, 2016, Iyyer et al, 2018, Gupta et al, 2017, using general-purpose datasets of paraphrase sentence pairs. Data augmentation through neural paraphrasing models has been applied to various tasks such as sentiment analysis (Iyyer et al, 2018), intent classification (Roy and Grangier, 2019), and span-based question answering (Yu et al, 2018a). Paraphrasing models may generate training examples that do not match the original label.…”
Section: Related Workmentioning
confidence: 99%
“…The general approach is to build a paraphrase generation model, usually a neural model (Prakash et al, 2016, Iyyer et al, 2018, Gupta et al, 2017, using general-purpose datasets of paraphrase sentence pairs. Data augmentation through neural paraphrasing models has been applied to various tasks such as sentiment analysis (Iyyer et al, 2018), intent classification (Roy and Grangier, 2019), and span-based question answering (Yu et al, 2018a). Paraphrasing models may generate training examples that do not match the original label.…”
Section: Related Workmentioning
confidence: 99%
“…We show the results that can be achieved on large automatically ranked corpora using a Sequence-to-Sequence model based on the Universal Transformer architecture as it has demonstrated superior performance over the past year in multiple generative tasks, such as abstractive summarization, machine translation and, of course, paraphrase generation. (Gupta et al, 2018;Mallinson et al, 2017;Gupta et al, 2018;Fu et al, 2019;Egonmwan and Chali, 2019;Roy and Grangier, 2019).…”
Section: Paraphrase Generationmentioning
confidence: 99%
“…These have included rule-based approaches (McKeown, 1979;Meteer and Shaked, 1988) and data-driven methods (Madnani and Dorr, 2010), with recently the most common approach being that the task is treated as a language translation task (Bannard and Callison-Burch, 2005;Barzilay and McKeown, 2001;Pang et al, 2003) -often performed using a bilingual corpus pivoting back and forth (Madnani and Dorr, 2010;Prakash et al, 2016;Mallinson et al, 2017). Other methods proposed include more recently the use of Deep Reinforcement Learning (Li et al, 2018) , supervised learning using sequence-to-sequence models (Gupta et al, 2018;Prakash et al, 2016) and unsupervised approaches (Bowman et al, 2016;Roy and Grangier, 2019).…”
Section: Related Workmentioning
confidence: 99%