Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.480
|View full text |Cite
|
Sign up to set email alerts
|

Learning to Selectively Learn for Weakly-supervised Paraphrase Generation

Abstract: Paraphrase generation is a longstanding NLP task that has diverse applications for downstream NLP tasks. However, the effectiveness of existing efforts predominantly relies on large amounts of golden labeled data. Though unsupervised endeavors have been proposed to address this issue, they may fail to generate meaningful paraphrases due to the lack of supervision signals. In this work, we go beyond the existing paradigms and propose a novel approach to generate high-quality paraphrases with weak supervision da… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(9 citation statements)
references
References 44 publications
(13 reference statements)
0
8
0
Order By: Relevance
“…In our experiments, we use the performance gain on both NDCG and Hit Rate as the reward for offline experiment, and use the change of simulated rating as the reward for online experiment. Following [11,54,57], the policy network for data augmentation is updated on a delayed reward received after feeding the generated augmented data to the recommender.…”
Section: Learning Augmentation Policymentioning
confidence: 99%
“…In our experiments, we use the performance gain on both NDCG and Hit Rate as the reward for offline experiment, and use the change of simulated rating as the reward for online experiment. Following [11,54,57], the policy network for data augmentation is updated on a delayed reward received after feeding the generated augmented data to the recommender.…”
Section: Learning Augmentation Policymentioning
confidence: 99%
“…Compared to the conventional reinforcement learning methods which consider the generators as the policy models, our work models the policy as a meta learner to accomplish a data selection objective. Our work is mostly related to (Ding et al, 2021), but we adopt a very different reinforcement learning approach which is the key for effective selective learning.…”
Section: Related Workmentioning
confidence: 99%
“…Though such attempts have demonstrated certain efficacy in handling instance-wise feature selection, they only deal with non timeseries data in non NLP domains, while the focus of our work is to deal with noisy labeled pairs in paraphrase generation tasks. Our work is mostly related to the instance-level active data acquisition approaches (Yoon et al, 2020;Ding et al, 2021), which are mostly adopted under the circumstances of data efficient or cost-sensitive learning or when dealing with noisy data.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations