2020
DOI: 10.1007/978-3-030-45439-5_36
|View full text |Cite
|
Sign up to set email alerts
|

Reinforced Rewards Framework for Text Style Transfer

Abstract: Style transfer deals with the algorithms to transfer the stylistic properties of a piece of text into that of another while ensuring that the core content is preserved. There has been a lot of interest in the field of text style transfer due to its wide application to tailored text generation. Existing works evaluate the style transfer models based on content preservation and transfer strength. In this work, we propose a reinforcement learning based framework that directly rewards the framework on these target… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
23
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 22 publications
(25 citation statements)
references
References 27 publications
(62 reference statements)
0
23
0
Order By: Relevance
“…Evaluation Following previous work He et al, 2020;Sancheti et al, 2020), we adopt the following strategies. The binary classifier TextCNN (Kim, 2014) is pre-trained to evaluate style strength; on the human references it has an accuracy of 87.0% (E&M) and 89.3% (F&R).…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Evaluation Following previous work He et al, 2020;Sancheti et al, 2020), we adopt the following strategies. The binary classifier TextCNN (Kim, 2014) is pre-trained to evaluate style strength; on the human references it has an accuracy of 87.0% (E&M) and 89.3% (F&R).…”
Section: Methodsmentioning
confidence: 99%
“…For the GPT-2 based model, we also add a classification confidence reward to the source sentence, similar to Eq. 4, since the model generates sentence x with the original style while generating the target sentence: BLEU Score Reward Following Sancheti et al (2020), we introduce a BLEU-based reward to foster content preservation as in Eq. 6, where y is the target style text obtained by greedily maximizing the distribution of model outputs at each time step, and y s is sampled from the distribution.…”
Section: Rewardsmentioning
confidence: 99%
See 1 more Smart Citation
“…Recent works have also modeled this in a self-supervised manner where rewriting (transfer) is achieved by utilizing corpus from the target style alone (Syed et al, 2020). These wide studies have also led to the curation and benchmarking of non-parallel dataset for various style dimensions, such as sentiment (Li et al, 2018), formality (Rao and Tetreault, 2018), politeness (Danescu-Niculescu-Mizil et al, 2013), excitement (Sancheti et al, 2020), etc. But availability of data with joint tagging across multiple styles is limited and has restricted the ability of existing approaches to scale from single-dimensional transfer to multiple style dimensions.…”
Section: Introductionmentioning
confidence: 99%
“…We then fine-tune GePpeTto with this perceptionlabelled data. In addition, inspired by the classifierbased reward used in style transfer tasks (Lample et al, 2019;Gong et al, 2019;Luo et al, 2019;Sancheti et al, 2020), we reward the model to push its classification confidence. We evaluate the new perception-enhanced models in comparison with the original GePpeTto by running both an automatic as well as a human evaluation on output generated by the various models.…”
Section: Introductionmentioning
confidence: 99%