Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.699
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Text Style Transfer with Padded Masked Language Models

Abstract: We propose MASKER, an unsupervised textediting method for style transfer. To tackle cases when no parallel source-target pairs are available, we train masked language models (MLMs) for both the source and the target domain. Then we find the text spans where the two models disagree the most in terms of likelihood. This allows us to identify the source tokens to delete to transform the source text to match the style of the target domain. The deleted tokens are replaced with the target MLM, and by using a padded … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
46
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 44 publications
(58 citation statements)
references
References 18 publications
0
46
0
Order By: Relevance
“…In addition to comparing end-to-end systems, we also compare LEWIS to concurrent editing and synthesis methods by Malmi et al (2019Malmi et al ( , 2020. Table 2 shows that training the same model (LaserTagger) on our data improves and BLEU by 9.5 (the accuracy difference is not directly comparable since Malmi et al (2020) used a BERT classifier and did not release model output). This suggests that our data synthesis procedure produces higher quality data than Malmi et al (2020).…”
Section: Resultsmentioning
confidence: 99%
“…In addition to comparing end-to-end systems, we also compare LEWIS to concurrent editing and synthesis methods by Malmi et al (2019Malmi et al ( , 2020. Table 2 shows that training the same model (LaserTagger) on our data improves and BLEU by 9.5 (the accuracy difference is not directly comparable since Malmi et al (2020) used a BERT classifier and did not release model output). This suggests that our data synthesis procedure produces higher quality data than Malmi et al (2020).…”
Section: Resultsmentioning
confidence: 99%
“…Finally, by allowing control through control codes and [BLANK]s, Polyjuice supports humangenerator collaboration, where a person specifies desired changes (e.g., perturb the sentence subject). Such collaboration is hard to imagine using automatic generators with no control, or with coarser control through predefined style attributes or labels (Madaan et al, 2020;Malmi et al, 2020). To our knowledge, prior work on controlled generation (Keskar et al, 2019;Dathathri et al, 2020) does not address counterfactual generation.…”
Section: Related Workmentioning
confidence: 99%
“…However, they are recently criticized for resulting in poor content preservation (Xu et al, 2018;Jin et al, 2019;Subramanian et al, 2018) and alternatively, translation-based models are proposed that use reconstruction and back-translation losses (e.g., Logeswaran et al (2018); Prabhumoye et al (2018)). Another line of work, focuses on manipulation methods that remove the style-specific attribute of text (e.g., ; Xu et al (2018)), while recent approaches use reinforcement learning (e.g., ; Gong et al (2019), probabilistic formulations (He et al, 2020), and masked language models (Malmi et al, 2020).…”
Section: Related Workmentioning
confidence: 99%