Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.124
|View full text |Cite
|
Sign up to set email alerts
|

Improving Formality Style Transfer with Context-Aware Rule Injection

Abstract: Models pre-trained on large-scale regular text corpora often do not work well for usergenerated data where the language styles differ significantly from the mainstream text.Here we present Context-Aware Rule Injection (CARI), an innovative method for formality style transfer (FST). CARI injects multiple rules into an end-to-end BERT-based encoder and decoder model. It learns to select optimal rules based on context. The intrinsic evaluation showed that CARI achieved the new highest performance on the FST bench… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(10 citation statements)
references
References 31 publications
0
7
0
Order By: Relevance
“…Previous studies (Rao and Tetreault, 2018;Niu et al, 2018; typically train seq2seq encoder-decoder models on this benchmark. Recent studies (Wang et al, 2019;Yao and Yu, 2021;Chawla and Yang, 2020;Lai et al, 2021) have deduced that fine-tuning large-scale pretrained models such as GPT-2 (Radford et al, 2019) and BART (Lewis et al, 2020) on the parallel corpora can improve the performance. To address the data-scarcity problem of parallel datasets, proposed three data augmentation techniques to augment pseudo-parallel data for training.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…Previous studies (Rao and Tetreault, 2018;Niu et al, 2018; typically train seq2seq encoder-decoder models on this benchmark. Recent studies (Wang et al, 2019;Yao and Yu, 2021;Chawla and Yang, 2020;Lai et al, 2021) have deduced that fine-tuning large-scale pretrained models such as GPT-2 (Radford et al, 2019) and BART (Lewis et al, 2020) on the parallel corpora can improve the performance. To address the data-scarcity problem of parallel datasets, proposed three data augmentation techniques to augment pseudo-parallel data for training.…”
Section: Related Workmentioning
confidence: 99%
“…There are some typical informal expressions in the parallel corpus, such as the use of slang words and abbreviations, capitalized words for emphasis, and spelling errors. Some existing studies (Wang et al, 2019;Yao and Yu, 2021) adopt editing rules to revise such informal expressions as a preprocessing step. Inspired by these, we propose the adoption of opposite rules to synthesize such noises.…”
Section: Data Perturbation Strategiesmentioning
confidence: 99%
See 2 more Smart Citations
“…To mitigate the consequent problem of noisy parallel data, the encoder was presented with an input which was a concatenation of the original informal sentence and its formal revision. Yao and Yu (2021) explored a similar architecture. The encoder's input was created by concatenating the original sentence and some supplementary information, which comprised an exhaustive list of all matched rules and the corresponding text alternatives, arranged as tuples.…”
Section: Formalitymentioning
confidence: 99%