“…However, prompt-based augmentation strategies are uncontrolled forms of generation, which may result in generation mistakes for labeled datasets (Sahu et al, 2022;Chen et al, 2022;Meng et al, 2022). In contrast, other recent studies have instead proposed language augmentation strategies that use complex, highly-controlled frameworks that often involve fine-tuning generators (Papangelis et al, 2021;Kulhánek et al, 2021;. Such complex augmentation frameworks require larger amounts of seed data to maintain a ground-truth language distribution (Rosenbaum et al, 2022b;Kim et al, 2021b), and are more costly than prompting PLMs (Chen et al, 2022).…”