2020
DOI: 10.48550/arxiv.2012.04698
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Generate Your Counterfactuals: Towards Controlled Counterfactual Generation for Text

Abstract: Machine Learning has seen tremendous growth recently, which has led to a larger adoption of ML systems for educational assessments, credit risk, healthcare, employment, criminal justice, to name a few. Trustworthiness of ML and NLP systems is a crucial aspect and requires guarantee that the decisions they make are fair and robust. Aligned with this, we propose a framework GYC, to generate a set of counterfactual text samples, which are crucial for testing these ML systems. Our main contributions include a) We … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
11
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 9 publications
(12 citation statements)
references
References 27 publications
(16 reference statements)
0
11
0
Order By: Relevance
“…Besides the aforementioned paraphrasing and style transfer, prior works have also successfully generated contrastive examples that are useful for model training, evaluation, and explanation. They usually rely on application-specific class labels (Ross et al, 2020;Madaan et al, 2020b;Sha et al, 2021;Akyürek et al, 2020) or heuristic perturbation strategies that needs to be expressed through pairs of original and perturbed sentences (Wu et al, 2021), which are expensive to generalize. Recently, Huang and Chang (2021) designed SynPG, a paraphraser that can mimic parse tree structures learned from non-paired sentences.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Besides the aforementioned paraphrasing and style transfer, prior works have also successfully generated contrastive examples that are useful for model training, evaluation, and explanation. They usually rely on application-specific class labels (Ross et al, 2020;Madaan et al, 2020b;Sha et al, 2021;Akyürek et al, 2020) or heuristic perturbation strategies that needs to be expressed through pairs of original and perturbed sentences (Wu et al, 2021), which are expensive to generalize. Recently, Huang and Chang (2021) designed SynPG, a paraphraser that can mimic parse tree structures learned from non-paired sentences.…”
Section: Related Workmentioning
confidence: 99%
“…Controllable text generation through semantic perturbations, which modifies sentences to match certain target attributes, has been widely applied to a variety of tasks, e.g., changing sentence styles (Reid and Zhong, 2021), mitigating dataset biases (Gardner et al, 2021), explaining model behaviors (Ross et al, 2020), and improving model generalization (Teney et al, 2020;Wu et al, 2021). Existing work trains controlled generators with task-specific data, e.g., training a style transferer requires instances labeled with positive and negative sentiments (Madaan et al, 2020b). As a result, * denotes equal contribution.…”
Section: Introductionmentioning
confidence: 99%
“…where an explicit protected attribute is often not present. In these domains, counterfactual augmentation is generally a manual process; recent work provides support [31,42] but not complete automation. Prediction sensitivity (defined in Section 3) can be viewed as a way of measuring counterfactual fairness.…”
Section: Background and Related Workmentioning
confidence: 99%
“…However, since they are trained using negative samples obtained from random contexts, they are also prone to the spurious pattern of content similarity. Adversarial or counterfactual data creation techniques have been proposed for applications such as evaluation (Gardner et al, 2020;Madaan et al, 2020), attacks (Ebrahimi et al, 2018;Wallace et al, 2019;Jin et al, 2020), explanations (Goodwin et al, 2020;Ross et al, 2020) or training models to be robust against spurious patterns and biases (Garg et al, 2019;Huang et al, 2020). Adversarial examples are crafted through operations such as adding noisy characters (Ebrahimi et al, 2018;Pruthi et al, 2019), paraphrasing (Iyyer et al, 2018), replacing with synonyms (Alzantot et al, 2018;Jin et al, 2020), rule based token-level transformations (Kryscinski et al, 2020), or inserting words relevant to the context (Zhang et al, 2019).…”
Section: Related Workmentioning
confidence: 99%