2021
DOI: 10.1609/aaai.v35i15.17594
|View full text |Cite
|
Sign up to set email alerts
|

Generate Your Counterfactuals: Towards Controlled Counterfactual Generation for Text

Abstract: Machine Learning has seen tremendous growth recently, which has led to a larger adaptation of ML systems for educational assessments, credit risk, healthcare, employment, criminal justice, to name a few. The trustworthiness of ML and NLP systems is a crucial aspect and requires a guarantee that the decisions they make are fair and robust. Aligned with this, we propose a novel framework GYC, to generate a set of exhaustive counterfactual text, which are crucial for testing these ML systems. Our main contributio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
4
2

Relationship

0
10

Authors

Journals

citations
Cited by 29 publications
(16 citation statements)
references
References 35 publications
0
8
0
Order By: Relevance
“…In the long term, we would like to make our model more explainable both to help us understand the strengths and weaknesses of our model and explain results to recruiters and human resources staff to allow them to adapt their needs in recruitment. To do so, we plan to explore counterfactual generation [28,29].…”
Section: Discussionmentioning
confidence: 99%
“…In the long term, we would like to make our model more explainable both to help us understand the strengths and weaknesses of our model and explain results to recruiters and human resources staff to allow them to adapt their needs in recruitment. To do so, we plan to explore counterfactual generation [28,29].…”
Section: Discussionmentioning
confidence: 99%
“…The proposed models by Madaan et al (2021) and Wu et al (2021) both generated counterfactual explanations in a conditional manner using GPT2. Madaan et al (2021) defines named-entity tags, semantic role labels, or sentiments as conditions for the words, while Wu et al (2021) controls the types and locations of perturbations in the text. Both approaches allow more targeted and controlled generation of counterfactual explanations in text.…”
Section: Related Work In Counterfactual Explanationsmentioning
confidence: 99%
“…Another research direction is the use of causal inference for measuring biases in LLMs, for example to analyze if the generated text by an LLM is affected considerably by only changing the protected attributes or categories in the input (Huang et al, 2020;Madaan et al, 2021;Cheng et al, 2021). In line with this idea, Huang et al ( 2020) used a sentiment classifier to quantify and reduce the sentiment bias existent in LLMs.…”
Section: Examples Of Bias Measure Studiesmentioning
confidence: 99%