2023
DOI: 10.1007/978-3-031-26409-2_8
|View full text |Cite
|
Sign up to set email alerts
|

Fooling Partial Dependence via Data Poisoning

Abstract: Many methods have been developed to understand complex predictive models and high expectations are placed on post-hoc model explainability. It turns out that such explanations are not robust nor trustworthy, and they can be fooled. This paper presents techniques for attacking Partial Dependence (plots, profiles, PDP), which are among the most popular methods of explaining any predictive model trained on tabular data. We showcase that PD can be manipulated in an adversarial manner, which is alarming, especially… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
2
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 30 publications
0
2
0
Order By: Relevance
“…The most popularly used toolkits that we can access from this review are DALEX and AIX360. DALEX21 22 is a library used by R Studio. It only supports a few functionalities (ie, local post-hoc and global post-hoc), whereas AIX36012 is a library used by Python.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…The most popularly used toolkits that we can access from this review are DALEX and AIX360. DALEX21 22 is a library used by R Studio. It only supports a few functionalities (ie, local post-hoc and global post-hoc), whereas AIX36012 is a library used by Python.…”
Section: Discussionmentioning
confidence: 99%
“…For this reason, it is necessary to have the skills and equipment to fill the gap from research to practice. To do so, XAI toolkits like AIX360,12 Alibi,14 Skater,15 H2O,16 17 InterpretML,18 19 EthicalML-XAI,19 20 DALEX,21 22 tf-explain,23 Investigate 24. Most interpretations and explanations are post hoc (local interpretable model-agnostic explanations (LIME) and SHapley Additive exPlanations (SHAP).…”
Section: Introductionmentioning
confidence: 99%
“…The ratio of majority to minority classes in the dataset is used to determine the effectiveness of the SDG in generating the target class which is the minority. Finally, the fairness, that is the bias introduced by the ML classifier algorithms as a result of training with synthetic data and bias introduced by real datasets were compared using dalex [30], an package in Python to check fairness. (e.g., malware 324 records and credit card fraud with 284808 records) 6.…”
Section: Related Workmentioning
confidence: 99%