2021
DOI: 10.48550/arxiv.2111.15603
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Human Imperceptible Attacks and Applications to Improve Fairness

Abstract: Modern neural networks are able to perform at least as well as humans in numerous tasks involving object classification and image generation. However, small perturbations which are imperceptible to humans may significantly degrade the performance of well-trained deep neural networks. We provide a Distributionally Robust Optimization (DRO) framework which integrates human-based image quality assessment methods to design optimal attacks that are imperceptible to humans but significantly damaging to deep neural n… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 61 publications
0
1
0
Order By: Relevance
“…Particularly, we define and compute Fairness Influence Function (FIF) that quantifies the contribution of individual and subset of features to the resulting bias. FIFs do not only allow practitioners to identify the features to act up on but also to quantify the effect of various affirmative [8,19,23,[45][46][47][48] or punitive actions [21,32,42] on the resulting bias.…”
Section: Introductionmentioning
confidence: 99%
“…Particularly, we define and compute Fairness Influence Function (FIF) that quantifies the contribution of individual and subset of features to the resulting bias. FIFs do not only allow practitioners to identify the features to act up on but also to quantify the effect of various affirmative [8,19,23,[45][46][47][48] or punitive actions [21,32,42] on the resulting bias.…”
Section: Introductionmentioning
confidence: 99%