2022
DOI: 10.1016/j.patcog.2021.108194
|View full text |Cite
|
Sign up to set email alerts
|

Towards robust explanations for deep neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
38
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 42 publications
(63 citation statements)
references
References 16 publications
0
38
0
Order By: Relevance
“…Instead of the current approach to debias models with data balancing, our debias approach can retrain models to de-emphasize focusing on sensitive concepts (e.g., faces). However, we caution about the dark pattern of debiasing explanations to make an unfair model appear fair by retraining its explanation to appear fair (e.g., [24,25]).…”
Section: Debiasing Explanations Against Social Biasmentioning
confidence: 99%
“…Instead of the current approach to debias models with data balancing, our debias approach can retrain models to de-emphasize focusing on sensitive concepts (e.g., faces). However, we caution about the dark pattern of debiasing explanations to make an unfair model appear fair by retraining its explanation to appear fair (e.g., [24,25]).…”
Section: Debiasing Explanations Against Social Biasmentioning
confidence: 99%
“…In the context of feature importance methods, [3,11] have proposed approaches to make gradient-based methods for DNNs significantly more robust. Anders et al [3] took inspiration from the field of manifold learning and proposed to project explanations along tangential directions of the data manifold.…”
Section: Robust Explanationsmentioning
confidence: 99%
“…Anders et al [3] took inspiration from the field of manifold learning and proposed to project explanations along tangential directions of the data manifold. Dombrowski et al [11] proposed three ways to improve the robustness of DNN explanations -(1) by training DNNs with weight decay; (2) by training using smoothed activation functions; and (3) by adding a regulariser for model's curvature in the training process. Similarly, Lakkaraju Table 1: A summary of robustness analysis scenarios for two types of post-hoc local explainability methods -feature importance and counterfactuals.…”
Section: Robust Explanationsmentioning
confidence: 99%
“…We consider such dependencies in stage 1, where the constraints in Eq. ( 5) encourage the connectivities, (10) where M j ∈ [0, 1] is the mask for the j-th edge. The constraints indicate that the selection of the j-th edge can lead to the selection of the k-th edge if they share a node [29], and controls the co-occurrence of the two edges.…”
Section: Optimization Problems For Graphsmentioning
confidence: 99%
“…Robustness in explanations is gaining attention [10], [22], [46]. In [10], the goal is to train neural networks for image classification that has robust explanations with malicious data manipulations.…”
Section: H Reproducibility Checklistmentioning
confidence: 99%