2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.01008
|View full text |Cite
|
Sign up to set email alerts
|

B-cos Networks: Alignment is All We Need for Interpretability

Abstract: We present a new direction for increasing the interpretability of deep neural networks (DNNs) by promoting weight-input alignment during training. For this, we propose to replace the linear transformations in DNNs by our novel B-cos transformation. As we show, a sequence (network) of such transformations induces a single linear transformation that faithfully summarises the full model computations. Moreover, the B-cos transformation is designed such that the weights align with relevant signals during optimisati… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 22 publications
(10 citation statements)
references
References 33 publications
0
10
0
Order By: Relevance
“…The last row of the table shows the results for an adapted model of M-ResNet aimed at making it more explainable. Recently, Bohle et al [11] proposed a so-called B-Cos transform which, when interchanged with linear transforms of neural networks, increases the networks' explainability by promoting the alignment of weightinput during training. The alignment pressure on the weights ensures that the model computations align with task-relevant features and therefore become explainable.…”
Section: State-of-the-art Results Reproductionmentioning
confidence: 99%
See 1 more Smart Citation
“…The last row of the table shows the results for an adapted model of M-ResNet aimed at making it more explainable. Recently, Bohle et al [11] proposed a so-called B-Cos transform which, when interchanged with linear transforms of neural networks, increases the networks' explainability by promoting the alignment of weightinput during training. The alignment pressure on the weights ensures that the model computations align with task-relevant features and therefore become explainable.…”
Section: State-of-the-art Results Reproductionmentioning
confidence: 99%
“…In this work, we make use of the iNNvestigate library [1] which implements many explainability methods and provides a common interface to evaluate 19 white-box approaches including Layer-wise Relevance Propagation (LRP) [9] using 12 different rules. In addition, we also evaluate explanations generated by the recently proposed B-Cos network adjustment [11].…”
Section: Explainability Methodsmentioning
confidence: 99%
“…As a bird's view analysis, there are two main distinctions between methods: Interpretable by-design architectures, and Post-Hoc explainability methods. The former searches to create algorithms that directly expose why a decision was made [2,3,5,8,22,35,53]. Our research study is based on the latter.…”
Section: Explainable Artificial Intelligencementioning
confidence: 99%
“…Aligning all used features with human concepts is still difficult, albeit more feasible than without a SLDD-Model. Future work could use a more interpretable feature extractor like B-cos Networks (Böhle et al, 2022) to alleviate that problem. The method could be improved via the feature selection or improvements for small datasets (Reinders et al, 2022).…”
Section: Limitations and Future Workmentioning
confidence: 99%
“…Our proposed method generates the SLDD-Model by utilizing glm-saga (Wong et al, 2021) to compute a sparse linear classifier for selected features, which we then finetune to the sparse structure. We apply feature selection instead of a transformation to reduce the computational load and preserve the original semantics of the features, which can improve interpretability (Tao et al, 2015), especially if a more interpretable model like B-cos Networks (Böhle et al, 2022) is used. Additionally, we propose a novel loss function for more diverse features, which is especially relevant when one class depends on very few features, since using more redundant features limits the total information available for the decision.…”
Section: Introductionmentioning
confidence: 99%