2021
DOI: 10.1109/tnnls.2020.3007259
|View full text |Cite
|
Sign up to set email alerts
|

Enhancing Explainability of Neural Networks Through Architecture Constraints

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
18
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 69 publications
(21 citation statements)
references
References 21 publications
0
18
0
Order By: Relevance
“…Whereas SHAP and LIME seek to explain complex models using a regression-like paradigm (i.e., a linear additive function), Explainable Neural Networks (XNNs) [144] use a more general formulation based on an "additive index model" [127]. Here, the algorithm seeks to return a function that describes how model predictions vary with changes to individual parameters (or, more recently, pairs of parameters [148]). As in LIME and SHAP, these models can help data scientists with the appropriate training to understand how changing a specific feature might change the model's prediction, albeit at the risk of inferring spurious correlations.…”
Section: Local Feature Importancementioning
confidence: 99%
“…Whereas SHAP and LIME seek to explain complex models using a regression-like paradigm (i.e., a linear additive function), Explainable Neural Networks (XNNs) [144] use a more general formulation based on an "additive index model" [127]. Here, the algorithm seeks to return a function that describes how model predictions vary with changes to individual parameters (or, more recently, pairs of parameters [148]). As in LIME and SHAP, these models can help data scientists with the appropriate training to understand how changing a specific feature might change the model's prediction, albeit at the risk of inferring spurious correlations.…”
Section: Local Feature Importancementioning
confidence: 99%
“…By contrast, other researches have tried to improve interpretability by changing the structure of the neural networks (NN). Yang et al [ 22 ] proposed the use of an explainable NN (xNN) subject to interpretability constraints in terms of the additivity, sparsity, orthogonality, and smoothness. A complex function is decomposed into sparse additive subnetworks and the projection indexes are forced to be mutually orthogonal such that the resulting subnetworks tend to be less confounded with each other.…”
Section: Related Studiesmentioning
confidence: 99%
“…Studies on the development and testing of an IML learning model have been conducted to improve transparency while maintaining a high-level learning ability by modifying the existing machine learning technologies or developing new ones. The technical approach for IML can be divided into the following: (a) explaining a decision of the learning model (ELM) [ 10 , 11 , 12 , 13 , 14 ] and (b) interpreting the learning model (ILM) [ 15 , 16 , 17 , 18 , 19 , 20 , 21 , 22 , 23 ].…”
Section: Introductionmentioning
confidence: 99%
“…Neural Interaction Transparency (NIT) (Tsang et al, 2018) is a framework that produces the same model as GAMI-Net, but by disentangling interactions within a FFNN. The Explainable Neural Network (xNN) (Vaughan et al, 2018), Adaptive xNN (AxNN) (Chen et al, 2020) and Enhanced xNN (ExNN) (Yang et al, 2020a) are all based on the Generalised Additive Index Model (GAIM):…”
Section: Related Workmentioning
confidence: 99%