2021
DOI: 10.48550/arxiv.2106.07410
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Model Explainability in Deep Learning Based Natural Language Processing

Abstract: Machine learning (ML) model explainability has received growing attention, especially in the area related to model risk and regulations. In this paper, we reviewed and compared some popular ML model explainability methodologies, especially those related to Natural Language Processing (NLP) models. We then applied one of the NLP explainability methods Layer-wise Relevance Propagation (LRP) to a NLP classification model. We used the LRP method to derive a relevance score for each word in an instance, which is a … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
2
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 17 publications
0
2
0
Order By: Relevance
“…In this study, a bibliographic search was conducted for publications that employed ML techniques (excluding DL) for the classification of binary data (interictal and ictal) in humans with epilepsy. It was chosen to use ML techniques because they are more feasible for performing model explainability analysis and observing the importance of features [ 60 ]. DL models, being “black-box”, make it difficult to understand predictions and are therefore less interpretable [ 61 ].…”
Section: Related Workmentioning
confidence: 99%
“…In this study, a bibliographic search was conducted for publications that employed ML techniques (excluding DL) for the classification of binary data (interictal and ictal) in humans with epilepsy. It was chosen to use ML techniques because they are more feasible for performing model explainability analysis and observing the importance of features [ 60 ]. DL models, being “black-box”, make it difficult to understand predictions and are therefore less interpretable [ 61 ].…”
Section: Related Workmentioning
confidence: 99%
“…Figure 6: Diagram of the perturbation based method (left, the figure is brought from[61]) and LRP (right, the figure is brought from[64]). Left: At each step, the least important word is removed by the saliency score using the leave-one-out technique.…”
mentioning
confidence: 99%