2020
DOI: 10.1101/2020.09.08.288068
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Global Importance Analysis: An Interpretability Method to Quantify Importance of Genomic Features in Deep Neural Networks

Abstract: Deep neural networks have demonstrated improved performance at predicting the sequence specificities of DNA- and RNA-binding proteins compared to previous methods that rely on k-mers and position weight matrices. For model interpretability, attribution methods have been employed to reveal learned patterns that resemble sequence motifs. First-order attribution methods only quantify the independent importance of single nucleotide variants in a given sequence – it does not provide the effect size of motifs (or th… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
25
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 17 publications
(25 citation statements)
references
References 68 publications
0
25
0
Order By: Relevance
“…This in combination with the large number overlap of Nrf1 and Tcf12 binding sites suggests Nrf1 and Tcf12 binding may be regulated by logic beyond the presence or absence of their DNA binding motifs. We also show that under class imbalanced data, EPE and DEPE generate robust estimates of transcription factor effects in contrast to Global Importance Analysis (15), which takes a related approach but uses a difference instead of a ratio to estimate the effect of a pattern (Fig S2).…”
Section: Resultsmentioning
confidence: 94%
See 3 more Smart Citations
“…This in combination with the large number overlap of Nrf1 and Tcf12 binding sites suggests Nrf1 and Tcf12 binding may be regulated by logic beyond the presence or absence of their DNA binding motifs. We also show that under class imbalanced data, EPE and DEPE generate robust estimates of transcription factor effects in contrast to Global Importance Analysis (15), which takes a related approach but uses a difference instead of a ratio to estimate the effect of a pattern (Fig S2).…”
Section: Resultsmentioning
confidence: 94%
“…In these cases, explicitly training models on each differential comparison between cell types would be computational expensive and time intensive. We also note that Expected Pattern Effect resembles other methods for extracting pattern effects from deep neural networks (3,15), except that we compute Expected Pattern Effect to permit the comparison of pattern effects between conditions, which is important for identification of cell type-specific or condition-specific sequence features and show that our use of ratio to compare effects is more robust to analysis of cell type-specific transcription factor activity under class imbalance.…”
Section: Discussionmentioning
confidence: 96%
See 2 more Smart Citations
“…More recently, a new wave of algorithms have been introduced that use deep neural networks to predict RBP binding sites ( Alipanahi et al , 2015 ; Ghanbari and Ohler, 2020 ; Grønning et al , 2020 ; Pan and Shen, 2018 ; Yan and Zhu, 2020 ). One challenge is to explain what these complex models have learned, although recently a multitude of methods for interpreting the learned models have been developed, for instance, based on in silico mutagenesis, predictions on synthetic sequences, gradient tracing and analyzing the convolutional filters ( Alipanahi et al , 2015 ; Ghanbari and Ohler, 2020 ; Koo et al , 2020 ; Pan and Shen, 2018 ). However, with the increasing number of model parameters and network complexity, the risk grows that such models could also learn experimental biases in the datasets.…”
Section: Introductionmentioning
confidence: 99%