2021
DOI: 10.21203/rs.3.rs-448572/v1
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Interpretable Machine Learning for Genomics

Abstract: High-throughput technologies such as next generation sequencing allow biologists to observe cell function with unprecedented resolution, but the resulting datasets are too large and complicated for humans to understand without the aid of advanced statistical methods. Machine learning (ML) algorithms, which are designed to automatically find patterns in data, are well suited to this task. Yet these models are often so complex as to be opaque, leaving researchers with few clues about underlying mechanisms. Inter… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 11 publications
(10 citation statements)
references
References 116 publications
(79 reference statements)
0
10
0
Order By: Relevance
“…Thus, SHAP is considered to be the most reliable metric for tree-based ML methods at the moment (Molnar, 2021). Furthermore, the SHAP values have a very intuitive explanation.…”
Section: Shapley Additive Explanations Valuesmentioning
confidence: 99%
“…Thus, SHAP is considered to be the most reliable metric for tree-based ML methods at the moment (Molnar, 2021). Furthermore, the SHAP values have a very intuitive explanation.…”
Section: Shapley Additive Explanations Valuesmentioning
confidence: 99%
“…Alternatively, the model-agnostic tools can be applied to the outputs of any AI algorithm, including very complex an opaque models, in order to provide interpretation on decision drivers for those models. According to Molnar [2021], this set of tools is post-hoc, i.e. they are applied after the model has been trained and they do not require access to its estimates or code, rather it only requires the ability to test the model.…”
Section: Model-specific Vs Model-agnosticmentioning
confidence: 99%
“…Algorithm 1: Permutation Feature Importance algorithm [Molnar, 2021], [Fisher et al, 2019] Initialization: Trained model f , feature matrix X, target vector y, error measure L(y, f ); Estimate the original model error = L(y, f (X)); For each feature: j = 0, 1, . .…”
Section: Feature Importancementioning
confidence: 99%
See 1 more Smart Citation
“…The lack of correlation may suggest that some models make the right prediction for the wrong reason (Kirchner, 2006). However, to fully understand the basis of the predictions, formal machine learning explanation methods for interpreting the model structures are required (Molnar, 2020). These methods include the SHAP method (Lundberg et al, 2017), the LIME method (Ribeiro et al, 2016), or the integrated gradients method (Sundararajan et al, 2017).…”
Section: Correlations Between Prediction Accuracy and Consistency Of ...mentioning
confidence: 99%