2021
DOI: 10.21203/rs.3.pex-1637/v1
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Deep-insight visible neural network (DI-VNN) for improving interpretability of a non-image deep learning model by data-driven ontology

Abstract: We aimed to provide a framework that organizes internal properties of a convolutional neural network (CNN) model using non-image data to be interpretable by human. The interface was represented as ontology map and network respectively by dimensional reduction and hierarchical clustering techniques. The applicability is to implement a prediction model either to classify categorical or to estimate numerical outcome, including but not limited to that using data from electronic health records. This pipeline harnes… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
14
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(14 citation statements)
references
References 21 publications
0
14
0
Order By: Relevance
“…It is because a more models in comparison would be more vulnerable to a multiple-testing effect relative to the number of datasets, i.e., the best model is found simply by chance [ 63 ]. To avoid such comparison, we considered three criteria for choosing algorithms in developing the models: (1) those commonly used in clinical prediction studies, i.e., logistic regression [ 58 ], which expects a linear predictor-outcome correlation; (2) those which commonly outperformed others (177 algorithms) across 121 datasets [ 64 ], which allow a non-linear predictor-outcome correlation; and (3) our proposed neural-network algorithm [ 65 ], which pursues moderate predictive performance and deeper interpretability. A sufficient sample size was also considered according to the PROBAST guidelines since a small sample size was vulnerable to overfitting [ 58 ].…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…It is because a more models in comparison would be more vulnerable to a multiple-testing effect relative to the number of datasets, i.e., the best model is found simply by chance [ 63 ]. To avoid such comparison, we considered three criteria for choosing algorithms in developing the models: (1) those commonly used in clinical prediction studies, i.e., logistic regression [ 58 ], which expects a linear predictor-outcome correlation; (2) those which commonly outperformed others (177 algorithms) across 121 datasets [ 64 ], which allow a non-linear predictor-outcome correlation; and (3) our proposed neural-network algorithm [ 65 ], which pursues moderate predictive performance and deeper interpretability. A sufficient sample size was also considered according to the PROBAST guidelines since a small sample size was vulnerable to overfitting [ 58 ].…”
Section: Methodsmentioning
confidence: 99%
“…In addition, we used a random-search method to tune values for the pre-defined hyperparameters. We also used those which were defined before conducting this study in a pre-registered protocol [ 65 ]. The randomness and pre-registration were deliberate to avoid a research bias, so-called “hypothesizing after the results are known (HARking)” [ 67 ].…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations