2020
DOI: 10.48550/arxiv.2007.04466
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

URSABench: Comprehensive Benchmarking of Approximate Bayesian Inference Methods for Deep Neural Networks

Abstract: While deep learning methods continue to improve in predictive accuracy on a wide range of application domains, significant issues remain with other aspects of their performance including their ability to quantify uncertainty and their robustness. Recent advances in approximate Bayesian inference hold significant promise for addressing these concerns, but the computational scalability of these methods can be problematic when applied to large-scale models. In this paper, we describe initial work on the developme… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3
1

Relationship

2
2

Authors

Journals

citations
Cited by 4 publications
(9 citation statements)
references
References 12 publications
0
9
0
Order By: Relevance
“…More recently, [12,11,13,14,15] presented large-scale evaluation studies of BDL methods with benchmarking on real-world datasets. These studies motivate data retention and distribution shift as generic protocols for evaluating predictive uncertainty.…”
Section: A Related Workmentioning
confidence: 99%
“…More recently, [12,11,13,14,15] presented large-scale evaluation studies of BDL methods with benchmarking on real-world datasets. These studies motivate data retention and distribution shift as generic protocols for evaluating predictive uncertainty.…”
Section: A Related Workmentioning
confidence: 99%
“…For our paper, we use the SGHMC implementation provided by Vadera et al [35]. We also utilize parts of the PyTorch implementation provided by Frankle and Carbin [10] for our experiments.…”
Section: A2 Implementation Detailsmentioning
confidence: 99%
“…As baselines we use Vanilla ResNet-20 and Deep Ensemble with ResNet-20 and M = 3 members. We used the Axolotl framework 2 and Keras for implementation.…”
Section: A Experimental Setup For Cifar-10mentioning
confidence: 99%
“…There exist many approaches towards uncertainty estimation, however, many of them are complex to train, lack good predictive performance, or are very resource-intensive [2]. One of the best performing approaches is Deep Ensemble [3], where multiple models are trained independently and used at inference, thus making it not viable on mobile platforms with low-latency requirements.…”
Section: Introductionmentioning
confidence: 99%