2021
DOI: 10.1038/s41467-021-24025-8
|View full text |Cite
|
Sign up to set email alerts
|

Predicting trends in the quality of state-of-the-art neural networks without access to training or testing data

Abstract: In many applications, one works with neural network models trained by someone else. For such pretrained models, one may not have access to training data or test data. Moreover, one may not know details about the model, e.g., the specifics of the training data, the loss function, the hyperparameter values, etc. Given one or many pretrained models, it is a challenge to say anything about the expected performance or quality of the models. Here, we address this challenge by providing a detailed meta-analysis of hu… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

3
47
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 37 publications
(70 citation statements)
references
References 22 publications
(35 reference statements)
3
47
0
Order By: Relevance
“…Broader picture. Overall, our analysis explains the success of previously-introduced metrics that combine norm information and shape/correlational information [6,7] as well as previouslyobserved peculiarities of norm-based metrics [15]. Our results also highlight the need to go beyond one-size-fits-all metrics based on upper bounds from generalization theory to describe the performance of SOTA NNs, as well as the evaluation of models by expensive retraining on test data.…”
Section: Introductionsupporting
confidence: 65%
See 3 more Smart Citations
“…Broader picture. Overall, our analysis explains the success of previously-introduced metrics that combine norm information and shape/correlational information [6,7] as well as previouslyobserved peculiarities of norm-based metrics [15]. Our results also highlight the need to go beyond one-size-fits-all metrics based on upper bounds from generalization theory to describe the performance of SOTA NNs, as well as the evaluation of models by expensive retraining on test data.…”
Section: Introductionsupporting
confidence: 65%
“…1 The Contest [2,3] made available a data set consisting of pre-trained NN models. This Contest data was more narrow than another corpus analyzed recently [6,7], which used Heavy-Tailed Self-Regularization (HT-SR) theory [8,9,10] to analyze hundreds of state-of-the-art (SOTA) models from CV and natural language processing (NLP). However, the Contest data was more detailed, in the sense there were lower quality models, models in which hyperparameters were suboptimally chosen, etc.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…Furthermore, we question whether a hold-out sample, proportional in size to our overall sample, would have been a better unbiased estimator (how can a sample with a size of around 30 be taken as representative of the whole population?). In the future, we will look to the works of Martin and Corneanu [40,41] that unlock estimating generalization performance directly from the characteristics of the model itself. We are already working on a criterion inspired by their work, which we call the network engagement criterion.…”
Section: Limitations and Future Directionsmentioning
confidence: 99%