2022
DOI: 10.48550/arxiv.2201.04315
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On the Statistical Complexity of Sample Amplification

Abstract: Given n i.i.d. samples drawn from an unknown distribution P , when is it possible to produce a larger set of n+m samples which cannot be distinguished from n+m i.i.d. samples drawn from P ? Axelrod et al. [AGSV20] formalized this question as the sample amplification problem, and gave optimal amplification procedures for discrete distributions and Gaussian location models. However, these procedures and associated lower bounds are tailored to the specific distribution classes, and a general statistical understan… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 12 publications
(14 reference statements)
0
1
0
Order By: Relevance
“…A larger generated sample than the training dataset will then include successively less information per event than the training data, and eventually the information in the generated events will saturate and be dominated by limitations from the network architecture and training. With this pattern in mind [1], we can define an amplification or GANplification factor [2,3] in terms of an effective sample size for a given surrogate model.…”
Section: Introductionmentioning
confidence: 99%
“…A larger generated sample than the training dataset will then include successively less information per event than the training data, and eventually the information in the generated events will saturate and be dominated by limitations from the network architecture and training. With this pattern in mind [1], we can define an amplification or GANplification factor [2,3] in terms of an effective sample size for a given surrogate model.…”
Section: Introductionmentioning
confidence: 99%