2018 IEEE 59th Annual Symposium on Foundations of Computer Science (FOCS) 2018
DOI: 10.1109/focs.2018.00064
|View full text |Cite
|
Sign up to set email alerts
|

Dispersion for Data-Driven Algorithm Design, Online Learning, and Private Optimization

Abstract: A crucial problem in modern data science is data-driven algorithm design, where the goal is to choose the best algorithm, or algorithm parameters, for a specific application domain. In practice, we often optimize over a parametric algorithm family, searching for parameters with high performance on a collection of typical problem instances. While effective in practice, these procedures generally have not come with provable guarantees. A recent line of work initiated by a seminal paper of Gupta and Roughgarden [… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

4
122
1

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 48 publications
(129 citation statements)
references
References 41 publications
4
122
1
Order By: Relevance
“…As we describe more in Section 3.1.2, this requires us to prove that with high probability, each function sequence from several infinite families of sequences is dispersed. This facet of our analysis is notably different from prior research by Balcan et al [2018a]: in their applications, it is enough to show that with high probability, a single, finite sequence of functions is dispersed. Our proofs thus necessitate that we carefully examine the structure of the utility functions that we analyze.…”
Section: Our Contributionsmentioning
confidence: 73%
See 1 more Smart Citation
“…As we describe more in Section 3.1.2, this requires us to prove that with high probability, each function sequence from several infinite families of sequences is dispersed. This facet of our analysis is notably different from prior research by Balcan et al [2018a]: in their applications, it is enough to show that with high probability, a single, finite sequence of functions is dispersed. Our proofs thus necessitate that we carefully examine the structure of the utility functions that we analyze.…”
Section: Our Contributionsmentioning
confidence: 73%
“…Rather, we must prove that for all type vectors, the dispersion property holds. This facet of our analysis is notably different from prior work by Balcan et al [2018a]: in their applications, it is enough to show that with high probability, a single, finite sequence of functions is dispersed. In contrast, we show that under mild assumptions, with high probability, each function sequence from an infinite family is dispersed.…”
Section: Dispersion and Pseudo-dimension Guaranteesmentioning
confidence: 78%
“…The authors of [37] studied the problem with smooth loss function and proposed using the 2 gradient-norm of a private estimator, i.e., ∇L(w priv , D) 2 , to measure the utility, which was then extended in [34,31] to the cases of non-smooth loss functions and high dimensional space. It is well known that 2 gradient-norm can estimate only the first-order stationary point (or critical point) 3 , and thus may lead to inferior generalization performance [10]. The authors of [28] are the first to show that the utility of general non-convex loss functions can also be measured in the same way as convex loss functions by the expected excess empirical risk.…”
Section: L(w D)mentioning
confidence: 99%
“…Here, we consider a simple private logistic regression model with 2 regularization trained on the Adult dataset [28]. The model is privatized by training with mini-batched projected SGD, then applying a Gaussian perturbation at the output using the method from [49,Algorithm 2] with default parameters 5 . The only hyperparameters tuned in this example are the regularization γ and the noise standard deviation σ, while the rest are fixed 6 .…”
Section: Private Logistic Regressionmentioning
confidence: 99%
“…Recent work on data-driven algorithm configuration has considered the problem of tuning the hyperparameters of combinatorial optimization algorithms while maintaining DP [5]. The setting considered in [5] assumes there is an underlying distribution of problem instances, and a sample from this distribution is used to select hyperparameters that will have good computational performance on future problem instances sampled from the same distribution. In this case, the authors consider a threat model where the whole sample of problem instances used to tune the algorithm needs to be protected.…”
Section: Related Workmentioning
confidence: 99%