2019 IEEE International Symposium on Information Theory (ISIT) 2019
DOI: 10.1109/isit.2019.8849614
|View full text |Cite
|
Sign up to set email alerts
|

Harmless interpolation of noisy data in regression

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

6
117
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 58 publications
(123 citation statements)
references
References 15 publications
6
117
0
Order By: Relevance
“…However, these calculations do not elucidate several crucial statistical phenomena, which are instead the main contribution of our work (see Section 2): optimality of large overparametrization, optimality of interpolators at high SNR ( 3 0 limit), the role of self-induced regularization, and the disappearance of the double descent at optimal overparametrization. Rate-optimal bounds on the generalization error of overparametrized linear models were recently derived in [12] (see also [51] for a different perspective).…”
Section: Learning Via Interpolationmentioning
confidence: 99%
“…However, these calculations do not elucidate several crucial statistical phenomena, which are instead the main contribution of our work (see Section 2): optimality of large overparametrization, optimality of interpolators at high SNR ( 3 0 limit), the role of self-induced regularization, and the disappearance of the double descent at optimal overparametrization. Rate-optimal bounds on the generalization error of overparametrized linear models were recently derived in [12] (see also [51] for a different perspective).…”
Section: Learning Via Interpolationmentioning
confidence: 99%
“…The initial version of this article [5] appeared concurrently with works of Hastie et al [11], Muthukumar et al [15], and Bartlett et al [3], all of which also study the behavior of the least squares/least norm predictor in overparameterized linear regression. Muthukumar et al [15] focus on the well-specified scenario (essentially p = D) and provide upper bounds on the risk that go to zero as p \rightar \infty . (A related variance analysis was carried out by Neal et al [16].)…”
mentioning
confidence: 99%
“…1.2), our scheme applies eigen-weighting matrix Λ * to incentivize the optimizer to place higher weight on promising eigen-directions. This eigen-weighting procedure has been shown in the single-task case to be extremely crucial to avail the benefit of overparameterization [6,30,33]: it captures an inductive bias that promotes certain features and demotes others. We show that the importance of eigen-weighting extends to the multi-task case as well.…”
Section: Our Contributionsmentioning
confidence: 99%
“…Overparameterized ML and double-descent The phenomenon of double-descent was first discovered by [6]. This paper and subsequent works on this topic [4,33,32,30,11] emphasize the importance of the right prior (sometimes referred to as inductive bias or regularization) to avail the benefits of overparameterization. However, an important question that arises is: where does this prior come from?…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation