2022
DOI: 10.48550/arxiv.2202.00834
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Nonlinear Initialization Methods for Low-Rank Neural Networks

Abstract: We study algorithms for learning low-rank neural networks -networks where the weight parameters are re-parameterized by products of two low-rank matrices. First, we present a provably efficient algorithm which learns an optimal low-rank approximation to a single-hidden-layer ReLU network up to additive error with probability ≥ 1 − δ, given access to noiseless samples with Gaussian marginals in polynomial time and samples. Thus, we provide the first example of an algorithm which can efficiently learn a neural n… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 37 publications
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?