2021
DOI: 10.48550/arxiv.2112.04364
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Generalization Error Bounds for Iterative Recovery Algorithms Unfolded as Neural Networks

Abstract: Motivated by the learned iterative soft thresholding algorithm (LISTA), we introduce a general class of neural networks suitable for sparse reconstruction from few linear measurements. By allowing a wide range of degrees of weight-sharing between the layers, we enable a unified analysis for very different neural network types, ranging from recurrent ones to networks more similar to standard feedforward neural networks. Based on training samples, via empirical risk minimization we aim at learning the optimal ne… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

2
6
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(8 citation statements)
references
References 23 publications
(53 reference statements)
2
6
0
Order By: Relevance
“…This result, whose proof relies on a Rademacher complexity analysis, suggests that model-based networks may exhibit better generalization capabilities than traditional neural networks, in line with empirical results [45]. Concretely, this generalization error bound for LISTA-like model-based networks depends on the number of layers only logarithmically, whereas generalization error bounds for traditional neural networks (albeit in classification settings) can scale exponentially in the number of layers [13], [109]. On the other hand, the dependence on m and n in (51) remains fairly strong, and it would be of interest to determine if it can be reduced, e.g., by exploiting sparsity (notice that the sparsity level s is absent in (51)).…”
Section: Theoretical Foundations Of Unfolding Techniquessupporting
confidence: 61%
See 2 more Smart Citations
“…This result, whose proof relies on a Rademacher complexity analysis, suggests that model-based networks may exhibit better generalization capabilities than traditional neural networks, in line with empirical results [45]. Concretely, this generalization error bound for LISTA-like model-based networks depends on the number of layers only logarithmically, whereas generalization error bounds for traditional neural networks (albeit in classification settings) can scale exponentially in the number of layers [13], [109]. On the other hand, the dependence on m and n in (51) remains fairly strong, and it would be of interest to determine if it can be reduced, e.g., by exploiting sparsity (notice that the sparsity level s is absent in (51)).…”
Section: Theoretical Foundations Of Unfolding Techniquessupporting
confidence: 61%
“…Learning-Theoretic Oriented Results: Another class of contributions concentrates on learning-theoretic aspects, studying how the generalization error -corresponding to the difference between the expected error and the empirical error -behaves as a function of various quantities relating to the learning problem, including the number of training samples. Works giving results of this kind include [13], [109], [23], [115].…”
Section: Theoretical Foundations Of Unfolding Techniquesmentioning
confidence: 99%
See 1 more Smart Citation
“…Learning-Theoretic Oriented Results: Another class of contributions concentrates on learning-theoretic aspects, studying how the generalization error -corresponding to the difference between the expected error and the empirical error -behaves as a function of various quantities relating to the learning problem, including the number of training samples. Works giving results of this kind include [14], [117], [26], [124].…”
Section: Theoretical Foundations Of Unfolding Methodsmentioning
confidence: 99%
“…For example, [44,45] provide convergence guarantees regarding variants of learned ISTA networks, while [26] studies the robustness of an unfolding network, produced by a forward-backward proximal interior point method. Moreover, [23,46,47] present generalization 1 error bounds, in terms of the Rademacher complexity of the hypothesis class consisting of the functions that a learnable ISTA network can implement. The concept of generalization error bounds is widely known in the field of statistical learning theory and studied by different -albeit connected -complexity terms, such as Rademacher complexity [48], Vapnik-Chervonenkis (VC) dimension [49], stability [50] and robustness [51].…”
Section: Related Workmentioning
confidence: 99%