LatinX in AI at Neural Information Processing Systems Conference 2021 2021
DOI: 10.52591/lxai202112071
|View full text |Cite
|
Sign up to set email alerts
|

Flexible Learning of Sparse Neural Networks via Constrained L0 Regularizations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
10
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(13 citation statements)
references
References 0 publications
0
10
0
Order By: Relevance
“…Ip i∈Ip l(w; x i , y i ), where I p is an index set of local data samples, I p := |I p | is the number of local data samples, I := P p=1 I p is the total number of data samples, l represents a loss function, and x i and y i represent data features and data labels, respectively. We note that most FL literature assumes W := R n , but more recently the importance of imposing hard constraints instead of soft constraints (i.e., penalizing in the objective function) has been discussed in the ML community [15,32].…”
Section: Federated Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…Ip i∈Ip l(w; x i , y i ), where I p is an index set of local data samples, I p := |I p | is the number of local data samples, I := P p=1 I p is the total number of data samples, l represents a loss function, and x i and y i represent data features and data labels, respectively. We note that most FL literature assumes W := R n , but more recently the importance of imposing hard constraints instead of soft constraints (i.e., penalizing in the objective function) has been discussed in the ML community [15,32].…”
Section: Federated Learningmentioning
confidence: 99%
“…Even though locally stored data are not shared with other agents, it is still possible to The aforementioned DP algorithms have been developed for unconstrained models; but, in general, constraints are necessary, for example, in most optimal control problems including distributed control of power flow. Also, imposing hard constraints on ML models (instead of penalizing constraints in the objective function as in [40,10]) is increasingly considered for purposes such as improving accuracy, explaining decisions suggested by ML models, promoting fairness, and observing some physical laws (e.g., [15,17,32,47,50]). This calls for developing DP algorithms suitable for the general constrained optimization model, like the form (1.1).…”
mentioning
confidence: 99%
“…We follow [22,23,34] to formulate the structured pruning task as a regularized learning problem, which aims to learn a sparse model. Let f (•; θ) be a model with parameters θ = {θj } n j=1 , where each θj is a group of parameters (e.g., an attention head) and n is the number of groups.…”
Section: Structured Pruning Using L0 Regularizationmentioning
confidence: 99%
“…( 1) using gradient descent because the gates are discrete. Louizos et al [34] propose a reparameterization trick to make the loss differentiable, which has been widely used in sparse model learning. Here we only introduce their final approach.…”
Section: Structured Pruning Using L0 Regularizationmentioning
confidence: 99%
See 1 more Smart Citation