2020 25th International Conference on Pattern Recognition (ICPR) 2021
DOI: 10.1109/icpr48806.2021.9412162
|View full text |Cite
|
Sign up to set email alerts
|

Learning sparse deep neural networks using efficient structured projections on convex constraints for green AI

Abstract: Deep neural networks (DNN) have been applied recently to different domains and perform better than classical state-of-the-art methods. However the high level of performances of DNNs is most often obtained with networks containing millions of parameters and for which training requires substantial computational power. To deal with this computational issue proximal regularization methods have been proposed in the literature but they are time consuming. In this paper, we propose instead a constrained approach. We … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
12
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
3
2

Relationship

3
5

Authors

Journals

citations
Cited by 13 publications
(12 citation statements)
references
References 29 publications
(33 reference statements)
0
12
0
Order By: Relevance
“…Total Loss Following the work developed by [34], which proposed a double descent algorithm, we replaced the thresholding by our ℓ 1,1 projection and devised a new double descent algorithm (See Barlaud and Guyard [35]) as follows :…”
Section: Reconstruction Loss λmentioning
confidence: 99%
“…Total Loss Following the work developed by [34], which proposed a double descent algorithm, we replaced the thresholding by our ℓ 1,1 projection and devised a new double descent algorithm (See Barlaud and Guyard [35]) as follows :…”
Section: Reconstruction Loss λmentioning
confidence: 99%
“…The goal is to compute the network weights W minimizing the total loss which includes both the classification loss and the reconstruction loss. To perform feature selection, as large datasets often present a relatively small number of informative features, we also want to sparsify the network, following the work proposed in [34]. Thus, instead of the classical computationally expensive lagrangian regularization approach [35], we propose to minimize the following constrained approach [36]: We use the Cross Entropy (CE) Loss for the classification loss ℋ.…”
Section: Methodsmentioning
confidence: 99%
“…The goal is to compute the network weights W minimizing the total loss which includes both the classification loss and the reconstruction loss. To perform feature selection, as large datasets often present a relatively small number of informative features, we also want to sparsify the network, following the work proposed in [34].…”
Section: Ssae Criterionmentioning
confidence: 99%
“…We compute the ℓ 1,1 constraint with the following algorithm: we first compute the radius t i and then project the rows using the ℓ 1 adaptive constraint t i [56]. We use a double descent algorithm with ℓ 1,1 projection [44] as follows:…”
Section: Supervised Autoencoder Neural Network Algorithmmentioning
confidence: 99%