2022
DOI: 10.1609/aaai.v36i8.20869
|View full text |Cite
|
Sign up to set email alerts
|

BATUDE: Budget-Aware Neural Network Compression Based on Tucker Decomposition

Abstract: Model compression is very important for the efficient deployment of deep neural network (DNN) models on resource-constrained devices. Among various model compression approaches, high-order tensor decomposition is particularly attractive and useful because the decomposed model is very small and fully structured. For this category of approaches, tensor ranks are the most important hyper-parameters that directly determine the architecture and task performance of the compressed DNN models. However, as an NP-hard p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(4 citation statements)
references
References 25 publications
0
4
0
Order By: Relevance
“…And when dealing with high-order tensors, there will be too many auxiliary variables used in the augmented Lagrangian algorithm, which will affect convergence. Without using nuclear norm, [54] just set the upper bound of ranks, therefore it is applicable to various tensor decompositions.…”
Section: Compression-aware Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…And when dealing with high-order tensors, there will be too many auxiliary variables used in the augmented Lagrangian algorithm, which will affect convergence. Without using nuclear norm, [54] just set the upper bound of ranks, therefore it is applicable to various tensor decompositions.…”
Section: Compression-aware Methodsmentioning
confidence: 99%
“…Recently, researches on compression-aware method emerge in large numbers and plenty of experiments show that with the premise of using the same tensor decomposition method, compression-aware method can outperform the other two methods [54,55,85]. Hence, we should pay more attention to it.…”
Section: Sparsity Measurementioning
confidence: 99%
“…Automatic Rank Selection. Our work is most closely related to (Gusak et al 2019;Liebenwein et al 2021;Li et al 2022;Yin et al 2022a), the state-of-the-art automatic rank selection solutions. Specifically, (Gusak et al 2019) proposes to utilize variational Bayesian matrix factorization to determine the ranks of the tensor decomposed DNNs in a multi-stage way.…”
Section: Related Workmentioning
confidence: 96%
“…Miao et al [36] recently proposed a budget-aware rank selection method that can calculate tensor ranks via one-shot training. Although it can automatically select the proper tensor ranks for each layer, it may obtain different ranks when the training environment changes, such as the dataset and training hyperparameters.…”
Section: Related Workmentioning
confidence: 99%