2022
DOI: 10.1609/aaai.v36i8.20869
|View full text |Cite
|
Sign up to set email alerts
|

BATUDE: Budget-Aware Neural Network Compression Based on Tucker Decomposition

Abstract: Model compression is very important for the efficient deployment of deep neural network (DNN) models on resource-constrained devices. Among various model compression approaches, high-order tensor decomposition is particularly attractive and useful because the decomposed model is very small and fully structured. For this category of approaches, tensor ranks are the most important hyper-parameters that directly determine the architecture and task performance of the compressed DNN models. However, as an NP-hard p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(2 citation statements)
references
References 25 publications
(45 reference statements)
0
2
0
Order By: Relevance
“…Automatic Rank Selection. Our work is most closely related to (Gusak et al 2019;Liebenwein et al 2021;Li et al 2022;Yin et al 2022a), the state-of-the-art automatic rank selection solutions. Specifically, (Gusak et al 2019) proposes to utilize variational Bayesian matrix factorization to determine the ranks of the tensor decomposed DNNs in a multi-stage way.…”
Section: Related Workmentioning
confidence: 96%
“…Automatic Rank Selection. Our work is most closely related to (Gusak et al 2019;Liebenwein et al 2021;Li et al 2022;Yin et al 2022a), the state-of-the-art automatic rank selection solutions. Specifically, (Gusak et al 2019) proposes to utilize variational Bayesian matrix factorization to determine the ranks of the tensor decomposed DNNs in a multi-stage way.…”
Section: Related Workmentioning
confidence: 96%
“…Miao et al [36] recently proposed a budget-aware rank selection method that can calculate tensor ranks via one-shot training. Although it can automatically select the proper tensor ranks for each layer, it may obtain different ranks when the training environment changes, such as the dataset and training hyperparameters.…”
Section: Related Workmentioning
confidence: 99%