2022
DOI: 10.1016/j.neunet.2022.06.026
|View full text |Cite
|
Sign up to set email alerts
|

Compressing speaker extraction model with ultra-low precision quantization and knowledge distillation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 17 publications
0
4
0
Order By: Relevance
“…Impressive results fire up expectations equally high to the quantum world. Data-driven weather and climate predictions apparently beat the best models (Pathak et al, 2022; Bi et al, 2022), and output data can be compressed by three orders of magnitude (Huang and Hoefler 2022). Similar successes are touted in literally any application area.…”
Section: Myth 2: Everything Will Be Deep Learning!mentioning
confidence: 99%
“…Impressive results fire up expectations equally high to the quantum world. Data-driven weather and climate predictions apparently beat the best models (Pathak et al, 2022; Bi et al, 2022), and output data can be compressed by three orders of magnitude (Huang and Hoefler 2022). Similar successes are touted in literally any application area.…”
Section: Myth 2: Everything Will Be Deep Learning!mentioning
confidence: 99%
“…Therefore, researchers have begun exploring methods for transferring knowledge from pretrained models to new models in order to speed up the training process and improve their performance. However, current knowledge transfer methods are typically based on networks of the same size (Yang et al, 2023 ), such as weight sharing, feature transfer, or knowledge transfer from deeper to shallower networks (Huang et al, 2022 ; Shi et al, 2023 ), such as knowledge distillation, and network pruning. There is a lack of knowledge transfer strategies from shallower to deeper networks.…”
Section: Introductionmentioning
confidence: 99%
“…power of scale [7,26]), they can provide warm-start for FL and enable better adaptation to local client distributions. Crucially, while very large PTFs with billions of parameters cannot be deployed in mobile devices, innovations in mobile hardware (equipped with GPU/TPU) [23] and advances in model compression/distillation [19,43,47] will make it possible to deploy smaller, yet equally effective models on clients' devices.…”
Section: Introductionmentioning
confidence: 99%