Interspeech 2022 2022
DOI: 10.21437/interspeech.2022-10584
|View full text |Cite
|
Sign up to set email alerts
|

On-demand compute reduction with stochastic wav2vec 2.0

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

1
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(3 citation statements)
references
References 0 publications
1
2
0
Order By: Relevance
“…Downstream results. First, the reported results in Table 1 are in-line with previous works evaluating partially pre-trained wav2vec 2.0 [18]. Speaker verification EER, for instance, are even better than those reported within the official SUPERB leaderboard with 6.0% against 4.1% in our experiments for the base wav2vec 2.0.…”
Section: Downstream Evaluationsupporting
confidence: 89%
See 1 more Smart Citation
“…Downstream results. First, the reported results in Table 1 are in-line with previous works evaluating partially pre-trained wav2vec 2.0 [18]. Speaker verification EER, for instance, are even better than those reported within the official SUPERB leaderboard with 6.0% against 4.1% in our experiments for the base wav2vec 2.0.…”
Section: Downstream Evaluationsupporting
confidence: 89%
“…SOTA SSL-trained models rely on the same CNN-based AFE combined with a large vanilla Transformer. Careful engineering of these two parts could lead to significant efficiency gains [18] and allow pre-training on mid-tier GPU (e.g., Nvidia Ti 80/90 families) [16].…”
Section: Introductionmentioning
confidence: 99%
“…Furthermore, they demonstrated that using a large amount of synthesized data can lead to better performance than with the small amount of real data used to train the synthesizer. Wu, Kim, Pan, Han, Weinberger and Artzi (2022) and Vyas, Hsu, Auli and Baevski (2022) proposed modifications to the wav2vec 2.0 architecture aimed at reducing the computational cost of pretraining and inference.…”
Section: Related Workmentioning
confidence: 99%