Interspeech 2022 2022
DOI: 10.21437/interspeech.2022-530
|View full text |Cite
|
Sign up to set email alerts
|

Autoregressive Co-Training for Learning Discrete Speech Representation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(7 citation statements)
references
References 0 publications
0
7
0
Order By: Relevance
“…To study the representations learned by our neural HMMs, we evaluate them on phone classification on Wall Street Journal (WSJ) and phone segmentation on TIMIT. We follow the same setting described in prior work [6,13,5], using Lib-riSpeech train-clean-360 for pre-training. Phone classification on WSJ is trained on 90% of si284, and evaluated on dev93.…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…To study the representations learned by our neural HMMs, we evaluate them on phone classification on Wall Street Journal (WSJ) and phone segmentation on TIMIT. We follow the same setting described in prior work [6,13,5], using Lib-riSpeech train-clean-360 for pre-training. Phone classification on WSJ is trained on 90% of si284, and evaluated on dev93.…”
Section: Methodsmentioning
confidence: 99%
“…Phone classification on WSJ is trained on 90% of si284, and evaluated on dev93. Following [5], we also evaluate phone cluster purity on WSJ si284. For phone segmentation on TIMIT, we do not follow the setting in other studies [24,22].…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations