Interspeech 2022 2022
DOI: 10.21437/interspeech.2022-226
|View full text |Cite
|
Sign up to set email alerts
|

Speech Sequence Embeddings using Nearest Neighbors Contrastive Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
1

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(12 citation statements)
references
References 0 publications
0
12
0
Order By: Relevance
“…In this work, we explore the setting where we have untranscribed speech of a target language to continue pretraining the self-supervised model f . In addition, we also explore pooling functions with trainable parameters, such as in [6,7,10]. We follow [9,10,19] and train the pooling function g with a contrastive loss.…”
Section: Task Overviewmentioning
confidence: 99%
See 3 more Smart Citations
“…In this work, we explore the setting where we have untranscribed speech of a target language to continue pretraining the self-supervised model f . In addition, we also explore pooling functions with trainable parameters, such as in [6,7,10]. We follow [9,10,19] and train the pooling function g with a contrastive loss.…”
Section: Task Overviewmentioning
confidence: 99%
“…In addition, we also explore pooling functions with trainable parameters, such as in [6,7,10]. We follow [9,10,19] and train the pooling function g with a contrastive loss. Specifically, we use NTXent [20] which is defined as…”
Section: Task Overviewmentioning
confidence: 99%
See 2 more Smart Citations
“…One way to do this is to employ the method of Algayres, Nabli, Sagot, and Dupoux (2023), who used the contrastive approach for acoustic word embeddings initiated by Livescu and colleagues to train a classifier without labeled training data (Kamper, Jansen, & Goldwater, 2016; Settle & Livescu, 2016). Here, we describe briefly how it works.…”
Section: Introductionmentioning
confidence: 99%