2020
DOI: 10.1016/j.csl.2019.101026
|View full text |Cite
|
Sign up to set email alerts
|

State-of-the-art speaker recognition with neural network embeddings in NIST SRE18 and Speakers in the Wild evaluations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
88
0
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
2
2

Relationship

2
5

Authors

Journals

citations
Cited by 97 publications
(89 citation statements)
references
References 7 publications
0
88
0
1
Order By: Relevance
“…The auxiliary network in our DFL formulation is the ResNet-34-LDE network described in [14,15,5]. It is a ResNet-34 residual network with Learnable Dictionary Encoding (LDE) pooling and Angular Softmax loss function.…”
Section: Residual Networkmentioning
confidence: 99%
See 2 more Smart Citations
“…The auxiliary network in our DFL formulation is the ResNet-34-LDE network described in [14,15,5]. It is a ResNet-34 residual network with Learnable Dictionary Encoding (LDE) pooling and Angular Softmax loss function.…”
Section: Residual Networkmentioning
confidence: 99%
“…Total parameters for ETDNN and FTDNN are 10M and 17M respectively. A summary of those networks can be found in [5].…”
Section: X-vector Networkmentioning
confidence: 99%
See 1 more Smart Citation
“…Total parameters for ETDNN and FTDNN are 10M and 17M respectively. More details on the networks and the pipeline can be found in [3,13].…”
Section: X-vector Architecturesmentioning
confidence: 99%
“…One approach to improve the robustness of SV systems is to train them on data created by artificially adding noise to the original training data or simulating the reverberant speech. This method, known as data augmentation, has proven to be effective in improving the performance of SV systems yielding state-of-the-art (SOTA) results on various tasks [2,3]. However, such simulation strategies do not take into account the amount and type of degradation the test utterances can have.…”
Section: Introductionmentioning
confidence: 99%