2014
DOI: 10.1109/taslp.2014.2308473
|View full text |Cite
|
Sign up to set email alerts
|

On the use of i–vector posterior distributions in Probabilistic Linear Discriminant Analysis

Abstract: The i-vector extraction process is affected by several factors such as the noise level, the acoustic content of the observed features, the channel mismatch between the training conditions and the test data, and the duration of the analyzed speech segment. These factors influence both the i-vector estimate and its uncertainty, represented by the i-vector posterior covariance. This paper presents a new PLDA model that, unlike the standard one, exploits the intrinsic i-vector uncertainty. Since the recognition ac… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
50
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 60 publications
(51 citation statements)
references
References 22 publications
(33 reference statements)
1
50
0
Order By: Relevance
“…Generative and discriminative models are two general approaches for language recognition based on i-vectors. Although the reported results using discriminative methods such as multiclass logistic regression and support vector machines are comparable to those of using generative models [22] such as Gaussian and probabilistic linear discriminant analysis (PLDA) models, the generative models provide an appropriate framework to benefit from the uncertainty in the i-vector extraction process through the posterior covariance matrix of the i-vector [23]. PLDA [24], originally studied in image processing, has been very successful in speaker and language recognition.…”
Section: Plda Modelmentioning
confidence: 99%
See 2 more Smart Citations
“…Generative and discriminative models are two general approaches for language recognition based on i-vectors. Although the reported results using discriminative methods such as multiclass logistic regression and support vector machines are comparable to those of using generative models [22] such as Gaussian and probabilistic linear discriminant analysis (PLDA) models, the generative models provide an appropriate framework to benefit from the uncertainty in the i-vector extraction process through the posterior covariance matrix of the i-vector [23]. PLDA [24], originally studied in image processing, has been very successful in speaker and language recognition.…”
Section: Plda Modelmentioning
confidence: 99%
“…Fig.1 indicates a high correlation (0.98) between the proposed quality measure and utterance duration in the LRE15 database (described in Section 5.1). However, since the i-vector posterior covariance is also influenced by other factors such as background noise, channel type, incomplete transformations and the acoustic content of the utterance [19,23], we expect that the proposed quality measure captures more information about the quality of an utterance than its duration.…”
Section: Proposed I-vector Quality Measurementioning
confidence: 99%
See 1 more Smart Citation
“…Other work based on the aleatoric uncertainty concept has also given rise to uncertainty propagation approaches for speaker recognition. However, these approaches focused on the issue of computing representations with in-sufficient data, caused by utterances with different, possibly short durations [9][10][11][12][13][14][15].…”
Section: Introductionmentioning
confidence: 99%
“…The conventional speaker verification approach entails using i-vectors [3] and probabilistic linear discriminant analysis (PLDA) [2]. As a supervised learning method, i-vector requires sufficient statistics which are computed from a Gaussian Mixture Model-Universal Background Model (GMM-UBM), followed by a PLDA model to produce verification scores [3].…”
Section: Introductionmentioning
confidence: 99%