2009 IEEE International Conference on Acoustics, Speech and Signal Processing 2009
DOI: 10.1109/icassp.2009.4960555
|View full text |Cite
|
Sign up to set email alerts
|

The I4U system in NIST 2008 speaker recognition evaluation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2010
2010
2017
2017

Publication Types

Select...
4
2
1

Relationship

2
5

Authors

Journals

citations
Cited by 23 publications
(12 citation statements)
references
References 15 publications
0
12
0
Order By: Relevance
“…However, there are some common practice that we can follow. To facilitate discussion, we present here the results of the latest NIST 2008 speaker recognition evaluation submission by the I4U consortium [138]. All the classifiers of I4U used short-term spectral features and the focus was in the supervectors classifiers.…”
Section: Summary: Which Supervector Methods To Use?mentioning
confidence: 99%
See 1 more Smart Citation
“…However, there are some common practice that we can follow. To facilitate discussion, we present here the results of the latest NIST 2008 speaker recognition evaluation submission by the I4U consortium [138]. All the classifiers of I4U used short-term spectral features and the focus was in the supervectors classifiers.…”
Section: Summary: Which Supervector Methods To Use?mentioning
confidence: 99%
“…They should be augmented with nuisance attribute projection (NAP) [28] and test normalization (T-norm) [14]. Table 1: Performance of individual classifiers and their fusion of I4U system on I4U's telephone quality development dataset [138]. UNC = Uncompensated, EIG = Eigenchannel, JFA = Joint factor analysis, GLDS = Generalized linear discriminant sequence, GSV = Gaussian supetvector, FT = Feature transformation, PSK = Probabilistic sequence kernel, BK = Bhattacharyya kernel.…”
Section: Summary: Which Supervector Methods To Use?mentioning
confidence: 99%
“…The size of the matrices becomes enormous when more sessions are available for each speaker in the development data. This is typically the case for speaker recognition where the number of utterances per speaker is usually in the range from ten to over a hundred [38,39]. In the following, we estimate the parameters μ; F; G; Σ f gof the PLDA model using the expectation maximization (EM) algorithm.…”
Section: Probabilistic Ldamentioning
confidence: 99%
“…In (38), K = log|M 2 |/2 − log|M 1 | is constant for the given set of parameters F; G; Σ f g . Though K diminishes when score normalization is applied, we could calculate the two log-determinant terms easily by using the property of eigenvalue decomposition.…”
Section: Plda Verification Scorementioning
confidence: 99%
“…It is generally agreed upon that the integration of different discriminative cues can improve the performance of language recognition [16]- [18]. In the IIR's submission to the 2009 NIST LRE [26], 7 language classifiers were developed for the language recognition, as follows:…”
Section: Description Of the Sub-systemsmentioning
confidence: 99%