2006
DOI: 10.1007/11613107_5
|View full text |Cite
|
Sign up to set email alerts
|

MLP Internal Representation as Discriminative Features for Improved Speaker Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
15
0

Year Published

2008
2008
2014
2014

Publication Types

Select...
3
3
2

Relationship

2
6

Authors

Journals

citations
Cited by 17 publications
(16 citation statements)
references
References 9 publications
1
15
0
Order By: Relevance
“…The most representative methods include some normalisation methods at the feature level, such as cepstral mean subtraction (CMS), H-norm, etc. (Wu et al, 2008) and transformation techniques at the feature level, such as LDA-based and NLDA-based feature transformation (Wu et al, 2005a(Wu et al, ,b,c, 2008. Processing at the model level then uses a variety of normalisation and transformation techniques at the model level that is a middle stage of a standard recognition system.…”
Section: Introductionmentioning
confidence: 99%
“…The most representative methods include some normalisation methods at the feature level, such as cepstral mean subtraction (CMS), H-norm, etc. (Wu et al, 2008) and transformation techniques at the feature level, such as LDA-based and NLDA-based feature transformation (Wu et al, 2005a(Wu et al, ,b,c, 2008. Processing at the model level then uses a variety of normalisation and transformation techniques at the model level that is a middle stage of a standard recognition system.…”
Section: Introductionmentioning
confidence: 99%
“…In initial tests with TIMIT in [10] and noisy TINIT in [11 ] we showed that the performance of the features provided by this MLP architecture increases with the number of speakers which the MLP is trained to separate. Due to various factors, test results are subject to significant random fluctuation.…”
Section: Related Researchmentioning
confidence: 99%
“…Instead, we use a subset of speakers, which are most representative of the overall population, and therefore referred to as speaker basis, for MLP training. It is said to be "speaker basis selection" concerning how to select the optimal basis speakers (Wu et al, 2005a(Wu et al, , 2005bMorris et al, 2005). The basic idea for speaker basis selection is based on an approximated Kullback-Leibler (KL) distance between any two speakers, say, S i and S k .…”
Section: Nonlinear Discriminant Analysismentioning
confidence: 99%