2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2015
DOI: 10.1109/icassp.2015.7178787
|View full text |Cite
|
Sign up to set email alerts
|

Cluster adaptive training for deep neural network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
40
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 64 publications
(40 citation statements)
references
References 21 publications
0
40
0
Order By: Relevance
“…Other subspace methods include cluster adaptive training (CAT) [113], [114] and factorized hidden layer (FHL) [115], [116], where the transformations are confined into the speaker subspace. Similar to the eigenvoice [117] or cluster adaptive training [118] in the Gaussian mixture model era, CAT [113], [114] in DNN training constructs multiple DNNs to form the bases of a canonical parametric space. During adaptation, an interpolation vector which is associated to a target speaker or environment is estimated online to combine the multiple DNN bases into a single adapted DNN.…”
Section: A Acoustic Model Adaptationmentioning
confidence: 99%
See 1 more Smart Citation
“…Other subspace methods include cluster adaptive training (CAT) [113], [114] and factorized hidden layer (FHL) [115], [116], where the transformations are confined into the speaker subspace. Similar to the eigenvoice [117] or cluster adaptive training [118] in the Gaussian mixture model era, CAT [113], [114] in DNN training constructs multiple DNNs to form the bases of a canonical parametric space. During adaptation, an interpolation vector which is associated to a target speaker or environment is estimated online to combine the multiple DNN bases into a single adapted DNN.…”
Section: A Acoustic Model Adaptationmentioning
confidence: 99%
“…An issue in the CAT-style methods is that the bases are fullrank matrices, which require very large amount of training data. Therefore, the number of bases in CAT is usually constrained to few [113], [114]. A solution is to use FHL [115], [116] which constrains the bases to be rank-1 matrices.…”
Section: A Acoustic Model Adaptationmentioning
confidence: 99%
“…During adaptation, an interpolation vector, specific to a particular acoustic condition, is used to combine the multiple sub-networks into a single adapted DNN. We refer to this factorized DNN training as cluster adaptive training (CAT) following [11]. CAT was initially proposed for GMM-HMM acoustic models [24], and later extended to DNN by introducing multiple canonical weight matrices for a DNN layer as depicted in Fig.…”
Section: Cluster Adaptive Trainingmentioning
confidence: 99%
“…In [11], [12] and [13], multiple weight matrices or sub-networks are constructed to form the bases of a canonical parametric space. During adaptation, an interpolation vector, specific to a particular acoustic condition, is used to combine the multiple sub-networks into a single adapted DNN.…”
Section: Cluster Adaptive Trainingmentioning
confidence: 99%
See 1 more Smart Citation