2017
DOI: 10.1016/j.csl.2017.01.006
|View full text |Cite
|
Sign up to set email alerts
|

Regularization of neural network model with distance metric learning for i-vector based spoken language identification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
7
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
3
1

Relationship

4
4

Authors

Journals

citations
Cited by 16 publications
(7 citation statements)
references
References 12 publications
0
7
0
Order By: Relevance
“…In this paper, we also take the DCNN as our baseline modeling architecture. As explained in introduction, although the DCNN model for classification is try to learn the input-target mapping function, we can regard the processing as two coupled functions of feature extraction and classifier modeling as we did before [21]. The coupling network and classification score calculation are illustrated in Fig.…”
Section: Deep Convolutional Neural Network For Aedmentioning
confidence: 99%
See 1 more Smart Citation
“…In this paper, we also take the DCNN as our baseline modeling architecture. As explained in introduction, although the DCNN model for classification is try to learn the input-target mapping function, we can regard the processing as two coupled functions of feature extraction and classifier modeling as we did before [21]. The coupling network and classification score calculation are illustrated in Fig.…”
Section: Deep Convolutional Neural Network For Aedmentioning
confidence: 99%
“…For example, in a large category of machine learning, feature learning takes into account of intra-and inter-class pair-wise distance measurements [14,15,16]. In the DL framework, nonlinear distance metric learning has been proposed for different applications [17,18,19,20,21], they all take a similar idea in feature extraction with the pair-wise Siamese network models as originally proposed in [23,24,19]. As a further generalization of the idea based on pair-wise Siamese network for feature extraction, triplet loss was proposed [22].…”
Section: Introductionmentioning
confidence: 99%
“…The i-vectors were 400dimensional vectors that obtained on the full-length duration utterances (Average 7.6s) with the script of Kaldi toolkit [17]. For SVM classifier, we used the radial basis function (RBF) kernel and a grid search with cross-validation following the work [18]. The DNN model were with two hidden layers with 512 neurons for each, and a dropout of 0.3 was applied.…”
Section: Implementation Of Baseline Systemsmentioning
confidence: 99%
“…The generated data were created by adding random uniformly distributed noise over the interval [-1, 1] to the original i-vector data. We also compared the proposed method with conventional methods, i.e., cosine distance, support vector machine (SVM) with linear and radial basis function (RBF) kernels [19]. The optimal model parameters of SVMs were obtained based on a grid search with cross-validation.…”
Section: Implementation Of Baseline Systemsmentioning
confidence: 99%
“…To improve the generalization of the model, regularization methods, such as weight decay [16], dropout [17], data augmentation, have been proposed. For i-vector-based LID tasks, previous works have already investigated DNN with dropout [15] and distance metric learning [18,19], with limited training data.…”
Section: Introductionmentioning
confidence: 99%