2008
DOI: 10.1155/2008/674859
|View full text |Cite
|
Sign up to set email alerts
|

Using SVM as Back-End Classifier for Language Identification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2011
2011
2022
2022

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 11 publications
(5 citation statements)
references
References 6 publications
0
5
0
Order By: Relevance
“…Given an MFCC frame, the mixture component is found which produces the highest likelihood score, and the index of that component becomes the token for that frame (Figure 2). For a stream of input frames, a stream of component indices will be produced, on which language modelling followed by back-end classification can be performed, as is common in audio LID [15], [16]. For the NIST 1996 12 language evaluation task [17], [15] report a minimum error rate of 17%, which is higher than standard PRLM techniques.…”
Section: A Audio Language Identificationmentioning
confidence: 99%
See 1 more Smart Citation
“…Given an MFCC frame, the mixture component is found which produces the highest likelihood score, and the index of that component becomes the token for that frame (Figure 2). For a stream of input frames, a stream of component indices will be produced, on which language modelling followed by back-end classification can be performed, as is common in audio LID [15], [16]. For the NIST 1996 12 language evaluation task [17], [15] report a minimum error rate of 17%, which is higher than standard PRLM techniques.…”
Section: A Audio Language Identificationmentioning
confidence: 99%
“…9. Showing the mean error-rate across six different codebook sizes (8,16,32,64, 128 and 256 codes) for VLID systems trained on visual speech recorded in two different ways. The upper trace is the error when trained on three separate recitals of the UN declaration in English, spoken by the same speaker, and designed to be identical in rendition.…”
Section: Bias Due To Speaking Rate and Recording Conditionsmentioning
confidence: 99%
“…The best-achieved accuracy by the kernel polynomial function is 0.9979712 (99.8 %), which is the same as from the linear model. The next model, RBF (Gaussian) kernel, comes from a family of kernels where a distance measure is smoothed by an exponential function (Suo et al, 2008). RBF is the most used type of kernel function, mainly because it has a localised and finite response along the entire x-axis.…”
Section: Research Resultsmentioning
confidence: 99%
“…Linear kernels are the simplest kernel functions, represented by the product x, y [23]. This kernel is the basic kernel that is most often used by SVM because with this kernel, SVM divides data linearly.…”
Section: Linear Kernelmentioning
confidence: 99%
“…The RBF kernel is a kernel family where distance measurements are smoothed by radials function (exponential function) [23]. The RBF kernel is denoted as in Equation (4),…”
Section: Polynomial Kernelmentioning
confidence: 99%