2009
DOI: 10.1016/j.specom.2008.07.003
|View full text |Cite
|
Sign up to set email alerts
|

Broad phonetic classification using discriminative Bayesian networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
13
0

Year Published

2009
2009
2017
2017

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 14 publications
(13 citation statements)
references
References 54 publications
(54 reference statements)
0
13
0
Order By: Relevance
“…We would seek to further improve the results by incorporating techniques used by other authors, in particular the use of commitee classifiers to combined a number of representations with different parameters. Additionally a hierarchical classification could be implemented to reduce broad phoneme class confusions [5,19,20]. There is also scope for further tuning within the method presented by weighting the sector sum and frame average, or allowing the number of frames to be different for each sector.…”
Section: Conclusion and Discussionmentioning
confidence: 99%
“…We would seek to further improve the results by incorporating techniques used by other authors, in particular the use of commitee classifiers to combined a number of representations with different parameters. Additionally a hierarchical classification could be implemented to reduce broad phoneme class confusions [5,19,20]. There is also scope for further tuning within the method presented by weighting the sector sum and frame average, or allowing the number of frames to be different for each sector.…”
Section: Conclusion and Discussionmentioning
confidence: 99%
“…For building AF classifiers, different statistical classifiers have been used, such as artificial neural networks (ANNs) (e.g., King and Taylor 2000;Kirchhoff 1998;Scharenborg et al 2007;Siniscalchi et al 2008) and FFNNs (Manjunath and Sreenivasa Rao 2016), HMMs (e.g., Kirchhoff 1999;Manjunath and Sreenivasa Rao 2016), linear dynamic models (Frankel 2003), k nearest neighbor (k-NNs) (Naess et al 2011) and dynamic Bayesian Networks (e.g., Pernkopf et al 2009;Frankel et al 2007b;Jyothi 2013). Niyogi et al (1999), Pruthi and Espy-Wilson (2007), Yoon et al (2010), Scharenborg et al (2007), and Schutte and Glass (2005) used support vector machines (SVMs) for the classification of articulatory-acoustic features, among other reasons because these show good generalization from a small amount of high-dimensional training data.…”
Section: Support Vector Machines (Svms)mentioning
confidence: 99%
“…This mapping is quite plausible, since the difference between a release realized with friction and a release realized as burst and friction is only salient within the first milliseconds of the release, i.e., the steepness of its amplitude rise. In order to distinguish friction from fricatives and releases from plosives that might be realized with a burst, plosives are mapped on to the sequence 'closure' and 'burst+release' (Frankel et al 2007b;Pernkopf et al 2009). We also opt for the latter sequence of values, for two reasons.…”
Section: Articulatory-acoustic Feature Valuesmentioning
confidence: 99%
See 1 more Smart Citation
“…This should facilitate the process of matching these phonemes to their occurrence in the transcript. The approach of combining ZCR and wavelet analysis has been used to classify segments of speech signals into broad phonetic categories including silence, voiced, unvoiced, and plosive release [4]. The results of the CWT were more straightforward and easier to interpret than the spectrogram results.…”
Section: Zero Crossing Rate and Frequency Analysismentioning
confidence: 99%