2011 IEEE Workshop on Automatic Speech Recognition &Amp; Understanding 2011
DOI: 10.1109/asru.2011.6163915
|View full text |Cite
|
Sign up to set email alerts
|

Improving reverberant VTS for hands-free robust speech recognition

Abstract: Abstract-Model-based approaches to handling additive background noise and channel distortion, such as Vector Taylor Series (VTS), have been intensively studied and extended in a number of ways. In previous work, VTS has been extended to handle both reverberant and background noise, yielding the Reverberant VTS (RVTS) scheme. In this work, rather than assuming the observation vector is generated by the reverberation of a sequence of background noise corrupted speech vectors, as in RVTS, the observation vector i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2012
2012
2016
2016

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(7 citation statements)
references
References 16 publications
0
7
0
Order By: Relevance
“…The convolutive nature of reverberation induces a long-term correlation between a current observation and past observations of reverberant speech. This longterm correlation has been exploited to mitigate the effect of reverberation directly on the speech signal (i.e., speech [9][10][11][12] or feature [13,14] dereverberation) or on the acoustic model used for recognition [15,16]. The REVERB challenge [17] was organized to evaluate recent progress in the field of reverberant speech enhancement (SE) and recognition.…”
Section: Y(t) = H(t) * S(t) + N(t)mentioning
confidence: 99%
“…The convolutive nature of reverberation induces a long-term correlation between a current observation and past observations of reverberant speech. This longterm correlation has been exploited to mitigate the effect of reverberation directly on the speech signal (i.e., speech [9][10][11][12] or feature [13,14] dereverberation) or on the acoustic model used for recognition [15,16]. The REVERB challenge [17] was organized to evaluate recent progress in the field of reverberant speech enhancement (SE) and recognition.…”
Section: Y(t) = H(t) * S(t) + N(t)mentioning
confidence: 99%
“…7 Bayesian network representation of reverberant VTS (Subsection 6.6.1) a before and b after approximation via an extended observation vector. The figure is based on [15] performs an online adaptation based on the best partial path [15]. It should be pointed out here that there is a variety of other approximations to the statistics of the logsum of (mixtures of ) Gaussian random variables (as seen in Subsections 4.2, 4.5, 6.1, 6.2, 6.6.1), ranging from different PMC methods [10] to maximum [64], piecewise linear [65], and other analytical approximations [66][67][68][69][70].…”
Section: Convolutive Model Adaptationmentioning
confidence: 99%
“…In this article, we review and examine for several uncertainty decoding [1][2][3][4][5], missing feature [6][7][8][9], and model adaptation techniques [10][11][12][13][14][15][16][17][18][19] how their compensation rules can be formulated as an approximated or modified Bayesian decoding rule. In order to illustrate the formalism, we also present the corresponding Bayesian network representations.…”
Section: Introductionmentioning
confidence: 99%
“…Assuming the speech and the noise LMPSCs to be mutually independent, the predictive PDF may be factorized as (8) Thus, the first required model is an a priori model to stochastically describe the trajectory of the LMPSC vectors of the clean speech signal and the noise signal, respectively.…”
Section: Overview Of Feature Enhancementmentioning
confidence: 99%
“…The approach proposed by Wang and Gales [8], [9] also attempts a dynamic compensation, however, quite differently. By defining a spliced vector comprising several adjacent feature vectors, a frame-by-frame compensation is achieved, thus reducing the computational effort.…”
Section: Introductionmentioning
confidence: 99%