2010 IEEE Spoken Language Technology Workshop 2010
DOI: 10.1109/slt.2010.5700834
|View full text |Cite
|
Sign up to set email alerts
|

Query language modeling for voice search

Abstract: The paper presents an empirical exploration of google.com query stream language modeling. We describe the normalization of the typed query stream resulting in out-of-vocabulary (OoV) rates below 1% for a one million word vocabulary. We present a comprehensive set of experiments that guided the design decisions for a voice search service. In the process we re-discovered a less known interaction between Kneser-Ney smoothing and entropy pruning, and found empirical evidence that hints at non-stationarity of the q… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
19
0

Year Published

2011
2011
2019
2019

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 26 publications
(19 citation statements)
references
References 10 publications
(7 reference statements)
0
19
0
Order By: Relevance
“…We use maximum likelihood (ML) trained single mixture Gaussians for our weakAM. We use a baseline LM that is sufficiently small (21 million n-grams) to allow for sub-real time lattice generation on the training data with a small memory footprint, without compromising on its strength: as shown in [12], it takes much larger LMs to get a significant relative gain in WER.…”
Section: Experimental Setup and Resultsmentioning
confidence: 99%
“…We use maximum likelihood (ML) trained single mixture Gaussians for our weakAM. We use a baseline LM that is sufficiently small (21 million n-grams) to allow for sub-real time lattice generation on the training data with a small memory footprint, without compromising on its strength: as shown in [12], it takes much larger LMs to get a significant relative gain in WER.…”
Section: Experimental Setup and Resultsmentioning
confidence: 99%
“…For example, Google's English voice search model was trained from a 230 billion word corpus with 1 million unique words [9]. There are a few directions that researchers and engineers have been recently exploring to improve the quality of the LM for voice search applications.…”
Section: Language Modelingmentioning
confidence: 99%
“…The first one is the scarcity of training data. For instance, in the voice search task, it is easy to collect data corresponding to the target task by accessing text-input query logs [8]. In the spontaneous speech task, on the other hand, the data corresponding to the target task must be obtained by manually transcribing speech.…”
Section: Introductionmentioning
confidence: 99%