2003
DOI: 10.1016/s0885-2308(02)00052-9
|View full text |Cite
|
Sign up to set email alerts
|

The effect of pruning and compression on graphical representations of the output of a speech recognizer

Abstract: Large vocabulary continuous speech recognition can benefit from an efficient data structure for representing a large number of acoustic hypotheses compactly. Word graphs or lattices have been chosen as such an efficient interface between acoustic recognition engines and subsequent language processing modules. This paper first investigates the effect of pruning during acoustic decoding on the quality of word lattices and shows that by combining different pruning options (at the model level and word level), we c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2008
2008
2010
2010

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 23 publications
(23 reference statements)
0
2
0
Order By: Relevance
“…Additional simplification algorithms have been proposed to assist in robot path planning [17], classifying the topology of surfaces [3], speech recognition [13], and improving the computational complexity and memory requirements of dense graph processing algorithms [12,18,6].…”
Section: Related Workmentioning
confidence: 99%
“…Additional simplification algorithms have been proposed to assist in robot path planning [17], classifying the topology of surfaces [3], speech recognition [13], and improving the computational complexity and memory requirements of dense graph processing algorithms [12,18,6].…”
Section: Related Workmentioning
confidence: 99%
“…If no pruning is done, the lattice can be highly accurate but also exorbitantly large. One of the most effective lattice pruning methods is the beam pruning during decoding [5]. Another very common method is to further compress the lattice after it has been generated, where a word lattice is usually converted to a word graph and all the time alignment information is discarded, such as the finite state automata determination and minimization [6] and the confusion network approach [7].…”
Section: Introductionmentioning
confidence: 99%