IEEE Workshop on Automatic Speech Recognition and Understanding, 2001. ASRU '01.
DOI: 10.1109/asru.2001.1034620
|View full text |Cite
|
Sign up to set email alerts
|

Incremental language models for speech recognition using finite-state transducers

Abstract: In the context of the weighted finite-state transducer approach to speech recognition, we investigate a novel decoding strategy to deal with very large n-gram language models often used in large-vocabulary systems. In particular, we present an alternative to full, static expansion and optimization of the finite-state transducer network. This alternative is useful when the individual knowledge sources, modeled as transducers, are too large to be composed and optimized. While the recognition decoder perceives a … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 30 publications
(8 citation statements)
references
References 5 publications
(10 reference statements)
0
8
0
Order By: Relevance
“…The most commonly used asynchronous method is A * [8,9,10]. • in the field of WFST-based speech recognition, several algorithms have been proposed in order to reduce time and memory problem: [11], [12] propose on the fly composition algorithms while [13] propose to factor the language models into smaller components.…”
Section: Classical Approachesmentioning
confidence: 99%
“…The most commonly used asynchronous method is A * [8,9,10]. • in the field of WFST-based speech recognition, several algorithms have been proposed in order to reduce time and memory problem: [11], [12] propose on the fly composition algorithms while [13] propose to factor the language models into smaller components.…”
Section: Classical Approachesmentioning
confidence: 99%
“…The WFST Guni represents a uni-gram language model and WFST G tri/uni represents a tri-gram model divided by the uni-gram probability. This scheme as studied in [3] allows for the inclusion of some static language model information and the addition of this look-ahead information can improve the search performance.…”
Section: Wfst Combinations Evaluatedmentioning
confidence: 99%
“…A drawback of this WFST approach is access to original knowledge sources is lost once the final network has been composed and optimized. On-the-fly composition and optimization algorithms have been developed by others [3,4,5,6] as a method of increasing flexibility within the WFST paradigm. However, one disadvantage with such on-line algorithms is some of the optimization powers available in the static equivalents are sacrificed.…”
Section: Introductionmentioning
confidence: 99%
“…One of the possible solutions to this problem is to perform on-the-fly transducer composition during decoding. Acoustical, phonetic and lexical resources may still be composed and optimised off-line, while the language model transducer is locally, dynamically composed at run time [3,19,9]. By using this approach, we can avoid composing part of the search space which is not traversed by any hypotheses.…”
Section: Future Developmentmentioning
confidence: 99%