1995 International Conference on Acoustics, Speech, and Signal Processing
DOI: 10.1109/icassp.1995.479408
|View full text |Cite
|
Sign up to set email alerts
|

Non-deterministic stochastic language models for speech recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
29
0
2

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 33 publications
(31 citation statements)
references
References 3 publications
0
29
0
2
Order By: Relevance
“…We denote by P(w, ′ ) the probability of generating the word w and visiting the state ′ given that the state is occupied. There is no requirement that the language model be deterministic (in the sense of Riccardi et al, 1995). One advantage of accommodating a general finite state machine as a language model is that it enables us to use our search software for training as well as recognition.…”
Section: The Language Modelmentioning
confidence: 99%
“…We denote by P(w, ′ ) the probability of generating the word w and visiting the state ′ given that the state is occupied. There is no requirement that the language model be deterministic (in the sense of Riccardi et al, 1995). One advantage of accommodating a general finite state machine as a language model is that it enables us to use our search software for training as well as recognition.…”
Section: The Language Modelmentioning
confidence: 99%
“…Two of the methods-Variable Length Markov Models (VLMMs) (Guyon & Pereira, 1995) and Variable N-gram Stochastic Automata (VNSA) (Riccardi, Bocchieri & Pieraccini, 1995) are the most recent developments reported in the literature. The third method, called Refined Probabilistic Finite Automata (RPFA), which will be introduced in this paper, is based on the highly successful text compression algorithm called Dynamic Markov Compression (DMC) (Cormack & Horspool, 1987).…”
Section: Introductionmentioning
confidence: 99%
“…Then an SLU module that uses the output of the ASR is trained and optimized for the understanding performance. However, as it has been pointed out many times, the hypothesis that gives a better recognition performance does not always yield a better understanding performance [108,127,38]. If the end goal of SLS is to understand what the user means and respond accordingly, both modules can be optimized jointly such that the system "understands better what it recognizes" and "recognizes better what it understands".…”
Section: Motivationmentioning
confidence: 99%
“…SELMs are important because as it has been argued multiple times [108,127] improving WER does not always yield an improvement on the understanding performance. This is especially important for spoken language systems which extracts semantic structures out of the utterances.…”
Section: Chapter 9 Deep Encodings For Semantic Language Modelsmentioning
confidence: 99%