2010
DOI: 10.1007/978-3-642-16202-2_6
|View full text |Cite
|
Sign up to set email alerts
|

Expansion of WFST-Based Dialog Management for Handling Multiple ASR Hypotheses

Abstract: We proposed a weighted finite-state transducer-based dialog manager (WFSTDM) which was a platform for expandable and adaptable dialog systems. In this platform, all rules and/or models for dialog management (DM) are expressed in WFST form, and the WFSTs are used to accomplish various tasks via multiple modalities. With this framework, we constructed a statistical dialog system using the user concept and system action tags which were acquired from an annotated corpus of human-to-human spoken dialogs as input an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 12 publications
(9 reference statements)
0
2
0
Order By: Relevance
“…In [12] a dialog system is described which transforms textual user utterances into response sentences using weighted FSTs, with the goal to be able to run a full back-and-forth dialog with the users. It was extended by [13] to accept n-best hypotheses from a triphone model acoustic model, which were combined with an additional 3-gram language model, as input. Eesen [14] introduced FST-decoding to models outputting character-based Connectionist Temporal Classification (CTC) [15] labels, similar to the Quartznet model of Scribosermo.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…In [12] a dialog system is described which transforms textual user utterances into response sentences using weighted FSTs, with the goal to be able to run a full back-and-forth dialog with the users. It was extended by [13] to accept n-best hypotheses from a triphone model acoustic model, which were combined with an additional 3-gram language model, as input. Eesen [14] introduced FST-decoding to models outputting character-based Connectionist Temporal Classification (CTC) [15] labels, similar to the Quartznet model of Scribosermo.…”
Section: Introductionmentioning
confidence: 99%
“…Alexa also uses FSTs for its skill kit, but keeps separate models for STT and NLU [16]. This work follows a very similar decoding approach as Eesen, which allows using recent CTC-based STT models (in difference to [10,11,13]), but alters the Grammar-FST (explained in the next chapters) to embed NLU information into it, similar to the semantic tagging of [9,10,11], which allows combining the two distinct STT+NLU models into a single SLU decoder.…”
Section: Introductionmentioning
confidence: 99%