2006 IEEE Spoken Language Technology Workshop 2006
DOI: 10.1109/slt.2006.326844
|View full text |Cite
|
Sign up to set email alerts
|

Online Supervised Learning of Non-Understanding Recovery Policies

Abstract: Spoken dialog systems typically use a limited number of nonunderstanding recovery strategies and simple heuristic policies 1 to engage them (e.g. first ask user to repeat, then give help, then transfer to an operator). We propose a supervised, online method for learning a non-understanding recovery policy over a large set of recovery strategies. The approach consists of two steps: first, we construct runtime estimates for the likelihood of success of each recovery strategy, and then we use these estimates to c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2009
2009
2021
2021

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 17 publications
(8 citation statements)
references
References 6 publications
0
8
0
Order By: Relevance
“…These features help in capturing the task-specific characteristics of the events when social behavior should be performed. The 37 steps in the tutor's interaction plan are represented as binary features [Bohus et al 2006] indicating which step of the interaction is being executed. We included the information about the dialog state before and after the event using this feature representation.…”
Section: Featuresmentioning
confidence: 99%
See 1 more Smart Citation
“…These features help in capturing the task-specific characteristics of the events when social behavior should be performed. The 37 steps in the tutor's interaction plan are represented as binary features [Bohus et al 2006] indicating which step of the interaction is being executed. We included the information about the dialog state before and after the event using this feature representation.…”
Section: Featuresmentioning
confidence: 99%
“…The most popular current application of such agents is information access typically using voice-based IO over telephones and mobile devices. At the core of these agents are models that represent the structure of the interaction [Rudnicky and Xu 1999;Freedman 2000;Bangalore et al 2006] and perform functions such as belief update [Williams 2007], confidence annotation [Bohus and Rudnicky 2007], timing [Raux and Eskenazi 2009], and action selection [Paek and Horvitz 2000;Bohus et al 2006;Williams and Young 2007].…”
Section: Introductionmentioning
confidence: 99%
“…While some authors have made progress towards learning dialogue behaviour from human-machine interaction, they still rely on some form of simulation or delayed re-training. [11], for example, describe a spoken dialogue system that learns to optimize its non-understanding recovery strategies on-line through interactions with human users based on pre-trained logistic regression models. The system is re-trained every day.…”
Section: From Virtual Assistants To Interactive Conversational Robotsmentioning
confidence: 99%
“…Early work that addressed confidence annotation of input to the dialogue manager illustrated how low-level ASR features interact with features from natural language understanding and dialogue management. Confidence annotation has been done through linear regression models whose predictors include parse and dialogue state, relying on at most a dozen features to learn a binary threshold [41], [42]. When learning was done over multiple days for nine recovery strategies, system performance improved [42].…”
Section: Related Workmentioning
confidence: 99%
“…Confidence annotation has been done through linear regression models whose predictors include parse and dialogue state, relying on at most a dozen features to learn a binary threshold [41], [42]. When learning was done over multiple days for nine recovery strategies, system performance improved [42]. Confidence scores can also be continuous.…”
Section: Related Workmentioning
confidence: 99%