2004
DOI: 10.1145/1035112.1035113
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating discourse understanding in spoken dialogue systems

Abstract: This paper describes a method for creating an evaluation measure for discourse understanding in spoken dialogue systems. Discourse understanding means utterance understanding taking the context into account. Since the measure needs to be determined based on its correlation with the system's performance, conventional measures, such as the concept error rate, cannot be easily applied. Using the multiple linear regression analysis, we have previously shown that the weighted sum of various metrics concerning dialo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2006
2006
2016
2016

Publication Types

Select...
3
3
3

Relationship

4
5

Authors

Journals

citations
Cited by 14 publications
(10 citation statements)
references
References 12 publications
0
10
0
Order By: Relevance
“…Note that, although the values of correlation ratio seem rather low, the correlation often becomes low when it comes to subjective judgments (Higashinaka et al, 2004). Considering that we deal with chat-oriented dialogues, which are less restricted than task-oriented ones, we consider the current values of correlation ratio to be acceptable.…”
Section: Analyzing the Impact Of Error Typesmentioning
confidence: 99%
“…Note that, although the values of correlation ratio seem rather low, the correlation often becomes low when it comes to subjective judgments (Higashinaka et al, 2004). Considering that we deal with chat-oriented dialogues, which are less restricted than task-oriented ones, we consider the current values of correlation ratio to be acceptable.…”
Section: Analyzing the Impact Of Error Typesmentioning
confidence: 99%
“…Two metrics, Update precision and Update accuracy measure the accuracy and precision of updates to the top scoring hypothesis from one turn to the next. For more details, see Higashinaka et al (2004), which finds these metrics to be highly correlated with dialog success in their data.…”
Section: Tracker Output and Evaluation Metricsmentioning
confidence: 99%
“…1 Architecture of a spoken dialogue system. This figure is a modified version of the diagram we used in [11].…”
Section: Discourse Understanding In Spoken Dialogue Systemsmentioning
confidence: 99%
“…Although Higashinaka et al [11] proposed creating an evaluation measure for discourse understanding by finding a measure that correlates closely with the performance of a dialogue system, the measure assumes that the system holds a single dialogue state. The best measure they propose is based on the precision of the update of a dialogue state (called update precision), which is difficult to calculate when a system has multiple dialogue states because it is not clear whether the sequence of dialogue states with a different understanding history can be used to calculate the update precision.…”
Section: Concept Error Rate (Cer)mentioning
confidence: 99%