Proceedings of the 23rd Annual Meeting of the Special Interest Group on Discourse and Dialogue 2022
DOI: 10.18653/v1/2022.sigdial-1.30
|View full text |Cite
|
Sign up to set email alerts
|

A Systematic Evaluation of Response Selection for Open Domain Dialogue

Behnam Hedayatnia,
Di Jin,
Yang Liu
et al.

Abstract: Recent progress on neural approaches for language processing has triggered a resurgence of interest on building intelligent open-domain chatbots. However, even the state-of-the-art neural chatbots cannot produce satisfying responses for every turn in a dialog. A practical solution is to generate multiple response candidates for the same context, and then perform response ranking/selection to determine which candidate is the best. Previous work in response selection typically trains response rankers using synth… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
0
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 29 publications
0
0
0
Order By: Relevance
“…In a study by Hedayatnia et al (2022), they demonstrated that using a human-chatbot dataset, where responses were generated by multiple response generators and then annotated by humans for training RS (response selection) models, led to improved performance compared to models trained on synthetically generated datasets. Unfortunately, the dataset they used could not be made public due to privacy concerns, as it contained real-user dialogs.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…In a study by Hedayatnia et al (2022), they demonstrated that using a human-chatbot dataset, where responses were generated by multiple response generators and then annotated by humans for training RS (response selection) models, led to improved performance compared to models trained on synthetically generated datasets. Unfortunately, the dataset they used could not be made public due to privacy concerns, as it contained real-user dialogs.…”
Section: Related Workmentioning
confidence: 99%
“…These synthetically curated test sets are not sufficient representations of real-world inference time candidates that are generated by dialog models. Hedayatnia et al (2022) demonstrated the effectiveness of training on machine-generated candidates from real user interactions over using synthetic candidates for response selection. However this data is not publicly available.…”
Section: Introductionmentioning
confidence: 99%