2019
DOI: 10.48550/arxiv.1904.10635
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Better Automatic Evaluation of Open-Domain Dialogue Systems with Contextualized Embeddings

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
15
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(15 citation statements)
references
References 18 publications
0
15
0
Order By: Relevance
“…Learning-based metrics. This type of metric always consists of one or more training models, such as ADEM [24], RUBER [47], PONE [18], and BERT-RUBER [10]. Following previous work [10,23], we select BERT-RUBER as the sole representative learningbased metric in this study given its superior performance.…”
Section: Baseline Metricsmentioning
confidence: 99%
See 4 more Smart Citations
“…Learning-based metrics. This type of metric always consists of one or more training models, such as ADEM [24], RUBER [47], PONE [18], and BERT-RUBER [10]. Following previous work [10,23], we select BERT-RUBER as the sole representative learningbased metric in this study given its superior performance.…”
Section: Baseline Metricsmentioning
confidence: 99%
“…This type of metric always consists of one or more training models, such as ADEM [24], RUBER [47], PONE [18], and BERT-RUBER [10]. Following previous work [10,23], we select BERT-RUBER as the sole representative learningbased metric in this study given its superior performance. Since the performance of learning-based models could be influenced by the pre-prepared training dataset [23], we train and tune the model based on the specific dataset we use.…”
Section: Baseline Metricsmentioning
confidence: 99%
See 3 more Smart Citations