Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.13
|View full text |Cite
|
Sign up to set email alerts
|

Transferable Dialogue Systems and User Simulators

Abstract: One of the difficulties in training dialogue systems is the lack of training data. We explore the possibility of creating dialogue data through the interaction between a dialogue system and a user simulator. Our goal is to develop a modelling framework that can incorporate new dialogue scenarios through self-play between the two agents. In this framework, we first pre-train the two agents on a collection of source domain dialogues, which equips the agents to converse with each other via natural language. With … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 23 publications
(12 citation statements)
references
References 23 publications
0
9
0
Order By: Relevance
“…We calculate the turn accuracy (ACC), joint goal accuracy (JGA) and combined scores (Comb) for these tasks, and compute the average score as the overall metric to measure the model performance. Due to few-shot learning gains increasing attention in various dialog tasks to assess model capability [14,26,90], we conduct experiments under both a full-data setting that uses all training data to fine-tune the models and a few-shot setting that uses only 10% of training data to fine-tune the 3, SPACE-3 outperforms all baselines on all datasets. On the full-data setting and the few-shot setting, SPACE-3 surpasses the PPTOD * with 1.75 and 2.48 absolute average score improvement respectively, indicating the SPACE-3 have better adaptability on all types of dialog tasks.…”
Section: Overall Comparison With Pcmsmentioning
confidence: 99%
“…We calculate the turn accuracy (ACC), joint goal accuracy (JGA) and combined scores (Comb) for these tasks, and compute the average score as the overall metric to measure the model performance. Due to few-shot learning gains increasing attention in various dialog tasks to assess model capability [14,26,90], we conduct experiments under both a full-data setting that uses all training data to fine-tune the models and a few-shot setting that uses only 10% of training data to fine-tune the 3, SPACE-3 outperforms all baselines on all datasets. On the full-data setting and the few-shot setting, SPACE-3 surpasses the PPTOD * with 1.75 and 2.48 absolute average score improvement respectively, indicating the SPACE-3 have better adaptability on all types of dialog tasks.…”
Section: Overall Comparison With Pcmsmentioning
confidence: 99%
“…Wang et al (2020) regarded the dialogue act as a sequence generation task to assist the system in generating replies at each step of the dialogue process. To focus on more diverse dialogue acts, Zhang et al (2020c) devised a data augmentation method that considers multiple dialogue acts to generate system responses simultaneously Tseng et al (2021) designed to first capture the correct dialogue act through the dialogue state, then represent it as a sequence of tokens and generate it through an LSTM. Their strategy for learning dialogue acts can handle multiple different acts during a conversation simultaneously.…”
Section: Response Generationmentioning
confidence: 99%
“…JOUST (Tseng et al, 2021) designed a strategy to obtain the current dialogue act through the dialog state. We denote one of their methods as JOUST+ RL-turn-R. NoisyChannel (Liu et al, 2021) uses the noisy channel model (Yu et al, 2017) to decode dialogue act to generate a higher quality response.…”
Section: End-to-end Modelmentioning
confidence: 99%
“…1 Our code is publicly available at https://github. com/nu-dialogue/post-processing-networks However, since each module is processed sequentially, errors in the preceding module can easily propagate to the following ones, and the performance of the entire system cannot be optimized (Tseng et al, 2021). This results in low dialogue performance of the entire system .…”
Section: Introductionmentioning
confidence: 99%