Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.273
|View full text |Cite
|
Sign up to set email alerts
|

MinTL: Minimalist Transfer Learning for Task-Oriented Dialogue Systems

Abstract: In this paper, we propose Minimalist Transfer Learning (MinTL) to simplify the system design process of task-oriented dialogue systems and alleviate the over-dependency on annotated data. MinTL is a simple yet effective transfer learning framework, which allows us to plug-and-play pre-trained seq2seq models, and jointly learn dialogue state tracking and dialogue response generation. Unlike previous approaches, which use a copy mechanism to "carryover" the old dialogue states to the new one, we introduce Levens… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
105
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
4
1

Relationship

2
8

Authors

Journals

citations
Cited by 89 publications
(107 citation statements)
references
References 32 publications
(38 reference statements)
1
105
1
Order By: Relevance
“…In this paper, we model task-oriented dialogue systems as a seq2seq generation task (Lei et al, 2018;Lin et al, 2020b;Byrne et al, 2020;Lin et al, 2021) that generates both API-calls and system responses. As shown in Figure 2, the model takes as input a dialogue history, which is the concate-nation of user intents and current dialogue states, and then uses its API-call returns, which can be empty or system speech-acts, to generate its system response.…”
Section: Task-oriented Dialogue Modellingmentioning
confidence: 99%
“…In this paper, we model task-oriented dialogue systems as a seq2seq generation task (Lei et al, 2018;Lin et al, 2020b;Byrne et al, 2020;Lin et al, 2021) that generates both API-calls and system responses. As shown in Figure 2, the model takes as input a dialogue history, which is the concate-nation of user intents and current dialogue states, and then uses its API-call returns, which can be empty or system speech-acts, to generate its system response.…”
Section: Task-oriented Dialogue Modellingmentioning
confidence: 99%
“…RL methods have been used effectively to optimize end-to-end DSs in (Dhingra et al, 2017;Zhao et al, 2019), although using rule-based USs or a fixed corpus for interaction. Recent works utilise powerful transformers such as GPT-2 (Peng et al, 2020;Hosseini-Asl et al, 2020) or T5 (Lin et al, 2020b) for dialogue modeling and reach stateof-the-art performance; however, the area of having a user simulator involved during training is unexplored. By comparison, this work uses a learned US as the environment for RL.…”
Section: Related Workmentioning
confidence: 99%
“…Dialogue Systems are categorized into chit-chat (Vinyals and Le, 2015;Serban et al, 2016) and task-oriented (Williams and Young, 2007;Young et al, 2013); in this paper we focus on the latter. Task-oriented dialogue systems are further classified into: modularized (Levin et al, 2000;Hori et al, 2009;Lee et al, 2009), retrieval (Henderson et al, 2019; end-to-end (Bordes and Weston, 2017;Eric et al, 2017a;Eric and Manning, 2017;Madotto et al, 2018;Madotto et al, 2020a;Neelakantan et al, 2019;He et al, 2020) and hybrid (Shu et al, 2018;Lei et al, 2018;Zhang et al, 2019a;Mehri et al, 2019;Peng et al, 2020a;Ham et al, 2020;Hosseini-Asl et al, 2020;Le et al, 2020;Lin et al, 2020). To the best of our knowledge, these methods use either DST/S-ACT annotations, template responses, or all/partial KB as the input to the model, where instead we only use the dialogue history.…”
Section: Related Workmentioning
confidence: 99%