2022
DOI: 10.1609/aaai.v36i10.21320
|View full text |Cite
|
Sign up to set email alerts
|

GALAXY: A Generative Pre-trained Model for Task-Oriented Dialog with Semi-supervised Learning and Explicit Policy Injection

Abstract: Pre-trained models have proved to be powerful in enhancing task-oriented dialog systems. However, current pre-training methods mainly focus on enhancing dialog understanding and generation tasks while neglecting the exploitation of dialog policy. In this paper, we propose GALAXY, a novel pre-trained dialog model that explicitly learns dialog policy from limited labeled dialogs and large-scale unlabeled dialog corpora via semi-supervised learning. Specifically, we introduce a dialog act prediction task for poli… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
41
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
4
2

Relationship

2
8

Authors

Journals

citations
Cited by 52 publications
(42 citation statements)
references
References 67 publications
1
41
0
Order By: Relevance
“…Finally, we encourage the establishment of better dialog act taxonomies that are backed by learning sciences research. As outlined in §5.6 and in He et al (2022), a unified taxonomy may also strongly aid in transfer learning.…”
Section: User Study With a Learning Interfacementioning
confidence: 99%
“…Finally, we encourage the establishment of better dialog act taxonomies that are backed by learning sciences research. As outlined in §5.6 and in He et al (2022), a unified taxonomy may also strongly aid in transfer learning.…”
Section: User Study With a Learning Interfacementioning
confidence: 99%
“…SOLOIST [70] parameterized a task bot using a Transformer-based auto-regressive language model, which subsumed different dialog modules into a single neural model and was pre-trained on two TOD datasets. SPACE [31] proposed to use consistency regularization loss to learn dialog policy from labeled and unlabeled dialog corpora via a semi-supervised manner. To exploit more heterogeneous TOD corpora, PPTOD [88] converted different TOD tasks into the text-to-text generation task with taskspecific prompts on T5 model [76].…”
Section: Pre-trained Conversation Modelsmentioning
confidence: 99%
“…One possible solution is to employ a regularisation to encourage HSEMEC to consider intact post topics and emotion information to generate both coherent and emotional responses. In addition, dialogue pre-training techniques [40,41] may help the dialogue systems learn better dialogue patterns.…”
Section: Incoherent Responsesmentioning
confidence: 99%