Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Confere 2015
DOI: 10.3115/v1/p15-2130
|View full text |Cite
|
Sign up to set email alerts
|

Multi-domain Dialog State Tracking using Recurrent Neural Networks

Abstract: Dialog state tracking is a key component of many modern dialog systems, most of which are designed with a single, welldefined domain in mind. This paper shows that dialog data drawn from different dialog domains can be used to train a general belief tracking model which can operate across all of these domains, exhibiting superior performance to each of the domainspecific models. We propose a training procedure which uses out-of-domain data to initialise belief tracking models for entirely new domains. This pro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
78
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
5
5

Relationship

3
7

Authors

Journals

citations
Cited by 151 publications
(80 citation statements)
references
References 11 publications
(19 reference statements)
2
78
0
Order By: Relevance
“…2.1.2 Hard and Soft Parameter Sharing. Most MTL approaches share the same base structure for feature extraction [1,9,12,18,26,27,33,[51][52][53] and then continue to branch out, intertwine or widen the model's parameter space. Sharing is an essential part of MTL and can be categorized as hard sharing or soft sharing.…”
Section: 11mentioning
confidence: 99%
“…2.1.2 Hard and Soft Parameter Sharing. Most MTL approaches share the same base structure for feature extraction [1,9,12,18,26,27,33,[51][52][53] and then continue to branch out, intertwine or widen the model's parameter space. Sharing is an essential part of MTL and can be categorized as hard sharing or soft sharing.…”
Section: 11mentioning
confidence: 99%
“…In machine learning in general much research has looked at adaptation of statistical models [21,22,23] however research into adaptation of SDS components to new domains [24,25,26,27,28] or user behaviour [29] presents its own challenges and is comparatively nascent. Research into these questions is growing though [30], and will continue to given the natural progression towards multi-domain SDS [31,32,33].…”
Section: Related Workmentioning
confidence: 99%
“…These models depend on delexicalization, using generic tags to replace specific slot types and values, and handcrafted semantic dictionaries. In practice, it is difficult to scale these models for every slot type and recent state-of-the-art models for DST use deep learning based methods to learn general representations for user and system utterances and previous system actions, and predict the turn state (Henderson et al, 2013(Henderson et al, , 2014bMrkšić et al, 2015Hori et al, 2016;Liu and Lane, 2017;Dernoncourt et al, 2017;Chen et al, 2016). However, these systems are found to perform poorly on rare and unknown slot-value pairs which was recently addressed through local slot-specific encoders (Zhong et al, 2018) and pointer network (Xu and Hu, 2018).…”
Section: Related Workmentioning
confidence: 99%