Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.148
|View full text |Cite
|
Sign up to set email alerts
|

Structured Attention for Unsupervised Dialogue Structure Induction

Abstract: Inducing a meaningful structural representation from one or a set of dialogues is a crucial but challenging task in computational linguistics. Advancement made in this area is critical for dialogue system design and discourse analysis. It can also be extended to solve grammatical inference. In this work, we propose to incorporate structured attention layers into a Variational Recurrent Neural Network (VRNN) model with discrete latent states to learn dialogue structure in an unsupervised fashion. Compared to a … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

2
7

Authors

Journals

citations
Cited by 17 publications
(32 citation statements)
references
References 21 publications
(22 reference statements)
0
10
0
Order By: Relevance
“…Existing works put more emphasis on unsupervised learning of dialogue structures. Representative ones include training language models based on Hidden Markov Models (HMMs) (Chotimongkol, 2008) or Varia-tional Recurrent Neural Networks (VRNNs) (Shi et al, 2019;Qiu et al, 2020) to reconstruct the original dialogues. The structure built upon the latent states is then evaluated in downstream tasks like dialogue policy learning.…”
Section: Dialoguementioning
confidence: 99%
See 1 more Smart Citation
“…Existing works put more emphasis on unsupervised learning of dialogue structures. Representative ones include training language models based on Hidden Markov Models (HMMs) (Chotimongkol, 2008) or Varia-tional Recurrent Neural Networks (VRNNs) (Shi et al, 2019;Qiu et al, 2020) to reconstruct the original dialogues. The structure built upon the latent states is then evaluated in downstream tasks like dialogue policy learning.…”
Section: Dialoguementioning
confidence: 99%
“…By reconstructing the original dialogues with discrete latent variable models, we can extract a structure representing the transition among the variables. In this direction, people have tried Hidden Markov Models (Chotimongkol, 2008;Ritter et al, 2010;Zhai and Williams, 2014), Variational Auto-Encoders (VAEs) (Kingma and Welling, 2013), and its recurrent version Variational Recurrent Neural Networks (VRNNs) (Chung et al, 2015;Shi et al, 2019;Qiu et al, 2020) Cann et al, 2017;Howard and Ruder, 2018;Peters et al, 2018;Devlin et al, 2019), the Transformer architecture (Vaswani et al, 2017) can be trained on generic corpora and adapted to specific downstream tasks. In dialogue systems, Wu et al (2020) pre-trained the BERT model (Devlin et al, 2019) on task-oriented dialogues for intent recognition, dialogue state tracking, dialogue act prediction, and response selection.…”
Section: Related Workmentioning
confidence: 99%
“…The research into dialog history modeling is much more extensive for open-domain dialog than task-oriented dialog (Tian et al, 2017). In the former, dialog history representation has been explored by, e.g., representing the entire dialog history as a linear sequence of tokens (Sordoni et al, 2015), using a fixed-size window to represent only the recent dialog history (Li et al, 2016), designing hierarchical repre- sentations (Serban et al, 2016;Xing et al, 2018;, leveraging structured attention (Qiu et al, 2020;Su et al, 2019) as well as summarizing (Xu et al, 2021) or re-writing (Xu et al, 2020) dialog history to handle long dialogs.…”
Section: Context Modeling For Dialogmentioning
confidence: 99%
“…Both are related to understanding multi-party conversation structures but they are different tasks. Dialogue structure learning aims to discover latent dialogue topics and construct an implicit utterance dependency tree to represent a multi-party dialogue's turn taking (Qiu et al, 2020), while the goal of conversation disentanglement is to learn an explicit dividing scheme that separates intermingled messages into sessions.…”
Section: Related Workmentioning
confidence: 99%