Interspeech 2018 2018
DOI: 10.21437/interspeech.2018-2527
|View full text |Cite
|
Sign up to set email alerts
|

Conversational Analysis Using Utterance-level Attention-based Bidirectional Recurrent Neural Networks

Abstract: Recent approaches for dialogue act recognition have shown that context from preceding utterances is important to classify the subsequent one. It was shown that the performance improves rapidly when the context is taken into account. We propose an utterance-level attention-based bidirectional recurrent neural network (Utt-Att-BiRNN) model to analyze the importance of preceding utterances to classify the current one. In our setup, the BiRNN is given the input set of current and preceding utterances. Our model ou… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
32
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 20 publications
(40 citation statements)
references
References 25 publications
0
32
0
Order By: Relevance
“…Bi-directional and Multi-layer Recurrent Models. In addition to our baseline recurrent models (LSTM and GRU), we also test their bi-directional and multi-layer variants, both of which have previously been explored within DA classification studies (Kumar et al 2017;Bothe et al 2018a;Chen et al 2018;Ribeiro et al 2019). The bi-directional models (Bi-LSTM and Bi-GRU) process the input sequence in the forwards and then backwards directions.…”
Section: Supervised Model Variantsmentioning
confidence: 99%
See 1 more Smart Citation
“…Bi-directional and Multi-layer Recurrent Models. In addition to our baseline recurrent models (LSTM and GRU), we also test their bi-directional and multi-layer variants, both of which have previously been explored within DA classification studies (Kumar et al 2017;Bothe et al 2018a;Chen et al 2018;Ribeiro et al 2019). The bi-directional models (Bi-LSTM and Bi-GRU) process the input sequence in the forwards and then backwards directions.…”
Section: Supervised Model Variantsmentioning
confidence: 99%
“…In some cases, the contextual information may also include 'future' utterances or DA labels, in other words, those that appear after the current utterance requiring classification; though, the utility of such future information for real-time applications such as dialogue systems is questionable. Within DA classification research, it has been widely shown that including such contextual information yields improved performance over single-sentence approaches (Lee and Dernoncourt 2016;Liu and Lane 2017;Bothe et al 2018a;Ribeiro et al 2019). The advantage of including contextual information is clear when considering the nature of dialogue as a sequence of utterances.…”
Section: Introductionmentioning
confidence: 99%
“…Figure 1. Attention mechanism (Bothe et al, 2018) In the context of our work, the attention mechanism gives RNNs a look at the whole sentence, which would not be the case without it. This mechanism helps the network decide which words to give greater weight to, as Bahdanau et al (2014) showed.…”
Section: Attention Mechanismmentioning
confidence: 99%
“…Following the majority of previous works, we produce frame-level predictions in this work. As attention mechanism has achieved excellent performance in various tasks like emotion recognition [10], conversational analysis [11], and speech recognition [12], we assume it can also benefit AED as it helps Liang He is the corresponding author. the model focus on when target events take place.…”
Section: Introductionmentioning
confidence: 99%